diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7328 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 200.0, + "global_step": 5205, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00019212295869356388, + "grad_norm": 27.250099182128906, + "learning_rate": 1.1494252873563217e-06, + "loss": 12.760068893432617, + "step": 1 + }, + { + "epoch": 0.0009606147934678194, + "grad_norm": 27.193613052368164, + "learning_rate": 5.747126436781608e-06, + "loss": 12.752052307128906, + "step": 5 + }, + { + "epoch": 0.0019212295869356388, + "grad_norm": 23.87885093688965, + "learning_rate": 1.1494252873563217e-05, + "loss": 12.383438110351562, + "step": 10 + }, + { + "epoch": 0.002881844380403458, + "grad_norm": 15.234784126281738, + "learning_rate": 1.7241379310344825e-05, + "loss": 11.199150085449219, + "step": 15 + }, + { + "epoch": 0.0038424591738712775, + "grad_norm": 15.339820861816406, + "learning_rate": 2.2988505747126433e-05, + "loss": 9.962273406982423, + "step": 20 + }, + { + "epoch": 0.004803073967339097, + "grad_norm": 7.6476054191589355, + "learning_rate": 2.8735632183908045e-05, + "loss": 8.961764526367187, + "step": 25 + }, + { + "epoch": 0.005763688760806916, + "grad_norm": 5.635213851928711, + "learning_rate": 3.448275862068965e-05, + "loss": 8.127005004882813, + "step": 30 + }, + { + "epoch": 0.0067243035542747355, + "grad_norm": 4.58963680267334, + "learning_rate": 4.022988505747126e-05, + "loss": 7.473291015625, + "step": 35 + }, + { + "epoch": 0.007684918347742555, + "grad_norm": 2.4760098457336426, + "learning_rate": 4.5977011494252866e-05, + "loss": 6.995676422119141, + "step": 40 + }, + { + "epoch": 0.008645533141210375, + "grad_norm": 1.5408024787902832, + "learning_rate": 5.172413793103448e-05, + "loss": 6.636921691894531, + "step": 45 + }, + { + "epoch": 0.009606147934678195, + "grad_norm": 1.112014651298523, + "learning_rate": 5.747126436781609e-05, + "loss": 6.349236297607422, + "step": 50 + }, + { + "epoch": 0.010566762728146013, + "grad_norm": 0.9658083319664001, + "learning_rate": 6.32183908045977e-05, + "loss": 6.117427825927734, + "step": 55 + }, + { + "epoch": 0.011527377521613832, + "grad_norm": 1.8757588863372803, + "learning_rate": 6.89655172413793e-05, + "loss": 5.927629852294922, + "step": 60 + }, + { + "epoch": 0.012487992315081652, + "grad_norm": 1.7964160442352295, + "learning_rate": 7.471264367816091e-05, + "loss": 5.770274353027344, + "step": 65 + }, + { + "epoch": 0.013448607108549471, + "grad_norm": 2.1291167736053467, + "learning_rate": 8.045977011494252e-05, + "loss": 5.6375282287597654, + "step": 70 + }, + { + "epoch": 0.01440922190201729, + "grad_norm": 3.0542609691619873, + "learning_rate": 8.620689655172413e-05, + "loss": 5.525964736938477, + "step": 75 + }, + { + "epoch": 0.01536983669548511, + "grad_norm": 3.1441893577575684, + "learning_rate": 9.195402298850573e-05, + "loss": 5.427687072753907, + "step": 80 + }, + { + "epoch": 0.01633045148895293, + "grad_norm": 1.3293317556381226, + "learning_rate": 9.770114942528733e-05, + "loss": 5.335831832885742, + "step": 85 + }, + { + "epoch": 0.01729106628242075, + "grad_norm": 2.560171127319336, + "learning_rate": 0.00010344827586206896, + "loss": 5.258418273925781, + "step": 90 + }, + { + "epoch": 0.01825168107588857, + "grad_norm": 2.012646436691284, + "learning_rate": 0.00010919540229885056, + "loss": 5.183747863769531, + "step": 95 + }, + { + "epoch": 0.01921229586935639, + "grad_norm": 2.6587507724761963, + "learning_rate": 0.00011494252873563218, + "loss": 5.119210052490234, + "step": 100 + }, + { + "epoch": 0.020172910662824207, + "grad_norm": 2.3047592639923096, + "learning_rate": 0.00012068965517241378, + "loss": 5.057864761352539, + "step": 105 + }, + { + "epoch": 0.021133525456292025, + "grad_norm": 2.4375789165496826, + "learning_rate": 0.0001264367816091954, + "loss": 5.000635147094727, + "step": 110 + }, + { + "epoch": 0.022094140249759846, + "grad_norm": 2.1402318477630615, + "learning_rate": 0.000132183908045977, + "loss": 4.947456359863281, + "step": 115 + }, + { + "epoch": 0.023054755043227664, + "grad_norm": 2.815868377685547, + "learning_rate": 0.0001379310344827586, + "loss": 4.898051071166992, + "step": 120 + }, + { + "epoch": 0.024015369836695485, + "grad_norm": 2.3648743629455566, + "learning_rate": 0.0001436781609195402, + "loss": 4.855741119384765, + "step": 125 + }, + { + "epoch": 0.024975984630163303, + "grad_norm": 2.230957508087158, + "learning_rate": 0.00014942528735632183, + "loss": 4.808645629882813, + "step": 130 + }, + { + "epoch": 0.025936599423631124, + "grad_norm": 1.3839224576950073, + "learning_rate": 0.00015517241379310346, + "loss": 4.757843780517578, + "step": 135 + }, + { + "epoch": 0.026897214217098942, + "grad_norm": 1.4279674291610718, + "learning_rate": 0.00016091954022988503, + "loss": 4.719438171386718, + "step": 140 + }, + { + "epoch": 0.027857829010566763, + "grad_norm": 1.4759711027145386, + "learning_rate": 0.00016666666666666666, + "loss": 4.6750732421875, + "step": 145 + }, + { + "epoch": 0.02881844380403458, + "grad_norm": 1.3809517621994019, + "learning_rate": 0.00017241379310344826, + "loss": 4.63501091003418, + "step": 150 + }, + { + "epoch": 0.029779058597502402, + "grad_norm": 1.6614536046981812, + "learning_rate": 0.00017816091954022986, + "loss": 4.603597640991211, + "step": 155 + }, + { + "epoch": 0.03073967339097022, + "grad_norm": 2.904686689376831, + "learning_rate": 0.00018390804597701147, + "loss": 4.574514770507813, + "step": 160 + }, + { + "epoch": 0.03170028818443804, + "grad_norm": 1.8000333309173584, + "learning_rate": 0.0001896551724137931, + "loss": 4.542916107177734, + "step": 165 + }, + { + "epoch": 0.03266090297790586, + "grad_norm": 2.0026073455810547, + "learning_rate": 0.00019540229885057467, + "loss": 4.509420013427734, + "step": 170 + }, + { + "epoch": 0.03362151777137368, + "grad_norm": 1.6088820695877075, + "learning_rate": 0.0002011494252873563, + "loss": 4.4883583068847654, + "step": 175 + }, + { + "epoch": 0.0345821325648415, + "grad_norm": 3.137084722518921, + "learning_rate": 0.00020689655172413793, + "loss": 4.462603759765625, + "step": 180 + }, + { + "epoch": 0.03554274735830932, + "grad_norm": 1.9390242099761963, + "learning_rate": 0.00021264367816091953, + "loss": 4.441267395019532, + "step": 185 + }, + { + "epoch": 0.03650336215177714, + "grad_norm": 2.0831611156463623, + "learning_rate": 0.00021839080459770113, + "loss": 4.420868682861328, + "step": 190 + }, + { + "epoch": 0.037463976945244955, + "grad_norm": 1.4936137199401855, + "learning_rate": 0.00022413793103448273, + "loss": 4.390471649169922, + "step": 195 + }, + { + "epoch": 0.03842459173871278, + "grad_norm": 2.1609437465667725, + "learning_rate": 0.00022988505747126436, + "loss": 4.372611999511719, + "step": 200 + }, + { + "epoch": 0.0393852065321806, + "grad_norm": 1.5894837379455566, + "learning_rate": 0.00023563218390804593, + "loss": 4.3513038635253904, + "step": 205 + }, + { + "epoch": 0.040345821325648415, + "grad_norm": 2.7555136680603027, + "learning_rate": 0.00024137931034482756, + "loss": 4.324166870117187, + "step": 210 + }, + { + "epoch": 0.04130643611911623, + "grad_norm": 1.7152142524719238, + "learning_rate": 0.00024712643678160916, + "loss": 4.298264694213867, + "step": 215 + }, + { + "epoch": 0.04226705091258405, + "grad_norm": 2.7435386180877686, + "learning_rate": 0.0002528735632183908, + "loss": 4.284121322631836, + "step": 220 + }, + { + "epoch": 0.043227665706051875, + "grad_norm": 1.9910717010498047, + "learning_rate": 0.00025862068965517237, + "loss": 4.270275497436524, + "step": 225 + }, + { + "epoch": 0.04418828049951969, + "grad_norm": 1.8952627182006836, + "learning_rate": 0.000264367816091954, + "loss": 4.251060867309571, + "step": 230 + }, + { + "epoch": 0.04514889529298751, + "grad_norm": 1.4182472229003906, + "learning_rate": 0.0002701149425287356, + "loss": 4.223833465576172, + "step": 235 + }, + { + "epoch": 0.04610951008645533, + "grad_norm": 2.826347589492798, + "learning_rate": 0.0002758620689655172, + "loss": 4.208561706542969, + "step": 240 + }, + { + "epoch": 0.04707012487992315, + "grad_norm": 2.5493690967559814, + "learning_rate": 0.00028160919540229883, + "loss": 4.194873428344726, + "step": 245 + }, + { + "epoch": 0.04803073967339097, + "grad_norm": 2.2824618816375732, + "learning_rate": 0.0002873563218390804, + "loss": 4.167812347412109, + "step": 250 + }, + { + "epoch": 0.04899135446685879, + "grad_norm": 1.4718077182769775, + "learning_rate": 0.00029310344827586203, + "loss": 4.150156402587891, + "step": 255 + }, + { + "epoch": 0.049951969260326606, + "grad_norm": 2.2785093784332275, + "learning_rate": 0.00029885057471264366, + "loss": 4.126807403564453, + "step": 260 + }, + { + "epoch": 0.05091258405379443, + "grad_norm": 1.5848008394241333, + "learning_rate": 0.00029999951546647263, + "loss": 4.111884689331054, + "step": 265 + }, + { + "epoch": 0.05187319884726225, + "grad_norm": 1.6441746950149536, + "learning_rate": 0.0002999975470543828, + "loss": 4.089701843261719, + "step": 270 + }, + { + "epoch": 0.052833813640730067, + "grad_norm": 1.9900462627410889, + "learning_rate": 0.00029999406450023966, + "loss": 4.0772346496582035, + "step": 275 + }, + { + "epoch": 0.053794428434197884, + "grad_norm": 1.938680648803711, + "learning_rate": 0.0002999890678391978, + "loss": 4.0556285858154295, + "step": 280 + }, + { + "epoch": 0.05475504322766571, + "grad_norm": 1.6333754062652588, + "learning_rate": 0.00029998255712169563, + "loss": 4.044793701171875, + "step": 285 + }, + { + "epoch": 0.05571565802113353, + "grad_norm": 1.8412508964538574, + "learning_rate": 0.00029997453241345533, + "loss": 4.03118896484375, + "step": 290 + }, + { + "epoch": 0.056676272814601344, + "grad_norm": 1.4252992868423462, + "learning_rate": 0.0002999649937954818, + "loss": 4.01298713684082, + "step": 295 + }, + { + "epoch": 0.05763688760806916, + "grad_norm": 2.278615951538086, + "learning_rate": 0.0002999539413640621, + "loss": 3.9946434020996096, + "step": 300 + }, + { + "epoch": 0.05859750240153699, + "grad_norm": 1.5454790592193604, + "learning_rate": 0.0002999413752307644, + "loss": 3.9831748962402345, + "step": 305 + }, + { + "epoch": 0.059558117195004805, + "grad_norm": 1.334790587425232, + "learning_rate": 0.0002999272955224369, + "loss": 3.9761894226074217, + "step": 310 + }, + { + "epoch": 0.06051873198847262, + "grad_norm": 1.9382845163345337, + "learning_rate": 0.0002999117023812064, + "loss": 3.9631324768066407, + "step": 315 + }, + { + "epoch": 0.06147934678194044, + "grad_norm": 2.01582932472229, + "learning_rate": 0.000299894595964477, + "loss": 3.9521259307861327, + "step": 320 + }, + { + "epoch": 0.06243996157540826, + "grad_norm": 1.8427826166152954, + "learning_rate": 0.0002998759764449286, + "loss": 3.9423927307128905, + "step": 325 + }, + { + "epoch": 0.06340057636887608, + "grad_norm": 1.6123930215835571, + "learning_rate": 0.0002998558440105148, + "loss": 3.935696029663086, + "step": 330 + }, + { + "epoch": 0.0643611911623439, + "grad_norm": 1.9502496719360352, + "learning_rate": 0.0002998341988644614, + "loss": 3.923968505859375, + "step": 335 + }, + { + "epoch": 0.06532180595581172, + "grad_norm": 1.2648119926452637, + "learning_rate": 0.0002998110412252641, + "loss": 3.9131423950195314, + "step": 340 + }, + { + "epoch": 0.06628242074927954, + "grad_norm": 1.5318907499313354, + "learning_rate": 0.0002997863713266866, + "loss": 3.9074745178222656, + "step": 345 + }, + { + "epoch": 0.06724303554274735, + "grad_norm": 1.3065940141677856, + "learning_rate": 0.0002997601894177576, + "loss": 3.8965118408203123, + "step": 350 + }, + { + "epoch": 0.06820365033621517, + "grad_norm": 1.2381218671798706, + "learning_rate": 0.00029973249576276914, + "loss": 3.890979766845703, + "step": 355 + }, + { + "epoch": 0.069164265129683, + "grad_norm": 1.5115997791290283, + "learning_rate": 0.0002997032906412732, + "loss": 3.8882400512695314, + "step": 360 + }, + { + "epoch": 0.07012487992315082, + "grad_norm": 1.6171698570251465, + "learning_rate": 0.0002996725743480793, + "loss": 3.876806640625, + "step": 365 + }, + { + "epoch": 0.07108549471661864, + "grad_norm": 1.0753288269042969, + "learning_rate": 0.00029964034719325147, + "loss": 3.860057067871094, + "step": 370 + }, + { + "epoch": 0.07204610951008646, + "grad_norm": 1.4012713432312012, + "learning_rate": 0.0002996066095021048, + "loss": 3.857683563232422, + "step": 375 + }, + { + "epoch": 0.07300672430355427, + "grad_norm": 1.765068769454956, + "learning_rate": 0.0002995713616152028, + "loss": 3.8544296264648437, + "step": 380 + }, + { + "epoch": 0.07396733909702209, + "grad_norm": 1.5474210977554321, + "learning_rate": 0.0002995346038883532, + "loss": 3.849739837646484, + "step": 385 + }, + { + "epoch": 0.07492795389048991, + "grad_norm": 1.443384051322937, + "learning_rate": 0.0002994963366926048, + "loss": 3.845722961425781, + "step": 390 + }, + { + "epoch": 0.07588856868395773, + "grad_norm": 1.4993858337402344, + "learning_rate": 0.0002994565604142439, + "loss": 3.8389007568359377, + "step": 395 + }, + { + "epoch": 0.07684918347742556, + "grad_norm": 1.4684888124465942, + "learning_rate": 0.00029941527545478976, + "loss": 3.8267383575439453, + "step": 400 + }, + { + "epoch": 0.07780979827089338, + "grad_norm": 1.2325655221939087, + "learning_rate": 0.00029937248223099136, + "loss": 3.818804168701172, + "step": 405 + }, + { + "epoch": 0.0787704130643612, + "grad_norm": 1.0254677534103394, + "learning_rate": 0.00029932818117482245, + "loss": 3.812041473388672, + "step": 410 + }, + { + "epoch": 0.07973102785782901, + "grad_norm": 1.3646955490112305, + "learning_rate": 0.0002992823727334776, + "loss": 3.810979461669922, + "step": 415 + }, + { + "epoch": 0.08069164265129683, + "grad_norm": 1.6398770809173584, + "learning_rate": 0.00029923505736936774, + "loss": 3.8059127807617186, + "step": 420 + }, + { + "epoch": 0.08165225744476465, + "grad_norm": 1.2734476327896118, + "learning_rate": 0.0002991862355601151, + "loss": 3.798663330078125, + "step": 425 + }, + { + "epoch": 0.08261287223823247, + "grad_norm": 1.4652912616729736, + "learning_rate": 0.00029913590779854886, + "loss": 3.7902076721191404, + "step": 430 + }, + { + "epoch": 0.08357348703170028, + "grad_norm": 1.0962094068527222, + "learning_rate": 0.00029908407459269977, + "loss": 3.786347198486328, + "step": 435 + }, + { + "epoch": 0.0845341018251681, + "grad_norm": 1.054427981376648, + "learning_rate": 0.0002990307364657954, + "loss": 3.780842590332031, + "step": 440 + }, + { + "epoch": 0.08549471661863593, + "grad_norm": 1.2044267654418945, + "learning_rate": 0.0002989758939562545, + "loss": 3.7756649017333985, + "step": 445 + }, + { + "epoch": 0.08645533141210375, + "grad_norm": 1.3676034212112427, + "learning_rate": 0.0002989195476176818, + "loss": 3.7702369689941406, + "step": 450 + }, + { + "epoch": 0.08741594620557157, + "grad_norm": 1.455597996711731, + "learning_rate": 0.00029886169801886237, + "loss": 3.76458740234375, + "step": 455 + }, + { + "epoch": 0.08837656099903939, + "grad_norm": 1.4684407711029053, + "learning_rate": 0.00029880234574375576, + "loss": 3.7579875946044923, + "step": 460 + }, + { + "epoch": 0.0893371757925072, + "grad_norm": 1.4192808866500854, + "learning_rate": 0.00029874149139149037, + "loss": 3.757953643798828, + "step": 465 + }, + { + "epoch": 0.09029779058597502, + "grad_norm": 0.8747047781944275, + "learning_rate": 0.00029867913557635704, + "loss": 3.7510101318359377, + "step": 470 + }, + { + "epoch": 0.09125840537944284, + "grad_norm": 1.4938651323318481, + "learning_rate": 0.0002986152789278031, + "loss": 3.746034622192383, + "step": 475 + }, + { + "epoch": 0.09221902017291066, + "grad_norm": 1.2561233043670654, + "learning_rate": 0.00029854992209042626, + "loss": 3.743701171875, + "step": 480 + }, + { + "epoch": 0.09317963496637849, + "grad_norm": 1.431540846824646, + "learning_rate": 0.0002984830657239673, + "loss": 3.736903762817383, + "step": 485 + }, + { + "epoch": 0.0941402497598463, + "grad_norm": 1.0810083150863647, + "learning_rate": 0.00029841471050330424, + "loss": 3.732843017578125, + "step": 490 + }, + { + "epoch": 0.09510086455331412, + "grad_norm": 0.9057182669639587, + "learning_rate": 0.00029834485711844515, + "loss": 3.7288368225097654, + "step": 495 + }, + { + "epoch": 0.09606147934678194, + "grad_norm": 1.559844970703125, + "learning_rate": 0.00029827350627452116, + "loss": 3.72186393737793, + "step": 500 + }, + { + "epoch": 0.09702209414024976, + "grad_norm": 1.3021472692489624, + "learning_rate": 0.00029820065869177954, + "loss": 3.7243560791015624, + "step": 505 + }, + { + "epoch": 0.09798270893371758, + "grad_norm": 0.9313806891441345, + "learning_rate": 0.0002981263151055762, + "loss": 3.7176589965820312, + "step": 510 + }, + { + "epoch": 0.0989433237271854, + "grad_norm": 1.2589117288589478, + "learning_rate": 0.0002980504762663683, + "loss": 3.712178039550781, + "step": 515 + }, + { + "epoch": 0.09990393852065321, + "grad_norm": 1.3793448209762573, + "learning_rate": 0.0002979731429397071, + "loss": 3.707489013671875, + "step": 520 + }, + { + "epoch": 0.10086455331412104, + "grad_norm": 1.6294807195663452, + "learning_rate": 0.0002978943159062295, + "loss": 3.7033920288085938, + "step": 525 + }, + { + "epoch": 0.10182516810758886, + "grad_norm": 1.3319975137710571, + "learning_rate": 0.0002978139959616507, + "loss": 3.7044357299804687, + "step": 530 + }, + { + "epoch": 0.10278578290105668, + "grad_norm": 1.1249921321868896, + "learning_rate": 0.00029773218391675594, + "loss": 3.6977405548095703, + "step": 535 + }, + { + "epoch": 0.1037463976945245, + "grad_norm": 1.7312142848968506, + "learning_rate": 0.00029764888059739255, + "loss": 3.698072814941406, + "step": 540 + }, + { + "epoch": 0.10470701248799232, + "grad_norm": 1.3000468015670776, + "learning_rate": 0.00029756408684446136, + "loss": 3.7032306671142576, + "step": 545 + }, + { + "epoch": 0.10566762728146013, + "grad_norm": 1.2404688596725464, + "learning_rate": 0.0002974778035139081, + "loss": 3.693822479248047, + "step": 550 + }, + { + "epoch": 0.10662824207492795, + "grad_norm": 0.9569908380508423, + "learning_rate": 0.00029739003147671536, + "loss": 3.6903202056884767, + "step": 555 + }, + { + "epoch": 0.10758885686839577, + "grad_norm": 1.3814141750335693, + "learning_rate": 0.00029730077161889304, + "loss": 3.6874603271484374, + "step": 560 + }, + { + "epoch": 0.10854947166186359, + "grad_norm": 1.3034733533859253, + "learning_rate": 0.00029721002484147, + "loss": 3.6825042724609376, + "step": 565 + }, + { + "epoch": 0.10951008645533142, + "grad_norm": 1.6188613176345825, + "learning_rate": 0.00029711779206048454, + "loss": 3.6817798614501953, + "step": 570 + }, + { + "epoch": 0.11047070124879924, + "grad_norm": 1.411993384361267, + "learning_rate": 0.0002970240742069755, + "loss": 3.677989959716797, + "step": 575 + }, + { + "epoch": 0.11143131604226705, + "grad_norm": 1.6159188747406006, + "learning_rate": 0.0002969288722269726, + "loss": 3.673622894287109, + "step": 580 + }, + { + "epoch": 0.11239193083573487, + "grad_norm": 1.2981064319610596, + "learning_rate": 0.000296832187081487, + "loss": 3.674951934814453, + "step": 585 + }, + { + "epoch": 0.11335254562920269, + "grad_norm": 1.0490260124206543, + "learning_rate": 0.0002967340197465017, + "loss": 3.665152740478516, + "step": 590 + }, + { + "epoch": 0.1143131604226705, + "grad_norm": 1.5031790733337402, + "learning_rate": 0.00029663437121296146, + "loss": 3.6685272216796876, + "step": 595 + }, + { + "epoch": 0.11527377521613832, + "grad_norm": 1.6380650997161865, + "learning_rate": 0.000296533242486763, + "loss": 3.6649364471435546, + "step": 600 + }, + { + "epoch": 0.11623439000960614, + "grad_norm": 1.1084094047546387, + "learning_rate": 0.0002964306345887447, + "loss": 3.658610153198242, + "step": 605 + }, + { + "epoch": 0.11719500480307397, + "grad_norm": 1.454255223274231, + "learning_rate": 0.0002963265485546764, + "loss": 3.655583953857422, + "step": 610 + }, + { + "epoch": 0.11815561959654179, + "grad_norm": 1.5477410554885864, + "learning_rate": 0.00029622098543524884, + "loss": 3.6575542449951173, + "step": 615 + }, + { + "epoch": 0.11911623439000961, + "grad_norm": 1.0675349235534668, + "learning_rate": 0.00029611394629606324, + "loss": 3.6511856079101563, + "step": 620 + }, + { + "epoch": 0.12007684918347743, + "grad_norm": 1.7654082775115967, + "learning_rate": 0.0002960054322176204, + "loss": 3.6485317230224608, + "step": 625 + }, + { + "epoch": 0.12103746397694524, + "grad_norm": 1.3815727233886719, + "learning_rate": 0.0002958954442953096, + "loss": 3.644425964355469, + "step": 630 + }, + { + "epoch": 0.12199807877041306, + "grad_norm": 1.0989855527877808, + "learning_rate": 0.000295783983639398, + "loss": 3.6440811157226562, + "step": 635 + }, + { + "epoch": 0.12295869356388088, + "grad_norm": 1.3829193115234375, + "learning_rate": 0.00029567105137501916, + "loss": 3.640919876098633, + "step": 640 + }, + { + "epoch": 0.1239193083573487, + "grad_norm": 1.3325438499450684, + "learning_rate": 0.00029555664864216156, + "loss": 3.6379947662353516, + "step": 645 + }, + { + "epoch": 0.12487992315081652, + "grad_norm": 1.112707257270813, + "learning_rate": 0.00029544077659565747, + "loss": 3.636188507080078, + "step": 650 + }, + { + "epoch": 0.12584053794428435, + "grad_norm": 1.1048657894134521, + "learning_rate": 0.0002953234364051708, + "loss": 3.637079620361328, + "step": 655 + }, + { + "epoch": 0.12680115273775217, + "grad_norm": 1.0728777647018433, + "learning_rate": 0.00029520462925518575, + "loss": 3.6303642272949217, + "step": 660 + }, + { + "epoch": 0.12776176753121998, + "grad_norm": 0.9134336113929749, + "learning_rate": 0.00029508435634499467, + "loss": 3.625618743896484, + "step": 665 + }, + { + "epoch": 0.1287223823246878, + "grad_norm": 1.0948387384414673, + "learning_rate": 0.00029496261888868586, + "loss": 3.6273883819580077, + "step": 670 + }, + { + "epoch": 0.12968299711815562, + "grad_norm": 1.1168876886367798, + "learning_rate": 0.0002948394181151314, + "loss": 3.6221595764160157, + "step": 675 + }, + { + "epoch": 0.13064361191162344, + "grad_norm": 1.1925864219665527, + "learning_rate": 0.0002947147552679748, + "loss": 3.6182037353515626, + "step": 680 + }, + { + "epoch": 0.13160422670509125, + "grad_norm": 1.2935006618499756, + "learning_rate": 0.00029458863160561837, + "loss": 3.6180152893066406, + "step": 685 + }, + { + "epoch": 0.13256484149855907, + "grad_norm": 1.0551207065582275, + "learning_rate": 0.0002944610484012105, + "loss": 3.614363098144531, + "step": 690 + }, + { + "epoch": 0.1335254562920269, + "grad_norm": 1.2916655540466309, + "learning_rate": 0.0002943320069426329, + "loss": 3.611812210083008, + "step": 695 + }, + { + "epoch": 0.1344860710854947, + "grad_norm": 1.084269404411316, + "learning_rate": 0.00029420150853248756, + "loss": 3.609844970703125, + "step": 700 + }, + { + "epoch": 0.13544668587896252, + "grad_norm": 1.6168688535690308, + "learning_rate": 0.0002940695544880836, + "loss": 3.6106407165527346, + "step": 705 + }, + { + "epoch": 0.13640730067243034, + "grad_norm": 1.294932246208191, + "learning_rate": 0.0002939361461414238, + "loss": 3.607910919189453, + "step": 710 + }, + { + "epoch": 0.1373679154658982, + "grad_norm": 0.9968737363815308, + "learning_rate": 0.0002938012848391915, + "loss": 3.6028270721435547, + "step": 715 + }, + { + "epoch": 0.138328530259366, + "grad_norm": 1.143962025642395, + "learning_rate": 0.0002936649719427367, + "loss": 3.6049777984619142, + "step": 720 + }, + { + "epoch": 0.13928914505283382, + "grad_norm": 1.4859670400619507, + "learning_rate": 0.00029352720882806267, + "loss": 3.6023681640625, + "step": 725 + }, + { + "epoch": 0.14024975984630164, + "grad_norm": 1.383272647857666, + "learning_rate": 0.00029338799688581146, + "loss": 3.5986564636230467, + "step": 730 + }, + { + "epoch": 0.14121037463976946, + "grad_norm": 1.3845287561416626, + "learning_rate": 0.00029324733752125054, + "loss": 3.597808074951172, + "step": 735 + }, + { + "epoch": 0.14217098943323728, + "grad_norm": 1.5982863903045654, + "learning_rate": 0.0002931052321542581, + "loss": 3.5963520050048827, + "step": 740 + }, + { + "epoch": 0.1431316042267051, + "grad_norm": 1.0727800130844116, + "learning_rate": 0.00029296168221930904, + "loss": 3.592487335205078, + "step": 745 + }, + { + "epoch": 0.1440922190201729, + "grad_norm": 1.4507306814193726, + "learning_rate": 0.0002928166891654604, + "loss": 3.595014190673828, + "step": 750 + }, + { + "epoch": 0.14505283381364073, + "grad_norm": 1.4457939863204956, + "learning_rate": 0.00029267025445633667, + "loss": 3.590290069580078, + "step": 755 + }, + { + "epoch": 0.14601344860710855, + "grad_norm": 1.1515284776687622, + "learning_rate": 0.0002925223795701149, + "loss": 3.5872840881347656, + "step": 760 + }, + { + "epoch": 0.14697406340057637, + "grad_norm": 1.0049299001693726, + "learning_rate": 0.00029237306599951007, + "loss": 3.5840923309326174, + "step": 765 + }, + { + "epoch": 0.14793467819404418, + "grad_norm": 1.628124475479126, + "learning_rate": 0.00029222231525176005, + "loss": 3.5856204986572267, + "step": 770 + }, + { + "epoch": 0.148895292987512, + "grad_norm": 1.5615901947021484, + "learning_rate": 0.0002920701288486099, + "loss": 3.5824256896972657, + "step": 775 + }, + { + "epoch": 0.14985590778097982, + "grad_norm": 1.0828890800476074, + "learning_rate": 0.00029191650832629694, + "loss": 3.5833717346191407, + "step": 780 + }, + { + "epoch": 0.15081652257444764, + "grad_norm": 0.8812053799629211, + "learning_rate": 0.00029176145523553517, + "loss": 3.581065368652344, + "step": 785 + }, + { + "epoch": 0.15177713736791545, + "grad_norm": 1.3092104196548462, + "learning_rate": 0.0002916049711414996, + "loss": 3.5770198822021486, + "step": 790 + }, + { + "epoch": 0.15273775216138327, + "grad_norm": 1.4702849388122559, + "learning_rate": 0.00029144705762381036, + "loss": 3.5817501068115236, + "step": 795 + }, + { + "epoch": 0.15369836695485112, + "grad_norm": 1.110189437866211, + "learning_rate": 0.0002912877162765169, + "loss": 3.5730361938476562, + "step": 800 + }, + { + "epoch": 0.15465898174831894, + "grad_norm": 0.9270662665367126, + "learning_rate": 0.00029112694870808155, + "loss": 3.5725852966308596, + "step": 805 + }, + { + "epoch": 0.15561959654178675, + "grad_norm": 1.3707542419433594, + "learning_rate": 0.00029096475654136395, + "loss": 3.5693943023681642, + "step": 810 + }, + { + "epoch": 0.15658021133525457, + "grad_norm": 1.4934515953063965, + "learning_rate": 0.000290801141413604, + "loss": 3.565943145751953, + "step": 815 + }, + { + "epoch": 0.1575408261287224, + "grad_norm": 1.0972487926483154, + "learning_rate": 0.00029063610497640576, + "loss": 3.5643775939941404, + "step": 820 + }, + { + "epoch": 0.1585014409221902, + "grad_norm": 1.2283886671066284, + "learning_rate": 0.0002904696488957204, + "loss": 3.5635589599609374, + "step": 825 + }, + { + "epoch": 0.15946205571565802, + "grad_norm": 1.164687156677246, + "learning_rate": 0.0002903017748518298, + "loss": 3.555510711669922, + "step": 830 + }, + { + "epoch": 0.16042267050912584, + "grad_norm": 0.9848024845123291, + "learning_rate": 0.0002901324845393294, + "loss": 3.5610126495361327, + "step": 835 + }, + { + "epoch": 0.16138328530259366, + "grad_norm": 1.4761382341384888, + "learning_rate": 0.00028996177966711097, + "loss": 3.5633079528808596, + "step": 840 + }, + { + "epoch": 0.16234390009606148, + "grad_norm": 1.265584945678711, + "learning_rate": 0.0002897896619583455, + "loss": 3.5593791961669923, + "step": 845 + }, + { + "epoch": 0.1633045148895293, + "grad_norm": 1.2358146905899048, + "learning_rate": 0.0002896161331504659, + "loss": 3.5586807250976564, + "step": 850 + }, + { + "epoch": 0.1642651296829971, + "grad_norm": 1.2075212001800537, + "learning_rate": 0.00028944119499514913, + "loss": 3.5556678771972656, + "step": 855 + }, + { + "epoch": 0.16522574447646493, + "grad_norm": 1.3434778451919556, + "learning_rate": 0.0002892648492582989, + "loss": 3.548540496826172, + "step": 860 + }, + { + "epoch": 0.16618635926993275, + "grad_norm": 0.7819045186042786, + "learning_rate": 0.00028908709772002765, + "loss": 3.5495044708251955, + "step": 865 + }, + { + "epoch": 0.16714697406340057, + "grad_norm": 1.3979746103286743, + "learning_rate": 0.00028890794217463863, + "loss": 3.5515655517578124, + "step": 870 + }, + { + "epoch": 0.16810758885686838, + "grad_norm": 1.0144678354263306, + "learning_rate": 0.0002887273844306076, + "loss": 3.547974395751953, + "step": 875 + }, + { + "epoch": 0.1690682036503362, + "grad_norm": 0.8473827242851257, + "learning_rate": 0.00028854542631056494, + "loss": 3.543077087402344, + "step": 880 + }, + { + "epoch": 0.17002881844380405, + "grad_norm": 0.9643390774726868, + "learning_rate": 0.0002883620696512769, + "loss": 3.546183776855469, + "step": 885 + }, + { + "epoch": 0.17098943323727187, + "grad_norm": 1.3916168212890625, + "learning_rate": 0.0002881773163036273, + "loss": 3.5453559875488283, + "step": 890 + }, + { + "epoch": 0.17195004803073968, + "grad_norm": 0.9916486740112305, + "learning_rate": 0.00028799116813259875, + "loss": 3.543218994140625, + "step": 895 + }, + { + "epoch": 0.1729106628242075, + "grad_norm": 0.9322313666343689, + "learning_rate": 0.0002878036270172538, + "loss": 3.5352169036865235, + "step": 900 + }, + { + "epoch": 0.17387127761767532, + "grad_norm": 1.0106613636016846, + "learning_rate": 0.000287614694850716, + "loss": 3.539842224121094, + "step": 905 + }, + { + "epoch": 0.17483189241114314, + "grad_norm": 0.9181408286094666, + "learning_rate": 0.00028742437354015073, + "loss": 3.53677978515625, + "step": 910 + }, + { + "epoch": 0.17579250720461095, + "grad_norm": 1.0180490016937256, + "learning_rate": 0.0002872326650067462, + "loss": 3.536122131347656, + "step": 915 + }, + { + "epoch": 0.17675312199807877, + "grad_norm": 1.140866994857788, + "learning_rate": 0.00028703957118569363, + "loss": 3.5390419006347655, + "step": 920 + }, + { + "epoch": 0.1777137367915466, + "grad_norm": 1.3825160264968872, + "learning_rate": 0.000286845094026168, + "loss": 3.534088897705078, + "step": 925 + }, + { + "epoch": 0.1786743515850144, + "grad_norm": 0.9901686310768127, + "learning_rate": 0.0002866492354913086, + "loss": 3.5334190368652343, + "step": 930 + }, + { + "epoch": 0.17963496637848222, + "grad_norm": 1.3177112340927124, + "learning_rate": 0.0002864519975581986, + "loss": 3.531676483154297, + "step": 935 + }, + { + "epoch": 0.18059558117195004, + "grad_norm": 1.5046802759170532, + "learning_rate": 0.0002862533822178456, + "loss": 3.5300270080566407, + "step": 940 + }, + { + "epoch": 0.18155619596541786, + "grad_norm": 0.8187685012817383, + "learning_rate": 0.00028605339147516113, + "loss": 3.527945709228516, + "step": 945 + }, + { + "epoch": 0.18251681075888568, + "grad_norm": 1.0654352903366089, + "learning_rate": 0.00028585202734894105, + "loss": 3.5284492492675783, + "step": 950 + }, + { + "epoch": 0.1834774255523535, + "grad_norm": 1.0617858171463013, + "learning_rate": 0.00028564929187184447, + "loss": 3.526523208618164, + "step": 955 + }, + { + "epoch": 0.1844380403458213, + "grad_norm": 0.9183857440948486, + "learning_rate": 0.00028544518709037363, + "loss": 3.5269630432128904, + "step": 960 + }, + { + "epoch": 0.18539865513928913, + "grad_norm": 1.3353052139282227, + "learning_rate": 0.000285239715064853, + "loss": 3.5191162109375, + "step": 965 + }, + { + "epoch": 0.18635926993275698, + "grad_norm": 1.1203348636627197, + "learning_rate": 0.0002850328778694088, + "loss": 3.5222145080566407, + "step": 970 + }, + { + "epoch": 0.1873198847262248, + "grad_norm": 1.1545228958129883, + "learning_rate": 0.0002848246775919478, + "loss": 3.518872833251953, + "step": 975 + }, + { + "epoch": 0.1882804995196926, + "grad_norm": 0.9494602084159851, + "learning_rate": 0.0002846151163341364, + "loss": 3.519129180908203, + "step": 980 + }, + { + "epoch": 0.18924111431316043, + "grad_norm": 0.8861366510391235, + "learning_rate": 0.0002844041962113792, + "loss": 3.517262268066406, + "step": 985 + }, + { + "epoch": 0.19020172910662825, + "grad_norm": 1.097005844116211, + "learning_rate": 0.00028419191935279793, + "loss": 3.5180564880371095, + "step": 990 + }, + { + "epoch": 0.19116234390009607, + "grad_norm": 0.9788710474967957, + "learning_rate": 0.00028397828790120965, + "loss": 3.5166107177734376, + "step": 995 + }, + { + "epoch": 0.19212295869356388, + "grad_norm": 1.3244202136993408, + "learning_rate": 0.0002837633040131055, + "loss": 3.5119117736816405, + "step": 1000 + }, + { + "epoch": 0.1930835734870317, + "grad_norm": 1.2634823322296143, + "learning_rate": 0.00028354696985862865, + "loss": 3.5120399475097654, + "step": 1005 + }, + { + "epoch": 0.19404418828049952, + "grad_norm": 1.3096829652786255, + "learning_rate": 0.00028332928762155225, + "loss": 3.5108917236328123, + "step": 1010 + }, + { + "epoch": 0.19500480307396734, + "grad_norm": 1.0718635320663452, + "learning_rate": 0.0002831102594992579, + "loss": 3.5096744537353515, + "step": 1015 + }, + { + "epoch": 0.19596541786743515, + "grad_norm": 1.0316826105117798, + "learning_rate": 0.00028288988770271297, + "loss": 3.5067817687988283, + "step": 1020 + }, + { + "epoch": 0.19692603266090297, + "grad_norm": 1.7893810272216797, + "learning_rate": 0.00028266817445644855, + "loss": 3.5084625244140626, + "step": 1025 + }, + { + "epoch": 0.1978866474543708, + "grad_norm": 1.3311545848846436, + "learning_rate": 0.0002824451219985369, + "loss": 3.5091777801513673, + "step": 1030 + }, + { + "epoch": 0.1988472622478386, + "grad_norm": 1.0418428182601929, + "learning_rate": 0.0002822207325805688, + "loss": 3.5157501220703127, + "step": 1035 + }, + { + "epoch": 0.19980787704130643, + "grad_norm": 1.1201485395431519, + "learning_rate": 0.00028199500846763116, + "loss": 3.5064178466796876, + "step": 1040 + }, + { + "epoch": 0.20076849183477424, + "grad_norm": 1.0694836378097534, + "learning_rate": 0.0002817679519382836, + "loss": 3.507163238525391, + "step": 1045 + }, + { + "epoch": 0.2017291066282421, + "grad_norm": 1.2998982667922974, + "learning_rate": 0.0002815395652845359, + "loss": 3.5081203460693358, + "step": 1050 + }, + { + "epoch": 0.2026897214217099, + "grad_norm": 1.3336325883865356, + "learning_rate": 0.0002813098508118247, + "loss": 3.499872589111328, + "step": 1055 + }, + { + "epoch": 0.20365033621517772, + "grad_norm": 0.7667920589447021, + "learning_rate": 0.0002810788108389901, + "loss": 3.499583435058594, + "step": 1060 + }, + { + "epoch": 0.20461095100864554, + "grad_norm": 1.072025179862976, + "learning_rate": 0.0002808464476982526, + "loss": 3.495106506347656, + "step": 1065 + }, + { + "epoch": 0.20557156580211336, + "grad_norm": 0.964080274105072, + "learning_rate": 0.0002806127637351892, + "loss": 3.4985504150390625, + "step": 1070 + }, + { + "epoch": 0.20653218059558118, + "grad_norm": 1.0775015354156494, + "learning_rate": 0.00028037776130871, + "loss": 3.4976402282714845, + "step": 1075 + }, + { + "epoch": 0.207492795389049, + "grad_norm": 1.0506839752197266, + "learning_rate": 0.00028014144279103406, + "loss": 3.4984210968017577, + "step": 1080 + }, + { + "epoch": 0.2084534101825168, + "grad_norm": 0.8290379643440247, + "learning_rate": 0.0002799038105676658, + "loss": 3.494451141357422, + "step": 1085 + }, + { + "epoch": 0.20941402497598463, + "grad_norm": 1.1757789850234985, + "learning_rate": 0.00027966486703737066, + "loss": 3.4927711486816406, + "step": 1090 + }, + { + "epoch": 0.21037463976945245, + "grad_norm": 1.0375887155532837, + "learning_rate": 0.0002794246146121512, + "loss": 3.4952877044677733, + "step": 1095 + }, + { + "epoch": 0.21133525456292027, + "grad_norm": 0.9264093637466431, + "learning_rate": 0.0002791830557172224, + "loss": 3.49503173828125, + "step": 1100 + }, + { + "epoch": 0.21229586935638808, + "grad_norm": 0.9268721342086792, + "learning_rate": 0.00027894019279098726, + "loss": 3.4857433319091795, + "step": 1105 + }, + { + "epoch": 0.2132564841498559, + "grad_norm": 1.1226038932800293, + "learning_rate": 0.00027869602828501234, + "loss": 3.488873291015625, + "step": 1110 + }, + { + "epoch": 0.21421709894332372, + "grad_norm": 1.1734631061553955, + "learning_rate": 0.00027845056466400297, + "loss": 3.487217330932617, + "step": 1115 + }, + { + "epoch": 0.21517771373679154, + "grad_norm": 0.9604414701461792, + "learning_rate": 0.0002782038044057783, + "loss": 3.487082672119141, + "step": 1120 + }, + { + "epoch": 0.21613832853025935, + "grad_norm": 0.9780157208442688, + "learning_rate": 0.0002779557500012462, + "loss": 3.48626708984375, + "step": 1125 + }, + { + "epoch": 0.21709894332372717, + "grad_norm": 1.1477540731430054, + "learning_rate": 0.0002777064039543784, + "loss": 3.4851287841796874, + "step": 1130 + }, + { + "epoch": 0.21805955811719502, + "grad_norm": 1.1044639348983765, + "learning_rate": 0.00027745576878218496, + "loss": 3.483640670776367, + "step": 1135 + }, + { + "epoch": 0.21902017291066284, + "grad_norm": 0.8308424949645996, + "learning_rate": 0.0002772038470146888, + "loss": 3.4800315856933595, + "step": 1140 + }, + { + "epoch": 0.21998078770413065, + "grad_norm": 1.204115390777588, + "learning_rate": 0.0002769506411949007, + "loss": 3.481386184692383, + "step": 1145 + }, + { + "epoch": 0.22094140249759847, + "grad_norm": 1.0885634422302246, + "learning_rate": 0.00027669615387879284, + "loss": 3.4802738189697267, + "step": 1150 + }, + { + "epoch": 0.2219020172910663, + "grad_norm": 1.1829088926315308, + "learning_rate": 0.0002764403876352736, + "loss": 3.4813682556152346, + "step": 1155 + }, + { + "epoch": 0.2228626320845341, + "grad_norm": 1.4106764793395996, + "learning_rate": 0.0002761833450461613, + "loss": 3.4792640686035154, + "step": 1160 + }, + { + "epoch": 0.22382324687800192, + "grad_norm": 1.074035406112671, + "learning_rate": 0.0002759250287061583, + "loss": 3.478520965576172, + "step": 1165 + }, + { + "epoch": 0.22478386167146974, + "grad_norm": 1.211889624595642, + "learning_rate": 0.00027566544122282496, + "loss": 3.478863525390625, + "step": 1170 + }, + { + "epoch": 0.22574447646493756, + "grad_norm": 0.9881852865219116, + "learning_rate": 0.0002754045852165529, + "loss": 3.4755462646484374, + "step": 1175 + }, + { + "epoch": 0.22670509125840538, + "grad_norm": 1.0194916725158691, + "learning_rate": 0.00027514246332053876, + "loss": 3.476274108886719, + "step": 1180 + }, + { + "epoch": 0.2276657060518732, + "grad_norm": 1.0759530067443848, + "learning_rate": 0.0002748790781807577, + "loss": 3.4741195678710937, + "step": 1185 + }, + { + "epoch": 0.228626320845341, + "grad_norm": 1.181774377822876, + "learning_rate": 0.0002746144324559368, + "loss": 3.4723861694335936, + "step": 1190 + }, + { + "epoch": 0.22958693563880883, + "grad_norm": 0.9129538536071777, + "learning_rate": 0.00027434852881752774, + "loss": 3.472820281982422, + "step": 1195 + }, + { + "epoch": 0.23054755043227665, + "grad_norm": 1.01265287399292, + "learning_rate": 0.0002740813699496804, + "loss": 3.472024917602539, + "step": 1200 + }, + { + "epoch": 0.23150816522574447, + "grad_norm": 1.2006109952926636, + "learning_rate": 0.0002738129585492153, + "loss": 3.4669479370117187, + "step": 1205 + }, + { + "epoch": 0.23246878001921228, + "grad_norm": 0.8117277026176453, + "learning_rate": 0.0002735432973255967, + "loss": 3.46884765625, + "step": 1210 + }, + { + "epoch": 0.2334293948126801, + "grad_norm": 0.9207207560539246, + "learning_rate": 0.0002732723890009051, + "loss": 3.4688953399658202, + "step": 1215 + }, + { + "epoch": 0.23439000960614795, + "grad_norm": 1.1179691553115845, + "learning_rate": 0.00027300023630980985, + "loss": 3.472461700439453, + "step": 1220 + }, + { + "epoch": 0.23535062439961577, + "grad_norm": 0.971973180770874, + "learning_rate": 0.00027272684199954137, + "loss": 3.467519760131836, + "step": 1225 + }, + { + "epoch": 0.23631123919308358, + "grad_norm": 0.8591349720954895, + "learning_rate": 0.0002724522088298637, + "loss": 3.4663330078125, + "step": 1230 + }, + { + "epoch": 0.2372718539865514, + "grad_norm": 1.1340241432189941, + "learning_rate": 0.0002721763395730462, + "loss": 3.466988372802734, + "step": 1235 + }, + { + "epoch": 0.23823246878001922, + "grad_norm": 0.7911289930343628, + "learning_rate": 0.00027189923701383627, + "loss": 3.463814544677734, + "step": 1240 + }, + { + "epoch": 0.23919308357348704, + "grad_norm": 1.2324531078338623, + "learning_rate": 0.0002716209039494304, + "loss": 3.46383056640625, + "step": 1245 + }, + { + "epoch": 0.24015369836695485, + "grad_norm": 1.0037308931350708, + "learning_rate": 0.0002713413431894466, + "loss": 3.4620521545410154, + "step": 1250 + }, + { + "epoch": 0.24111431316042267, + "grad_norm": 1.3981319665908813, + "learning_rate": 0.00027106055755589566, + "loss": 3.4639919281005858, + "step": 1255 + }, + { + "epoch": 0.2420749279538905, + "grad_norm": 1.0288619995117188, + "learning_rate": 0.00027077854988315285, + "loss": 3.4653533935546874, + "step": 1260 + }, + { + "epoch": 0.2430355427473583, + "grad_norm": 1.0437750816345215, + "learning_rate": 0.00027049532301792924, + "loss": 3.4617347717285156, + "step": 1265 + }, + { + "epoch": 0.24399615754082613, + "grad_norm": 1.018276572227478, + "learning_rate": 0.00027021087981924296, + "loss": 3.4626544952392577, + "step": 1270 + }, + { + "epoch": 0.24495677233429394, + "grad_norm": 1.2246990203857422, + "learning_rate": 0.0002699252231583904, + "loss": 3.458060073852539, + "step": 1275 + }, + { + "epoch": 0.24591738712776176, + "grad_norm": 1.440726637840271, + "learning_rate": 0.000269638355918917, + "loss": 3.4551132202148436, + "step": 1280 + }, + { + "epoch": 0.24687800192122958, + "grad_norm": 1.0400745868682861, + "learning_rate": 0.00026935028099658864, + "loss": 3.455486297607422, + "step": 1285 + }, + { + "epoch": 0.2478386167146974, + "grad_norm": 1.271278738975525, + "learning_rate": 0.00026906100129936173, + "loss": 3.4623786926269533, + "step": 1290 + }, + { + "epoch": 0.24879923150816521, + "grad_norm": 1.0657492876052856, + "learning_rate": 0.0002687705197473545, + "loss": 3.455769348144531, + "step": 1295 + }, + { + "epoch": 0.24975984630163303, + "grad_norm": 0.913031816482544, + "learning_rate": 0.00026847883927281715, + "loss": 3.4589412689208983, + "step": 1300 + }, + { + "epoch": 0.2507204610951009, + "grad_norm": 0.8772344589233398, + "learning_rate": 0.00026818596282010223, + "loss": 3.4558467864990234, + "step": 1305 + }, + { + "epoch": 0.2516810758885687, + "grad_norm": 1.1004761457443237, + "learning_rate": 0.00026789189334563507, + "loss": 3.455731964111328, + "step": 1310 + }, + { + "epoch": 0.2526416906820365, + "grad_norm": 1.0385078191757202, + "learning_rate": 0.00026759663381788407, + "loss": 3.4509082794189454, + "step": 1315 + }, + { + "epoch": 0.25360230547550433, + "grad_norm": 0.8879292011260986, + "learning_rate": 0.00026730018721733034, + "loss": 3.447312927246094, + "step": 1320 + }, + { + "epoch": 0.25456292026897215, + "grad_norm": 1.32045316696167, + "learning_rate": 0.0002670025565364379, + "loss": 3.4483909606933594, + "step": 1325 + }, + { + "epoch": 0.25552353506243997, + "grad_norm": 1.0469266176223755, + "learning_rate": 0.0002667037447796234, + "loss": 3.44598388671875, + "step": 1330 + }, + { + "epoch": 0.2564841498559078, + "grad_norm": 0.9312946796417236, + "learning_rate": 0.0002664037549632259, + "loss": 3.449735641479492, + "step": 1335 + }, + { + "epoch": 0.2574447646493756, + "grad_norm": 1.0092568397521973, + "learning_rate": 0.00026610259011547617, + "loss": 3.4483173370361326, + "step": 1340 + }, + { + "epoch": 0.2584053794428434, + "grad_norm": 0.905603289604187, + "learning_rate": 0.0002658002532764663, + "loss": 3.4480667114257812, + "step": 1345 + }, + { + "epoch": 0.25936599423631124, + "grad_norm": 1.152783989906311, + "learning_rate": 0.00026549674749811917, + "loss": 3.4437828063964844, + "step": 1350 + }, + { + "epoch": 0.26032660902977905, + "grad_norm": 0.9953619241714478, + "learning_rate": 0.00026519207584415705, + "loss": 3.444900131225586, + "step": 1355 + }, + { + "epoch": 0.2612872238232469, + "grad_norm": 0.8666988611221313, + "learning_rate": 0.00026488624139007154, + "loss": 3.4405364990234375, + "step": 1360 + }, + { + "epoch": 0.2622478386167147, + "grad_norm": 1.0244816541671753, + "learning_rate": 0.0002645792472230917, + "loss": 3.4408805847167967, + "step": 1365 + }, + { + "epoch": 0.2632084534101825, + "grad_norm": 1.211923599243164, + "learning_rate": 0.0002642710964421535, + "loss": 3.441836929321289, + "step": 1370 + }, + { + "epoch": 0.2641690682036503, + "grad_norm": 0.9250677824020386, + "learning_rate": 0.0002639617921578681, + "loss": 3.4430908203125, + "step": 1375 + }, + { + "epoch": 0.26512968299711814, + "grad_norm": 0.9860612154006958, + "learning_rate": 0.0002636513374924908, + "loss": 3.4425369262695313, + "step": 1380 + }, + { + "epoch": 0.26609029779058596, + "grad_norm": 1.1464399099349976, + "learning_rate": 0.00026333973557988923, + "loss": 3.4433387756347655, + "step": 1385 + }, + { + "epoch": 0.2670509125840538, + "grad_norm": 0.9755972027778625, + "learning_rate": 0.0002630269895655119, + "loss": 3.4360820770263674, + "step": 1390 + }, + { + "epoch": 0.2680115273775216, + "grad_norm": 0.7360960245132446, + "learning_rate": 0.00026271310260635633, + "loss": 3.4424560546875, + "step": 1395 + }, + { + "epoch": 0.2689721421709894, + "grad_norm": 1.3588593006134033, + "learning_rate": 0.0002623980778709374, + "loss": 3.4407089233398436, + "step": 1400 + }, + { + "epoch": 0.26993275696445723, + "grad_norm": 1.060451626777649, + "learning_rate": 0.0002620819185392551, + "loss": 3.4387184143066407, + "step": 1405 + }, + { + "epoch": 0.27089337175792505, + "grad_norm": 1.108783483505249, + "learning_rate": 0.00026176462780276246, + "loss": 3.440538787841797, + "step": 1410 + }, + { + "epoch": 0.27185398655139287, + "grad_norm": 1.033456802368164, + "learning_rate": 0.0002614462088643336, + "loss": 3.438457489013672, + "step": 1415 + }, + { + "epoch": 0.2728146013448607, + "grad_norm": 1.1797153949737549, + "learning_rate": 0.00026112666493823103, + "loss": 3.4355682373046874, + "step": 1420 + }, + { + "epoch": 0.2737752161383285, + "grad_norm": 0.8781239986419678, + "learning_rate": 0.00026080599925007355, + "loss": 3.437305450439453, + "step": 1425 + }, + { + "epoch": 0.2747358309317964, + "grad_norm": 0.812835156917572, + "learning_rate": 0.00026048421503680337, + "loss": 3.4358146667480467, + "step": 1430 + }, + { + "epoch": 0.2756964457252642, + "grad_norm": 1.6010043621063232, + "learning_rate": 0.00026016131554665377, + "loss": 3.4342422485351562, + "step": 1435 + }, + { + "epoch": 0.276657060518732, + "grad_norm": 1.0615707635879517, + "learning_rate": 0.000259837304039116, + "loss": 3.434612274169922, + "step": 1440 + }, + { + "epoch": 0.27761767531219983, + "grad_norm": 0.9948373436927795, + "learning_rate": 0.0002595121837849065, + "loss": 3.434070587158203, + "step": 1445 + }, + { + "epoch": 0.27857829010566765, + "grad_norm": 1.335551142692566, + "learning_rate": 0.000259185958065934, + "loss": 3.4336421966552733, + "step": 1450 + }, + { + "epoch": 0.27953890489913547, + "grad_norm": 1.2016087770462036, + "learning_rate": 0.00025885863017526613, + "loss": 3.432381439208984, + "step": 1455 + }, + { + "epoch": 0.2804995196926033, + "grad_norm": 1.0571645498275757, + "learning_rate": 0.00025853020341709646, + "loss": 3.4373619079589846, + "step": 1460 + }, + { + "epoch": 0.2814601344860711, + "grad_norm": 0.9469745755195618, + "learning_rate": 0.000258200681106711, + "loss": 3.433266448974609, + "step": 1465 + }, + { + "epoch": 0.2824207492795389, + "grad_norm": 1.0642951726913452, + "learning_rate": 0.00025787006657045477, + "loss": 3.4274715423583983, + "step": 1470 + }, + { + "epoch": 0.28338136407300674, + "grad_norm": 1.117781162261963, + "learning_rate": 0.000257538363145698, + "loss": 3.430649185180664, + "step": 1475 + }, + { + "epoch": 0.28434197886647455, + "grad_norm": 1.1161621809005737, + "learning_rate": 0.00025720557418080304, + "loss": 3.43245849609375, + "step": 1480 + }, + { + "epoch": 0.28530259365994237, + "grad_norm": 1.0228129625320435, + "learning_rate": 0.00025687170303508977, + "loss": 3.4293182373046873, + "step": 1485 + }, + { + "epoch": 0.2862632084534102, + "grad_norm": 0.8712406754493713, + "learning_rate": 0.00025653675307880225, + "loss": 3.424097442626953, + "step": 1490 + }, + { + "epoch": 0.287223823246878, + "grad_norm": 0.8777533173561096, + "learning_rate": 0.00025620072769307463, + "loss": 3.424530029296875, + "step": 1495 + }, + { + "epoch": 0.2881844380403458, + "grad_norm": 0.9582040309906006, + "learning_rate": 0.00025586363026989677, + "loss": 3.4286567687988283, + "step": 1500 + }, + { + "epoch": 0.28914505283381364, + "grad_norm": 1.0138081312179565, + "learning_rate": 0.0002555254642120802, + "loss": 3.4281455993652346, + "step": 1505 + }, + { + "epoch": 0.29010566762728146, + "grad_norm": 1.2888671159744263, + "learning_rate": 0.0002551862329332238, + "loss": 3.425878143310547, + "step": 1510 + }, + { + "epoch": 0.2910662824207493, + "grad_norm": 0.9225014448165894, + "learning_rate": 0.0002548459398576791, + "loss": 3.426416015625, + "step": 1515 + }, + { + "epoch": 0.2920268972142171, + "grad_norm": 0.7408014535903931, + "learning_rate": 0.00025450458842051616, + "loss": 3.4226699829101563, + "step": 1520 + }, + { + "epoch": 0.2929875120076849, + "grad_norm": 1.099754810333252, + "learning_rate": 0.0002541621820674882, + "loss": 3.42437858581543, + "step": 1525 + }, + { + "epoch": 0.29394812680115273, + "grad_norm": 1.210573434829712, + "learning_rate": 0.0002538187242549976, + "loss": 3.422506332397461, + "step": 1530 + }, + { + "epoch": 0.29490874159462055, + "grad_norm": 1.1748319864273071, + "learning_rate": 0.00025347421845006056, + "loss": 3.4207489013671877, + "step": 1535 + }, + { + "epoch": 0.29586935638808837, + "grad_norm": 1.2270821332931519, + "learning_rate": 0.00025312866813027195, + "loss": 3.4193565368652346, + "step": 1540 + }, + { + "epoch": 0.2968299711815562, + "grad_norm": 1.1345160007476807, + "learning_rate": 0.0002527820767837708, + "loss": 3.4215885162353517, + "step": 1545 + }, + { + "epoch": 0.297790585975024, + "grad_norm": 1.4022908210754395, + "learning_rate": 0.00025243444790920447, + "loss": 3.4179046630859373, + "step": 1550 + }, + { + "epoch": 0.2987512007684918, + "grad_norm": 0.9985158443450928, + "learning_rate": 0.0002520857850156936, + "loss": 3.4201751708984376, + "step": 1555 + }, + { + "epoch": 0.29971181556195964, + "grad_norm": 1.081823468208313, + "learning_rate": 0.0002517360916227968, + "loss": 3.416116714477539, + "step": 1560 + }, + { + "epoch": 0.30067243035542746, + "grad_norm": 1.2369301319122314, + "learning_rate": 0.000251385371260475, + "loss": 3.4199737548828124, + "step": 1565 + }, + { + "epoch": 0.3016330451488953, + "grad_norm": 0.8743671178817749, + "learning_rate": 0.0002510336274690557, + "loss": 3.4136940002441407, + "step": 1570 + }, + { + "epoch": 0.3025936599423631, + "grad_norm": 1.0848971605300903, + "learning_rate": 0.0002506808637991974, + "loss": 3.417308044433594, + "step": 1575 + }, + { + "epoch": 0.3035542747358309, + "grad_norm": 1.450103521347046, + "learning_rate": 0.0002503270838118537, + "loss": 3.41820068359375, + "step": 1580 + }, + { + "epoch": 0.3045148895292987, + "grad_norm": 0.9271462559700012, + "learning_rate": 0.0002499722910782374, + "loss": 3.4144283294677735, + "step": 1585 + }, + { + "epoch": 0.30547550432276654, + "grad_norm": 0.9813090562820435, + "learning_rate": 0.0002496164891797844, + "loss": 3.41821403503418, + "step": 1590 + }, + { + "epoch": 0.30643611911623436, + "grad_norm": 0.9819661378860474, + "learning_rate": 0.0002492596817081175, + "loss": 3.4166328430175783, + "step": 1595 + }, + { + "epoch": 0.30739673390970224, + "grad_norm": 1.1119061708450317, + "learning_rate": 0.0002489018722650103, + "loss": 3.4129749298095704, + "step": 1600 + }, + { + "epoch": 0.30835734870317005, + "grad_norm": 1.0902990102767944, + "learning_rate": 0.0002485430644623507, + "loss": 3.415445327758789, + "step": 1605 + }, + { + "epoch": 0.30931796349663787, + "grad_norm": 1.1950386762619019, + "learning_rate": 0.00024818326192210447, + "loss": 3.411751937866211, + "step": 1610 + }, + { + "epoch": 0.3102785782901057, + "grad_norm": 0.8721151351928711, + "learning_rate": 0.0002478224682762787, + "loss": 3.409122085571289, + "step": 1615 + }, + { + "epoch": 0.3112391930835735, + "grad_norm": 1.1784799098968506, + "learning_rate": 0.0002474606871668852, + "loss": 3.411494827270508, + "step": 1620 + }, + { + "epoch": 0.3121998078770413, + "grad_norm": 1.107187271118164, + "learning_rate": 0.00024709792224590356, + "loss": 3.4105316162109376, + "step": 1625 + }, + { + "epoch": 0.31316042267050914, + "grad_norm": 0.8363384008407593, + "learning_rate": 0.0002467341771752446, + "loss": 3.406443786621094, + "step": 1630 + }, + { + "epoch": 0.31412103746397696, + "grad_norm": 0.8663597702980042, + "learning_rate": 0.000246369455626713, + "loss": 3.409493637084961, + "step": 1635 + }, + { + "epoch": 0.3150816522574448, + "grad_norm": 0.7546355724334717, + "learning_rate": 0.00024600376128197047, + "loss": 3.403882598876953, + "step": 1640 + }, + { + "epoch": 0.3160422670509126, + "grad_norm": 1.0707752704620361, + "learning_rate": 0.00024563709783249877, + "loss": 3.405009460449219, + "step": 1645 + }, + { + "epoch": 0.3170028818443804, + "grad_norm": 1.0536484718322754, + "learning_rate": 0.00024526946897956194, + "loss": 3.4075836181640624, + "step": 1650 + }, + { + "epoch": 0.31796349663784823, + "grad_norm": 1.0901403427124023, + "learning_rate": 0.00024490087843416947, + "loss": 3.406299591064453, + "step": 1655 + }, + { + "epoch": 0.31892411143131605, + "grad_norm": 1.224731683731079, + "learning_rate": 0.00024453132991703844, + "loss": 3.408635711669922, + "step": 1660 + }, + { + "epoch": 0.31988472622478387, + "grad_norm": 0.9345903396606445, + "learning_rate": 0.00024416082715855627, + "loss": 3.4082077026367186, + "step": 1665 + }, + { + "epoch": 0.3208453410182517, + "grad_norm": 1.0254333019256592, + "learning_rate": 0.00024378937389874276, + "loss": 3.409267807006836, + "step": 1670 + }, + { + "epoch": 0.3218059558117195, + "grad_norm": 1.0035061836242676, + "learning_rate": 0.0002434169738872126, + "loss": 3.4060768127441405, + "step": 1675 + }, + { + "epoch": 0.3227665706051873, + "grad_norm": 0.8024983406066895, + "learning_rate": 0.0002430436308831374, + "loss": 3.4000919342041014, + "step": 1680 + }, + { + "epoch": 0.32372718539865514, + "grad_norm": 0.8286018371582031, + "learning_rate": 0.00024266934865520767, + "loss": 3.405823516845703, + "step": 1685 + }, + { + "epoch": 0.32468780019212296, + "grad_norm": 0.9187076091766357, + "learning_rate": 0.00024229413098159506, + "loss": 3.402732086181641, + "step": 1690 + }, + { + "epoch": 0.3256484149855908, + "grad_norm": 0.9258525967597961, + "learning_rate": 0.00024191798164991378, + "loss": 3.4013343811035157, + "step": 1695 + }, + { + "epoch": 0.3266090297790586, + "grad_norm": 1.202602744102478, + "learning_rate": 0.0002415409044571828, + "loss": 3.4034278869628904, + "step": 1700 + }, + { + "epoch": 0.3275696445725264, + "grad_norm": 1.0954785346984863, + "learning_rate": 0.00024116290320978724, + "loss": 3.4010189056396483, + "step": 1705 + }, + { + "epoch": 0.3285302593659942, + "grad_norm": 0.8299508690834045, + "learning_rate": 0.00024078398172344006, + "loss": 3.402983856201172, + "step": 1710 + }, + { + "epoch": 0.32949087415946204, + "grad_norm": 1.2064135074615479, + "learning_rate": 0.00024040414382314358, + "loss": 3.397635650634766, + "step": 1715 + }, + { + "epoch": 0.33045148895292986, + "grad_norm": 1.0353888273239136, + "learning_rate": 0.00024002339334315066, + "loss": 3.4012256622314454, + "step": 1720 + }, + { + "epoch": 0.3314121037463977, + "grad_norm": 0.9317356944084167, + "learning_rate": 0.00023964173412692631, + "loss": 3.3994606018066404, + "step": 1725 + }, + { + "epoch": 0.3323727185398655, + "grad_norm": 1.0982980728149414, + "learning_rate": 0.00023925917002710865, + "loss": 3.398478698730469, + "step": 1730 + }, + { + "epoch": 0.3333333333333333, + "grad_norm": 0.8598034977912903, + "learning_rate": 0.0002388757049054701, + "loss": 3.400804138183594, + "step": 1735 + }, + { + "epoch": 0.33429394812680113, + "grad_norm": 0.9004125595092773, + "learning_rate": 0.00023849134263287836, + "loss": 3.3936607360839846, + "step": 1740 + }, + { + "epoch": 0.33525456292026895, + "grad_norm": 1.0058287382125854, + "learning_rate": 0.00023810608708925755, + "loss": 3.397069549560547, + "step": 1745 + }, + { + "epoch": 0.33621517771373677, + "grad_norm": 1.0947821140289307, + "learning_rate": 0.00023771994216354857, + "loss": 3.398267364501953, + "step": 1750 + }, + { + "epoch": 0.3371757925072046, + "grad_norm": 1.1534372568130493, + "learning_rate": 0.00023733291175367046, + "loss": 3.3978240966796873, + "step": 1755 + }, + { + "epoch": 0.3381364073006724, + "grad_norm": 0.9231524467468262, + "learning_rate": 0.00023694499976648043, + "loss": 3.4001846313476562, + "step": 1760 + }, + { + "epoch": 0.3390970220941403, + "grad_norm": 1.1671085357666016, + "learning_rate": 0.0002365562101177349, + "loss": 3.3949043273925783, + "step": 1765 + }, + { + "epoch": 0.3400576368876081, + "grad_norm": 1.1204719543457031, + "learning_rate": 0.00023616654673204983, + "loss": 3.3966522216796875, + "step": 1770 + }, + { + "epoch": 0.3410182516810759, + "grad_norm": 0.9162036776542664, + "learning_rate": 0.00023577601354286094, + "loss": 3.3971607208251955, + "step": 1775 + }, + { + "epoch": 0.34197886647454373, + "grad_norm": 0.8886377811431885, + "learning_rate": 0.00023538461449238422, + "loss": 3.3945545196533202, + "step": 1780 + }, + { + "epoch": 0.34293948126801155, + "grad_norm": 0.8314016461372375, + "learning_rate": 0.00023499235353157603, + "loss": 3.397966766357422, + "step": 1785 + }, + { + "epoch": 0.34390009606147937, + "grad_norm": 0.9351712465286255, + "learning_rate": 0.0002345992346200932, + "loss": 3.3912559509277345, + "step": 1790 + }, + { + "epoch": 0.3448607108549472, + "grad_norm": 1.2932080030441284, + "learning_rate": 0.00023420526172625316, + "loss": 3.399237060546875, + "step": 1795 + }, + { + "epoch": 0.345821325648415, + "grad_norm": 1.092822551727295, + "learning_rate": 0.00023381043882699377, + "loss": 3.3947265625, + "step": 1800 + }, + { + "epoch": 0.3467819404418828, + "grad_norm": 1.2571241855621338, + "learning_rate": 0.0002334147699078333, + "loss": 3.3920978546142577, + "step": 1805 + }, + { + "epoch": 0.34774255523535064, + "grad_norm": 0.7598462104797363, + "learning_rate": 0.00023301825896282992, + "loss": 3.3873401641845704, + "step": 1810 + }, + { + "epoch": 0.34870317002881845, + "grad_norm": 0.9965523481369019, + "learning_rate": 0.00023262090999454194, + "loss": 3.390736389160156, + "step": 1815 + }, + { + "epoch": 0.34966378482228627, + "grad_norm": 0.9940029978752136, + "learning_rate": 0.00023222272701398664, + "loss": 3.387689208984375, + "step": 1820 + }, + { + "epoch": 0.3506243996157541, + "grad_norm": 1.051283597946167, + "learning_rate": 0.00023182371404060047, + "loss": 3.3928714752197267, + "step": 1825 + }, + { + "epoch": 0.3515850144092219, + "grad_norm": 0.9655299186706543, + "learning_rate": 0.00023142387510219814, + "loss": 3.3853427886962892, + "step": 1830 + }, + { + "epoch": 0.3525456292026897, + "grad_norm": 0.912288248538971, + "learning_rate": 0.00023102321423493192, + "loss": 3.3889575958251954, + "step": 1835 + }, + { + "epoch": 0.35350624399615754, + "grad_norm": 0.890324592590332, + "learning_rate": 0.00023062173548325112, + "loss": 3.388275146484375, + "step": 1840 + }, + { + "epoch": 0.35446685878962536, + "grad_norm": 0.819926381111145, + "learning_rate": 0.0002302194428998611, + "loss": 3.3917514801025392, + "step": 1845 + }, + { + "epoch": 0.3554274735830932, + "grad_norm": 1.0578113794326782, + "learning_rate": 0.0002298163405456824, + "loss": 3.384230041503906, + "step": 1850 + }, + { + "epoch": 0.356388088376561, + "grad_norm": 1.0350604057312012, + "learning_rate": 0.00022941243248980966, + "loss": 3.388728713989258, + "step": 1855 + }, + { + "epoch": 0.3573487031700288, + "grad_norm": 1.0110005140304565, + "learning_rate": 0.0002290077228094708, + "loss": 3.386165237426758, + "step": 1860 + }, + { + "epoch": 0.35830931796349663, + "grad_norm": 0.7671937346458435, + "learning_rate": 0.00022860221558998554, + "loss": 3.386688995361328, + "step": 1865 + }, + { + "epoch": 0.35926993275696445, + "grad_norm": 1.032853603363037, + "learning_rate": 0.00022819591492472438, + "loss": 3.385240173339844, + "step": 1870 + }, + { + "epoch": 0.36023054755043227, + "grad_norm": 0.8370668888092041, + "learning_rate": 0.00022778882491506725, + "loss": 3.3829574584960938, + "step": 1875 + }, + { + "epoch": 0.3611911623439001, + "grad_norm": 1.0862385034561157, + "learning_rate": 0.00022738094967036208, + "loss": 3.3890167236328126, + "step": 1880 + }, + { + "epoch": 0.3621517771373679, + "grad_norm": 1.2748796939849854, + "learning_rate": 0.00022697229330788312, + "loss": 3.384657287597656, + "step": 1885 + }, + { + "epoch": 0.3631123919308357, + "grad_norm": 1.163509726524353, + "learning_rate": 0.00022656285995278984, + "loss": 3.3868003845214845, + "step": 1890 + }, + { + "epoch": 0.36407300672430354, + "grad_norm": 0.832695484161377, + "learning_rate": 0.00022615265373808488, + "loss": 3.38619384765625, + "step": 1895 + }, + { + "epoch": 0.36503362151777136, + "grad_norm": 1.1879743337631226, + "learning_rate": 0.00022574167880457245, + "loss": 3.384031295776367, + "step": 1900 + }, + { + "epoch": 0.3659942363112392, + "grad_norm": 0.9494090676307678, + "learning_rate": 0.00022532993930081668, + "loss": 3.379372406005859, + "step": 1905 + }, + { + "epoch": 0.366954851104707, + "grad_norm": 0.7896311283111572, + "learning_rate": 0.00022491743938309936, + "loss": 3.3832141876220705, + "step": 1910 + }, + { + "epoch": 0.3679154658981748, + "grad_norm": 0.911125898361206, + "learning_rate": 0.0002245041832153786, + "loss": 3.3824745178222657, + "step": 1915 + }, + { + "epoch": 0.3688760806916426, + "grad_norm": 1.241249680519104, + "learning_rate": 0.0002240901749692461, + "loss": 3.3809112548828124, + "step": 1920 + }, + { + "epoch": 0.36983669548511044, + "grad_norm": 0.8434759974479675, + "learning_rate": 0.00022367541882388554, + "loss": 3.3806549072265626, + "step": 1925 + }, + { + "epoch": 0.37079731027857826, + "grad_norm": 1.0567084550857544, + "learning_rate": 0.00022325991896603018, + "loss": 3.380388641357422, + "step": 1930 + }, + { + "epoch": 0.37175792507204614, + "grad_norm": 1.1255742311477661, + "learning_rate": 0.00022284367958992065, + "loss": 3.3810966491699217, + "step": 1935 + }, + { + "epoch": 0.37271853986551395, + "grad_norm": 0.8697860240936279, + "learning_rate": 0.0002224267048972627, + "loss": 3.381666564941406, + "step": 1940 + }, + { + "epoch": 0.37367915465898177, + "grad_norm": 1.1864638328552246, + "learning_rate": 0.00022200899909718456, + "loss": 3.3785327911376952, + "step": 1945 + }, + { + "epoch": 0.3746397694524496, + "grad_norm": 0.8901607394218445, + "learning_rate": 0.00022159056640619457, + "loss": 3.3746604919433594, + "step": 1950 + }, + { + "epoch": 0.3756003842459174, + "grad_norm": 0.940137505531311, + "learning_rate": 0.00022117141104813876, + "loss": 3.3786911010742187, + "step": 1955 + }, + { + "epoch": 0.3765609990393852, + "grad_norm": 0.9261166453361511, + "learning_rate": 0.000220751537254158, + "loss": 3.377775192260742, + "step": 1960 + }, + { + "epoch": 0.37752161383285304, + "grad_norm": 0.9173313975334167, + "learning_rate": 0.00022033094926264548, + "loss": 3.375761795043945, + "step": 1965 + }, + { + "epoch": 0.37848222862632086, + "grad_norm": 1.0608283281326294, + "learning_rate": 0.00021990965131920358, + "loss": 3.3748458862304687, + "step": 1970 + }, + { + "epoch": 0.3794428434197887, + "grad_norm": 0.9284070134162903, + "learning_rate": 0.0002194876476766015, + "loss": 3.374094772338867, + "step": 1975 + }, + { + "epoch": 0.3804034582132565, + "grad_norm": 0.8284513354301453, + "learning_rate": 0.00021906494259473196, + "loss": 3.3742515563964846, + "step": 1980 + }, + { + "epoch": 0.3813640730067243, + "grad_norm": 1.1894856691360474, + "learning_rate": 0.00021864154034056832, + "loss": 3.3718952178955077, + "step": 1985 + }, + { + "epoch": 0.38232468780019213, + "grad_norm": 0.9039756655693054, + "learning_rate": 0.00021821744518812154, + "loss": 3.3747280120849608, + "step": 1990 + }, + { + "epoch": 0.38328530259365995, + "grad_norm": 0.8699926137924194, + "learning_rate": 0.00021779266141839699, + "loss": 3.372933197021484, + "step": 1995 + }, + { + "epoch": 0.38424591738712777, + "grad_norm": 0.9204120635986328, + "learning_rate": 0.00021736719331935127, + "loss": 3.370801544189453, + "step": 2000 + }, + { + "epoch": 0.3852065321805956, + "grad_norm": 1.0923502445220947, + "learning_rate": 0.00021694104518584886, + "loss": 3.374231719970703, + "step": 2005 + }, + { + "epoch": 0.3861671469740634, + "grad_norm": 0.8937914967536926, + "learning_rate": 0.00021651422131961884, + "loss": 3.374364471435547, + "step": 2010 + }, + { + "epoch": 0.3871277617675312, + "grad_norm": 0.8756441473960876, + "learning_rate": 0.0002160867260292115, + "loss": 3.3731971740722657, + "step": 2015 + }, + { + "epoch": 0.38808837656099904, + "grad_norm": 0.7807921171188354, + "learning_rate": 0.00021565856362995464, + "loss": 3.3712364196777345, + "step": 2020 + }, + { + "epoch": 0.38904899135446686, + "grad_norm": 1.0138747692108154, + "learning_rate": 0.00021522973844391024, + "loss": 3.370909881591797, + "step": 2025 + }, + { + "epoch": 0.3900096061479347, + "grad_norm": 0.9295784831047058, + "learning_rate": 0.00021480025479983077, + "loss": 3.369549560546875, + "step": 2030 + }, + { + "epoch": 0.3909702209414025, + "grad_norm": 0.7283740043640137, + "learning_rate": 0.00021437011703311545, + "loss": 3.3707386016845704, + "step": 2035 + }, + { + "epoch": 0.3919308357348703, + "grad_norm": 0.9124945402145386, + "learning_rate": 0.0002139393294857665, + "loss": 3.3750667572021484, + "step": 2040 + }, + { + "epoch": 0.3928914505283381, + "grad_norm": 1.0313608646392822, + "learning_rate": 0.00021350789650634525, + "loss": 3.369192123413086, + "step": 2045 + }, + { + "epoch": 0.39385206532180594, + "grad_norm": 0.904303252696991, + "learning_rate": 0.00021307582244992838, + "loss": 3.3705421447753907, + "step": 2050 + }, + { + "epoch": 0.39481268011527376, + "grad_norm": 0.9359715580940247, + "learning_rate": 0.0002126431116780639, + "loss": 3.3716297149658203, + "step": 2055 + }, + { + "epoch": 0.3957732949087416, + "grad_norm": 0.9251255989074707, + "learning_rate": 0.00021220976855872712, + "loss": 3.3701812744140627, + "step": 2060 + }, + { + "epoch": 0.3967339097022094, + "grad_norm": 0.8808803558349609, + "learning_rate": 0.00021177579746627643, + "loss": 3.369044876098633, + "step": 2065 + }, + { + "epoch": 0.3976945244956772, + "grad_norm": 0.8854672908782959, + "learning_rate": 0.0002113412027814094, + "loss": 3.3659637451171873, + "step": 2070 + }, + { + "epoch": 0.39865513928914503, + "grad_norm": 0.8999817371368408, + "learning_rate": 0.0002109059888911183, + "loss": 3.365467071533203, + "step": 2075 + }, + { + "epoch": 0.39961575408261285, + "grad_norm": 0.9394131898880005, + "learning_rate": 0.00021047016018864602, + "loss": 3.372660827636719, + "step": 2080 + }, + { + "epoch": 0.40057636887608067, + "grad_norm": 0.7608519792556763, + "learning_rate": 0.00021003372107344167, + "loss": 3.367539978027344, + "step": 2085 + }, + { + "epoch": 0.4015369836695485, + "grad_norm": 1.1321345567703247, + "learning_rate": 0.00020959667595111603, + "loss": 3.3716583251953125, + "step": 2090 + }, + { + "epoch": 0.4024975984630163, + "grad_norm": 0.9030200242996216, + "learning_rate": 0.00020915902923339722, + "loss": 3.3653472900390624, + "step": 2095 + }, + { + "epoch": 0.4034582132564842, + "grad_norm": 0.9246105551719666, + "learning_rate": 0.0002087207853380862, + "loss": 3.3666152954101562, + "step": 2100 + }, + { + "epoch": 0.404418828049952, + "grad_norm": 0.9399805665016174, + "learning_rate": 0.00020828194868901205, + "loss": 3.3659893035888673, + "step": 2105 + }, + { + "epoch": 0.4053794428434198, + "grad_norm": 0.8061392307281494, + "learning_rate": 0.00020784252371598732, + "loss": 3.366693115234375, + "step": 2110 + }, + { + "epoch": 0.40634005763688763, + "grad_norm": 1.1672389507293701, + "learning_rate": 0.00020740251485476345, + "loss": 3.3655033111572266, + "step": 2115 + }, + { + "epoch": 0.40730067243035545, + "grad_norm": 0.7972300052642822, + "learning_rate": 0.00020696192654698592, + "loss": 3.3643836975097656, + "step": 2120 + }, + { + "epoch": 0.40826128722382327, + "grad_norm": 0.6781191825866699, + "learning_rate": 0.00020652076324014927, + "loss": 3.363838958740234, + "step": 2125 + }, + { + "epoch": 0.4092219020172911, + "grad_norm": 1.1469253301620483, + "learning_rate": 0.00020607902938755252, + "loss": 3.3665115356445314, + "step": 2130 + }, + { + "epoch": 0.4101825168107589, + "grad_norm": 0.9588373303413391, + "learning_rate": 0.00020563672944825392, + "loss": 3.3640487670898436, + "step": 2135 + }, + { + "epoch": 0.4111431316042267, + "grad_norm": 0.8890582323074341, + "learning_rate": 0.00020519386788702602, + "loss": 3.361592102050781, + "step": 2140 + }, + { + "epoch": 0.41210374639769454, + "grad_norm": 0.9977627396583557, + "learning_rate": 0.0002047504491743107, + "loss": 3.3624588012695313, + "step": 2145 + }, + { + "epoch": 0.41306436119116235, + "grad_norm": 1.0691829919815063, + "learning_rate": 0.000204306477786174, + "loss": 3.3603302001953126, + "step": 2150 + }, + { + "epoch": 0.4140249759846302, + "grad_norm": 1.0000925064086914, + "learning_rate": 0.00020386195820426082, + "loss": 3.364809036254883, + "step": 2155 + }, + { + "epoch": 0.414985590778098, + "grad_norm": 1.147975206375122, + "learning_rate": 0.00020341689491574984, + "loss": 3.3625713348388673, + "step": 2160 + }, + { + "epoch": 0.4159462055715658, + "grad_norm": 0.9693566560745239, + "learning_rate": 0.00020297129241330817, + "loss": 3.3619319915771486, + "step": 2165 + }, + { + "epoch": 0.4169068203650336, + "grad_norm": 0.8027826547622681, + "learning_rate": 0.00020252515519504592, + "loss": 3.3606258392333985, + "step": 2170 + }, + { + "epoch": 0.41786743515850144, + "grad_norm": 0.7421839833259583, + "learning_rate": 0.0002020784877644709, + "loss": 3.3613494873046874, + "step": 2175 + }, + { + "epoch": 0.41882804995196926, + "grad_norm": 0.7940289378166199, + "learning_rate": 0.00020163129463044308, + "loss": 3.360422897338867, + "step": 2180 + }, + { + "epoch": 0.4197886647454371, + "grad_norm": 1.0238333940505981, + "learning_rate": 0.0002011835803071292, + "loss": 3.3607128143310545, + "step": 2185 + }, + { + "epoch": 0.4207492795389049, + "grad_norm": 1.0974775552749634, + "learning_rate": 0.00020073534931395697, + "loss": 3.3562847137451173, + "step": 2190 + }, + { + "epoch": 0.4217098943323727, + "grad_norm": 0.9029604196548462, + "learning_rate": 0.00020028660617556985, + "loss": 3.3622852325439454, + "step": 2195 + }, + { + "epoch": 0.42267050912584053, + "grad_norm": 1.0699923038482666, + "learning_rate": 0.00019983735542178086, + "loss": 3.359086608886719, + "step": 2200 + }, + { + "epoch": 0.42363112391930835, + "grad_norm": 0.9210760593414307, + "learning_rate": 0.00019938760158752725, + "loss": 3.360548400878906, + "step": 2205 + }, + { + "epoch": 0.42459173871277617, + "grad_norm": 1.0080957412719727, + "learning_rate": 0.00019893734921282448, + "loss": 3.3639129638671874, + "step": 2210 + }, + { + "epoch": 0.425552353506244, + "grad_norm": 0.668106734752655, + "learning_rate": 0.0001984866028427207, + "loss": 3.3550628662109374, + "step": 2215 + }, + { + "epoch": 0.4265129682997118, + "grad_norm": 0.9900498390197754, + "learning_rate": 0.00019803536702725044, + "loss": 3.3611160278320313, + "step": 2220 + }, + { + "epoch": 0.4274735830931796, + "grad_norm": 0.6974493861198425, + "learning_rate": 0.00019758364632138908, + "loss": 3.354916000366211, + "step": 2225 + }, + { + "epoch": 0.42843419788664744, + "grad_norm": 0.8031367063522339, + "learning_rate": 0.0001971314452850066, + "loss": 3.360482406616211, + "step": 2230 + }, + { + "epoch": 0.42939481268011526, + "grad_norm": 0.9039629697799683, + "learning_rate": 0.00019667876848282167, + "loss": 3.3570487976074217, + "step": 2235 + }, + { + "epoch": 0.4303554274735831, + "grad_norm": 0.7340952157974243, + "learning_rate": 0.0001962256204843556, + "loss": 3.3560821533203127, + "step": 2240 + }, + { + "epoch": 0.4313160422670509, + "grad_norm": 1.069008469581604, + "learning_rate": 0.00019577200586388618, + "loss": 3.358184814453125, + "step": 2245 + }, + { + "epoch": 0.4322766570605187, + "grad_norm": 0.8336465358734131, + "learning_rate": 0.00019531792920040133, + "loss": 3.359252166748047, + "step": 2250 + }, + { + "epoch": 0.4332372718539865, + "grad_norm": 1.0504704713821411, + "learning_rate": 0.0001948633950775532, + "loss": 3.3570930480957033, + "step": 2255 + }, + { + "epoch": 0.43419788664745435, + "grad_norm": 0.879917562007904, + "learning_rate": 0.00019440840808361174, + "loss": 3.3541915893554686, + "step": 2260 + }, + { + "epoch": 0.43515850144092216, + "grad_norm": 0.9227856397628784, + "learning_rate": 0.00019395297281141828, + "loss": 3.3502120971679688, + "step": 2265 + }, + { + "epoch": 0.43611911623439004, + "grad_norm": 0.7166210412979126, + "learning_rate": 0.0001934970938583393, + "loss": 3.3565288543701173, + "step": 2270 + }, + { + "epoch": 0.43707973102785785, + "grad_norm": 0.8305450677871704, + "learning_rate": 0.00019304077582622003, + "loss": 3.352977752685547, + "step": 2275 + }, + { + "epoch": 0.43804034582132567, + "grad_norm": 0.9930862188339233, + "learning_rate": 0.00019258402332133798, + "loss": 3.360272979736328, + "step": 2280 + }, + { + "epoch": 0.4390009606147935, + "grad_norm": 1.1689637899398804, + "learning_rate": 0.00019212684095435637, + "loss": 3.3537288665771485, + "step": 2285 + }, + { + "epoch": 0.4399615754082613, + "grad_norm": 0.7753242254257202, + "learning_rate": 0.00019166923334027765, + "loss": 3.356067657470703, + "step": 2290 + }, + { + "epoch": 0.4409221902017291, + "grad_norm": 0.8336865901947021, + "learning_rate": 0.00019121120509839692, + "loss": 3.3564964294433595, + "step": 2295 + }, + { + "epoch": 0.44188280499519694, + "grad_norm": 0.8085048198699951, + "learning_rate": 0.0001907527608522552, + "loss": 3.3494583129882813, + "step": 2300 + }, + { + "epoch": 0.44284341978866476, + "grad_norm": 1.051607370376587, + "learning_rate": 0.000190293905229593, + "loss": 3.351207733154297, + "step": 2305 + }, + { + "epoch": 0.4438040345821326, + "grad_norm": 0.7590753436088562, + "learning_rate": 0.00018983464286230327, + "loss": 3.3514862060546875, + "step": 2310 + }, + { + "epoch": 0.4447646493756004, + "grad_norm": 0.9674370288848877, + "learning_rate": 0.00018937497838638509, + "loss": 3.3499351501464845, + "step": 2315 + }, + { + "epoch": 0.4457252641690682, + "grad_norm": 0.8794429898262024, + "learning_rate": 0.0001889149164418963, + "loss": 3.3523765563964845, + "step": 2320 + }, + { + "epoch": 0.44668587896253603, + "grad_norm": 0.7959157228469849, + "learning_rate": 0.00018845446167290705, + "loss": 3.350267791748047, + "step": 2325 + }, + { + "epoch": 0.44764649375600385, + "grad_norm": 0.625482976436615, + "learning_rate": 0.0001879936187274529, + "loss": 3.348467254638672, + "step": 2330 + }, + { + "epoch": 0.44860710854947167, + "grad_norm": 1.1344565153121948, + "learning_rate": 0.00018753239225748796, + "loss": 3.35140380859375, + "step": 2335 + }, + { + "epoch": 0.4495677233429395, + "grad_norm": 1.0100090503692627, + "learning_rate": 0.0001870707869188375, + "loss": 3.3482070922851563, + "step": 2340 + }, + { + "epoch": 0.4505283381364073, + "grad_norm": 0.9724920988082886, + "learning_rate": 0.00018660880737115146, + "loss": 3.3504077911376955, + "step": 2345 + }, + { + "epoch": 0.4514889529298751, + "grad_norm": 0.8876535296440125, + "learning_rate": 0.0001861464582778572, + "loss": 3.352552032470703, + "step": 2350 + }, + { + "epoch": 0.45244956772334294, + "grad_norm": 1.3819876909255981, + "learning_rate": 0.00018568374430611242, + "loss": 3.3520408630371095, + "step": 2355 + }, + { + "epoch": 0.45341018251681076, + "grad_norm": 1.1260014772415161, + "learning_rate": 0.00018522067012675798, + "loss": 3.353636932373047, + "step": 2360 + }, + { + "epoch": 0.4543707973102786, + "grad_norm": 0.8487617373466492, + "learning_rate": 0.00018475724041427106, + "loss": 3.3487457275390624, + "step": 2365 + }, + { + "epoch": 0.4553314121037464, + "grad_norm": 1.024625539779663, + "learning_rate": 0.00018429345984671743, + "loss": 3.3499549865722655, + "step": 2370 + }, + { + "epoch": 0.4562920268972142, + "grad_norm": 0.6985915303230286, + "learning_rate": 0.0001838293331057048, + "loss": 3.349627685546875, + "step": 2375 + }, + { + "epoch": 0.457252641690682, + "grad_norm": 0.6404640674591064, + "learning_rate": 0.00018336486487633528, + "loss": 3.347215270996094, + "step": 2380 + }, + { + "epoch": 0.45821325648414984, + "grad_norm": 1.0900517702102661, + "learning_rate": 0.000182900059847158, + "loss": 3.346749114990234, + "step": 2385 + }, + { + "epoch": 0.45917387127761766, + "grad_norm": 0.9426947832107544, + "learning_rate": 0.00018243492271012202, + "loss": 3.3453174591064454, + "step": 2390 + }, + { + "epoch": 0.4601344860710855, + "grad_norm": 0.8978161215782166, + "learning_rate": 0.00018196945816052867, + "loss": 3.3470741271972657, + "step": 2395 + }, + { + "epoch": 0.4610951008645533, + "grad_norm": 1.0698800086975098, + "learning_rate": 0.00018150367089698452, + "loss": 3.3448417663574217, + "step": 2400 + }, + { + "epoch": 0.4620557156580211, + "grad_norm": 0.9989749193191528, + "learning_rate": 0.00018103756562135373, + "loss": 3.3498584747314455, + "step": 2405 + }, + { + "epoch": 0.46301633045148893, + "grad_norm": 1.1512004137039185, + "learning_rate": 0.0001805711470387105, + "loss": 3.344708251953125, + "step": 2410 + }, + { + "epoch": 0.46397694524495675, + "grad_norm": 0.7708789706230164, + "learning_rate": 0.00018010441985729183, + "loss": 3.341356658935547, + "step": 2415 + }, + { + "epoch": 0.46493756003842457, + "grad_norm": 0.8291172385215759, + "learning_rate": 0.00017963738878844966, + "loss": 3.3480514526367187, + "step": 2420 + }, + { + "epoch": 0.4658981748318924, + "grad_norm": 0.8072155714035034, + "learning_rate": 0.00017917005854660374, + "loss": 3.3407440185546875, + "step": 2425 + }, + { + "epoch": 0.4668587896253602, + "grad_norm": 1.1140457391738892, + "learning_rate": 0.00017870243384919364, + "loss": 3.3480880737304686, + "step": 2430 + }, + { + "epoch": 0.4678194044188281, + "grad_norm": 1.1302070617675781, + "learning_rate": 0.0001782345194166314, + "loss": 3.3469482421875, + "step": 2435 + }, + { + "epoch": 0.4687800192122959, + "grad_norm": 0.835529625415802, + "learning_rate": 0.00017776631997225365, + "loss": 3.341139221191406, + "step": 2440 + }, + { + "epoch": 0.4697406340057637, + "grad_norm": 0.8538889288902283, + "learning_rate": 0.0001772978402422742, + "loss": 3.342560958862305, + "step": 2445 + }, + { + "epoch": 0.47070124879923153, + "grad_norm": 0.904460608959198, + "learning_rate": 0.0001768290849557361, + "loss": 3.3424705505371093, + "step": 2450 + }, + { + "epoch": 0.47166186359269935, + "grad_norm": 0.8424951434135437, + "learning_rate": 0.00017636005884446397, + "loss": 3.3413028717041016, + "step": 2455 + }, + { + "epoch": 0.47262247838616717, + "grad_norm": 0.924486517906189, + "learning_rate": 0.00017589076664301637, + "loss": 3.3444580078125, + "step": 2460 + }, + { + "epoch": 0.473583093179635, + "grad_norm": 0.9826391935348511, + "learning_rate": 0.00017542121308863776, + "loss": 3.3431236267089846, + "step": 2465 + }, + { + "epoch": 0.4745437079731028, + "grad_norm": 1.2090203762054443, + "learning_rate": 0.00017495140292121084, + "loss": 3.344011688232422, + "step": 2470 + }, + { + "epoch": 0.4755043227665706, + "grad_norm": 0.7930108904838562, + "learning_rate": 0.0001744813408832088, + "loss": 3.3414871215820314, + "step": 2475 + }, + { + "epoch": 0.47646493756003844, + "grad_norm": 0.7430881857872009, + "learning_rate": 0.0001740110317196472, + "loss": 3.3419719696044923, + "step": 2480 + }, + { + "epoch": 0.47742555235350626, + "grad_norm": 1.0060484409332275, + "learning_rate": 0.0001735404801780362, + "loss": 3.3399085998535156, + "step": 2485 + }, + { + "epoch": 0.4783861671469741, + "grad_norm": 1.0594747066497803, + "learning_rate": 0.0001730696910083326, + "loss": 3.3429637908935548, + "step": 2490 + }, + { + "epoch": 0.4793467819404419, + "grad_norm": 0.9936187863349915, + "learning_rate": 0.0001725986689628921, + "loss": 3.3392757415771483, + "step": 2495 + }, + { + "epoch": 0.4803073967339097, + "grad_norm": 0.6973989009857178, + "learning_rate": 0.00017212741879642096, + "loss": 3.3416038513183595, + "step": 2500 + }, + { + "epoch": 0.4812680115273775, + "grad_norm": 0.5994371771812439, + "learning_rate": 0.00017165594526592833, + "loss": 3.3429832458496094, + "step": 2505 + }, + { + "epoch": 0.48222862632084534, + "grad_norm": 0.8716941475868225, + "learning_rate": 0.000171184253130678, + "loss": 3.339080810546875, + "step": 2510 + }, + { + "epoch": 0.48318924111431316, + "grad_norm": 0.7032945156097412, + "learning_rate": 0.00017071234715214045, + "loss": 3.340536880493164, + "step": 2515 + }, + { + "epoch": 0.484149855907781, + "grad_norm": 0.7946033477783203, + "learning_rate": 0.0001702402320939449, + "loss": 3.3366943359375, + "step": 2520 + }, + { + "epoch": 0.4851104707012488, + "grad_norm": 0.8639918565750122, + "learning_rate": 0.00016976791272183098, + "loss": 3.3365756988525392, + "step": 2525 + }, + { + "epoch": 0.4860710854947166, + "grad_norm": 0.7089393734931946, + "learning_rate": 0.0001692953938036008, + "loss": 3.3373767852783205, + "step": 2530 + }, + { + "epoch": 0.48703170028818443, + "grad_norm": 0.745219349861145, + "learning_rate": 0.00016882268010907087, + "loss": 3.339314270019531, + "step": 2535 + }, + { + "epoch": 0.48799231508165225, + "grad_norm": 0.733579695224762, + "learning_rate": 0.00016834977641002377, + "loss": 3.3369155883789063, + "step": 2540 + }, + { + "epoch": 0.48895292987512007, + "grad_norm": 0.808840274810791, + "learning_rate": 0.00016787668748016008, + "loss": 3.336173248291016, + "step": 2545 + }, + { + "epoch": 0.4899135446685879, + "grad_norm": 0.885305643081665, + "learning_rate": 0.00016740341809505017, + "loss": 3.34091796875, + "step": 2550 + }, + { + "epoch": 0.4908741594620557, + "grad_norm": 0.8833165764808655, + "learning_rate": 0.000166929973032086, + "loss": 3.339965057373047, + "step": 2555 + }, + { + "epoch": 0.4918347742555235, + "grad_norm": 0.948832094669342, + "learning_rate": 0.0001664563570704329, + "loss": 3.341967010498047, + "step": 2560 + }, + { + "epoch": 0.49279538904899134, + "grad_norm": 0.7376810908317566, + "learning_rate": 0.0001659825749909814, + "loss": 3.339089584350586, + "step": 2565 + }, + { + "epoch": 0.49375600384245916, + "grad_norm": 0.6202611327171326, + "learning_rate": 0.00016550863157629888, + "loss": 3.337108612060547, + "step": 2570 + }, + { + "epoch": 0.494716618635927, + "grad_norm": 1.3193303346633911, + "learning_rate": 0.00016503453161058123, + "loss": 3.337067413330078, + "step": 2575 + }, + { + "epoch": 0.4956772334293948, + "grad_norm": 0.9450001120567322, + "learning_rate": 0.00016456027987960466, + "loss": 3.3382850646972657, + "step": 2580 + }, + { + "epoch": 0.4966378482228626, + "grad_norm": 0.9356205463409424, + "learning_rate": 0.00016408588117067743, + "loss": 3.334228515625, + "step": 2585 + }, + { + "epoch": 0.49759846301633043, + "grad_norm": 0.9595420360565186, + "learning_rate": 0.00016361134027259136, + "loss": 3.3412109375, + "step": 2590 + }, + { + "epoch": 0.49855907780979825, + "grad_norm": 0.9596777558326721, + "learning_rate": 0.00016313666197557373, + "loss": 3.337894058227539, + "step": 2595 + }, + { + "epoch": 0.49951969260326606, + "grad_norm": 0.6967231631278992, + "learning_rate": 0.00016266185107123864, + "loss": 3.337984085083008, + "step": 2600 + }, + { + "epoch": 0.5004803073967339, + "grad_norm": 0.638558030128479, + "learning_rate": 0.00016218691235253893, + "loss": 3.3327606201171873, + "step": 2605 + }, + { + "epoch": 0.5014409221902018, + "grad_norm": 1.0894724130630493, + "learning_rate": 0.0001617118506137175, + "loss": 3.3385719299316405, + "step": 2610 + }, + { + "epoch": 0.5024015369836695, + "grad_norm": 0.9641280174255371, + "learning_rate": 0.00016123667065025914, + "loss": 3.3361286163330077, + "step": 2615 + }, + { + "epoch": 0.5033621517771374, + "grad_norm": 0.8381847143173218, + "learning_rate": 0.00016076137725884218, + "loss": 3.334253692626953, + "step": 2620 + }, + { + "epoch": 0.5043227665706052, + "grad_norm": 0.8258698582649231, + "learning_rate": 0.0001602859752372897, + "loss": 3.3326160430908205, + "step": 2625 + }, + { + "epoch": 0.505283381364073, + "grad_norm": 0.6313220262527466, + "learning_rate": 0.00015981046938452146, + "loss": 3.3321769714355467, + "step": 2630 + }, + { + "epoch": 0.5062439961575408, + "grad_norm": 0.6548064947128296, + "learning_rate": 0.0001593348645005054, + "loss": 3.3319812774658204, + "step": 2635 + }, + { + "epoch": 0.5072046109510087, + "grad_norm": 0.9846540093421936, + "learning_rate": 0.00015885916538620906, + "loss": 3.3337692260742187, + "step": 2640 + }, + { + "epoch": 0.5081652257444764, + "grad_norm": 0.9483662843704224, + "learning_rate": 0.00015838337684355121, + "loss": 3.335358810424805, + "step": 2645 + }, + { + "epoch": 0.5091258405379443, + "grad_norm": 0.7134802341461182, + "learning_rate": 0.0001579075036753533, + "loss": 3.333547592163086, + "step": 2650 + }, + { + "epoch": 0.5100864553314121, + "grad_norm": 0.6113592386245728, + "learning_rate": 0.00015743155068529102, + "loss": 3.3339920043945312, + "step": 2655 + }, + { + "epoch": 0.5110470701248799, + "grad_norm": 0.7905352115631104, + "learning_rate": 0.0001569555226778459, + "loss": 3.3328788757324217, + "step": 2660 + }, + { + "epoch": 0.5120076849183477, + "grad_norm": 0.7155654430389404, + "learning_rate": 0.0001564794244582567, + "loss": 3.3339805603027344, + "step": 2665 + }, + { + "epoch": 0.5129682997118156, + "grad_norm": 0.6985665559768677, + "learning_rate": 0.0001560032608324709, + "loss": 3.332271194458008, + "step": 2670 + }, + { + "epoch": 0.5139289145052833, + "grad_norm": 0.9106906056404114, + "learning_rate": 0.00015552703660709618, + "loss": 3.3307167053222657, + "step": 2675 + }, + { + "epoch": 0.5148895292987512, + "grad_norm": 1.00930917263031, + "learning_rate": 0.00015505075658935207, + "loss": 3.3333648681640624, + "step": 2680 + }, + { + "epoch": 0.515850144092219, + "grad_norm": 0.740381121635437, + "learning_rate": 0.0001545744255870212, + "loss": 3.335179901123047, + "step": 2685 + }, + { + "epoch": 0.5168107588856868, + "grad_norm": 0.5899009704589844, + "learning_rate": 0.00015409804840840088, + "loss": 3.3299629211425783, + "step": 2690 + }, + { + "epoch": 0.5177713736791547, + "grad_norm": 1.0662180185317993, + "learning_rate": 0.0001536216298622545, + "loss": 3.332952880859375, + "step": 2695 + }, + { + "epoch": 0.5187319884726225, + "grad_norm": 0.8961836099624634, + "learning_rate": 0.00015314517475776318, + "loss": 3.331147003173828, + "step": 2700 + }, + { + "epoch": 0.5196926032660903, + "grad_norm": 0.8946348428726196, + "learning_rate": 0.00015266868790447685, + "loss": 3.3304420471191407, + "step": 2705 + }, + { + "epoch": 0.5206532180595581, + "grad_norm": 1.0596193075180054, + "learning_rate": 0.0001521921741122661, + "loss": 3.3243175506591798, + "step": 2710 + }, + { + "epoch": 0.521613832853026, + "grad_norm": 1.1533540487289429, + "learning_rate": 0.00015171563819127342, + "loss": 3.329667663574219, + "step": 2715 + }, + { + "epoch": 0.5225744476464937, + "grad_norm": 0.8710746765136719, + "learning_rate": 0.00015123908495186464, + "loss": 3.3310035705566405, + "step": 2720 + }, + { + "epoch": 0.5235350624399616, + "grad_norm": 0.7002533078193665, + "learning_rate": 0.0001507625192045804, + "loss": 3.3324230194091795, + "step": 2725 + }, + { + "epoch": 0.5244956772334294, + "grad_norm": 1.0313575267791748, + "learning_rate": 0.00015028594576008773, + "loss": 3.3279163360595705, + "step": 2730 + }, + { + "epoch": 0.5254562920268973, + "grad_norm": 0.832625687122345, + "learning_rate": 0.00014980936942913113, + "loss": 3.331475830078125, + "step": 2735 + }, + { + "epoch": 0.526416906820365, + "grad_norm": 1.0033438205718994, + "learning_rate": 0.00014933279502248444, + "loss": 3.326156234741211, + "step": 2740 + }, + { + "epoch": 0.5273775216138329, + "grad_norm": 0.5835141539573669, + "learning_rate": 0.000148856227350902, + "loss": 3.325116729736328, + "step": 2745 + }, + { + "epoch": 0.5283381364073007, + "grad_norm": 0.5704639554023743, + "learning_rate": 0.00014837967122507015, + "loss": 3.3299407958984375, + "step": 2750 + }, + { + "epoch": 0.5292987512007685, + "grad_norm": 1.1036860942840576, + "learning_rate": 0.00014790313145555863, + "loss": 3.327469253540039, + "step": 2755 + }, + { + "epoch": 0.5302593659942363, + "grad_norm": 0.8866048455238342, + "learning_rate": 0.00014742661285277228, + "loss": 3.3271484375, + "step": 2760 + }, + { + "epoch": 0.5312199807877042, + "grad_norm": 0.8478221297264099, + "learning_rate": 0.00014695012022690205, + "loss": 3.3288875579833985, + "step": 2765 + }, + { + "epoch": 0.5321805955811719, + "grad_norm": 0.6032402515411377, + "learning_rate": 0.00014647365838787673, + "loss": 3.3231765747070314, + "step": 2770 + }, + { + "epoch": 0.5331412103746398, + "grad_norm": 1.1227121353149414, + "learning_rate": 0.00014599723214531434, + "loss": 3.326404571533203, + "step": 2775 + }, + { + "epoch": 0.5341018251681076, + "grad_norm": 0.9632681012153625, + "learning_rate": 0.0001455208463084737, + "loss": 3.32501106262207, + "step": 2780 + }, + { + "epoch": 0.5350624399615754, + "grad_norm": 0.6381242871284485, + "learning_rate": 0.00014504450568620557, + "loss": 3.328451919555664, + "step": 2785 + }, + { + "epoch": 0.5360230547550432, + "grad_norm": 0.818413496017456, + "learning_rate": 0.00014456821508690432, + "loss": 3.325485610961914, + "step": 2790 + }, + { + "epoch": 0.5369836695485111, + "grad_norm": 0.9857754707336426, + "learning_rate": 0.0001440919793184595, + "loss": 3.323963928222656, + "step": 2795 + }, + { + "epoch": 0.5379442843419788, + "grad_norm": 0.7225484848022461, + "learning_rate": 0.00014361580318820696, + "loss": 3.3251392364501955, + "step": 2800 + }, + { + "epoch": 0.5389048991354467, + "grad_norm": 0.8891339898109436, + "learning_rate": 0.00014313969150288083, + "loss": 3.3246475219726563, + "step": 2805 + }, + { + "epoch": 0.5398655139289145, + "grad_norm": 1.0610110759735107, + "learning_rate": 0.00014266364906856442, + "loss": 3.3243381500244142, + "step": 2810 + }, + { + "epoch": 0.5408261287223823, + "grad_norm": 0.7058207392692566, + "learning_rate": 0.00014218768069064214, + "loss": 3.32562141418457, + "step": 2815 + }, + { + "epoch": 0.5417867435158501, + "grad_norm": 1.1421717405319214, + "learning_rate": 0.00014171179117375082, + "loss": 3.3252777099609374, + "step": 2820 + }, + { + "epoch": 0.542747358309318, + "grad_norm": 0.7852960228919983, + "learning_rate": 0.00014123598532173113, + "loss": 3.3269309997558594, + "step": 2825 + }, + { + "epoch": 0.5437079731027857, + "grad_norm": 0.9386373162269592, + "learning_rate": 0.00014076026793757943, + "loss": 3.325482940673828, + "step": 2830 + }, + { + "epoch": 0.5446685878962536, + "grad_norm": 0.9038012027740479, + "learning_rate": 0.00014028464382339877, + "loss": 3.3256744384765624, + "step": 2835 + }, + { + "epoch": 0.5456292026897214, + "grad_norm": 0.6614556312561035, + "learning_rate": 0.0001398091177803509, + "loss": 3.3238311767578126, + "step": 2840 + }, + { + "epoch": 0.5465898174831892, + "grad_norm": 0.7346850037574768, + "learning_rate": 0.00013933369460860748, + "loss": 3.3215705871582033, + "step": 2845 + }, + { + "epoch": 0.547550432276657, + "grad_norm": 0.9192922115325928, + "learning_rate": 0.00013885837910730168, + "loss": 3.3206497192382813, + "step": 2850 + }, + { + "epoch": 0.5485110470701249, + "grad_norm": 0.8936217427253723, + "learning_rate": 0.00013838317607448004, + "loss": 3.325290298461914, + "step": 2855 + }, + { + "epoch": 0.5494716618635928, + "grad_norm": 0.9450289607048035, + "learning_rate": 0.00013790809030705354, + "loss": 3.324379730224609, + "step": 2860 + }, + { + "epoch": 0.5504322766570605, + "grad_norm": 0.868273913860321, + "learning_rate": 0.00013743312660074962, + "loss": 3.3241962432861327, + "step": 2865 + }, + { + "epoch": 0.5513928914505284, + "grad_norm": 0.6488667130470276, + "learning_rate": 0.00013695828975006336, + "loss": 3.3240901947021486, + "step": 2870 + }, + { + "epoch": 0.5523535062439962, + "grad_norm": 0.4751059114933014, + "learning_rate": 0.00013648358454820957, + "loss": 3.3235042572021483, + "step": 2875 + }, + { + "epoch": 0.553314121037464, + "grad_norm": 1.1512478590011597, + "learning_rate": 0.00013600901578707402, + "loss": 3.3230491638183595, + "step": 2880 + }, + { + "epoch": 0.5542747358309318, + "grad_norm": 0.8698762059211731, + "learning_rate": 0.00013553458825716502, + "loss": 3.324365997314453, + "step": 2885 + }, + { + "epoch": 0.5552353506243997, + "grad_norm": 0.8694410920143127, + "learning_rate": 0.0001350603067475655, + "loss": 3.324441146850586, + "step": 2890 + }, + { + "epoch": 0.5561959654178674, + "grad_norm": 0.8007134199142456, + "learning_rate": 0.00013458617604588418, + "loss": 3.3229400634765627, + "step": 2895 + }, + { + "epoch": 0.5571565802113353, + "grad_norm": 0.765385091304779, + "learning_rate": 0.00013411220093820773, + "loss": 3.323167419433594, + "step": 2900 + }, + { + "epoch": 0.5581171950048031, + "grad_norm": 0.4617004096508026, + "learning_rate": 0.00013363838620905188, + "loss": 3.320812225341797, + "step": 2905 + }, + { + "epoch": 0.5590778097982709, + "grad_norm": 0.7083408832550049, + "learning_rate": 0.00013316473664131347, + "loss": 3.324010467529297, + "step": 2910 + }, + { + "epoch": 0.5600384245917387, + "grad_norm": 0.8876819014549255, + "learning_rate": 0.00013269125701622243, + "loss": 3.3241298675537108, + "step": 2915 + }, + { + "epoch": 0.5609990393852066, + "grad_norm": 0.806339681148529, + "learning_rate": 0.00013221795211329281, + "loss": 3.319907379150391, + "step": 2920 + }, + { + "epoch": 0.5619596541786743, + "grad_norm": 0.6743384599685669, + "learning_rate": 0.00013174482671027526, + "loss": 3.319793701171875, + "step": 2925 + }, + { + "epoch": 0.5629202689721422, + "grad_norm": 0.6159988641738892, + "learning_rate": 0.00013127188558310823, + "loss": 3.323670196533203, + "step": 2930 + }, + { + "epoch": 0.56388088376561, + "grad_norm": 0.9810566902160645, + "learning_rate": 0.0001307991335058702, + "loss": 3.3209304809570312, + "step": 2935 + }, + { + "epoch": 0.5648414985590778, + "grad_norm": 0.913245677947998, + "learning_rate": 0.00013032657525073122, + "loss": 3.320903778076172, + "step": 2940 + }, + { + "epoch": 0.5658021133525456, + "grad_norm": 0.6053956151008606, + "learning_rate": 0.00012985421558790473, + "loss": 3.3186370849609377, + "step": 2945 + }, + { + "epoch": 0.5667627281460135, + "grad_norm": 0.6618540287017822, + "learning_rate": 0.00012938205928559964, + "loss": 3.3157257080078124, + "step": 2950 + }, + { + "epoch": 0.5677233429394812, + "grad_norm": 0.5338152647018433, + "learning_rate": 0.00012891011110997187, + "loss": 3.3171661376953123, + "step": 2955 + }, + { + "epoch": 0.5686839577329491, + "grad_norm": 0.7981716394424438, + "learning_rate": 0.0001284383758250767, + "loss": 3.3200511932373047, + "step": 2960 + }, + { + "epoch": 0.5696445725264169, + "grad_norm": 0.8445151448249817, + "learning_rate": 0.00012796685819282009, + "loss": 3.319135284423828, + "step": 2965 + }, + { + "epoch": 0.5706051873198847, + "grad_norm": 0.6141214370727539, + "learning_rate": 0.0001274955629729111, + "loss": 3.3177207946777343, + "step": 2970 + }, + { + "epoch": 0.5715658021133525, + "grad_norm": 1.2424954175949097, + "learning_rate": 0.0001270244949228136, + "loss": 3.322146987915039, + "step": 2975 + }, + { + "epoch": 0.5725264169068204, + "grad_norm": 0.7265079021453857, + "learning_rate": 0.00012655365879769826, + "loss": 3.3195030212402346, + "step": 2980 + }, + { + "epoch": 0.5734870317002881, + "grad_norm": 0.5697556734085083, + "learning_rate": 0.00012608305935039475, + "loss": 3.319179153442383, + "step": 2985 + }, + { + "epoch": 0.574447646493756, + "grad_norm": 0.6604142189025879, + "learning_rate": 0.00012561270133134344, + "loss": 3.317556381225586, + "step": 2990 + }, + { + "epoch": 0.5754082612872238, + "grad_norm": 0.6255987882614136, + "learning_rate": 0.00012514258948854773, + "loss": 3.3171707153320313, + "step": 2995 + }, + { + "epoch": 0.5763688760806917, + "grad_norm": 0.6028192639350891, + "learning_rate": 0.00012467272856752593, + "loss": 3.317168426513672, + "step": 3000 + }, + { + "epoch": 0.5773294908741594, + "grad_norm": 0.8794112801551819, + "learning_rate": 0.0001242031233112634, + "loss": 3.3190616607666015, + "step": 3005 + }, + { + "epoch": 0.5782901056676273, + "grad_norm": 0.6320680975914001, + "learning_rate": 0.00012373377846016493, + "loss": 3.315277099609375, + "step": 3010 + }, + { + "epoch": 0.579250720461095, + "grad_norm": 0.5714786052703857, + "learning_rate": 0.0001232646987520064, + "loss": 3.3192188262939455, + "step": 3015 + }, + { + "epoch": 0.5802113352545629, + "grad_norm": 0.7069401144981384, + "learning_rate": 0.00012279588892188736, + "loss": 3.317308807373047, + "step": 3020 + }, + { + "epoch": 0.5811719500480308, + "grad_norm": 0.7392922639846802, + "learning_rate": 0.000122327353702183, + "loss": 3.317064666748047, + "step": 3025 + }, + { + "epoch": 0.5821325648414986, + "grad_norm": 1.0391535758972168, + "learning_rate": 0.0001218590978224966, + "loss": 3.313740539550781, + "step": 3030 + }, + { + "epoch": 0.5830931796349664, + "grad_norm": 0.7290921211242676, + "learning_rate": 0.00012139112600961159, + "loss": 3.317361831665039, + "step": 3035 + }, + { + "epoch": 0.5840537944284342, + "grad_norm": 0.5767372250556946, + "learning_rate": 0.00012092344298744383, + "loss": 3.3169986724853517, + "step": 3040 + }, + { + "epoch": 0.5850144092219021, + "grad_norm": 1.0307767391204834, + "learning_rate": 0.00012045605347699411, + "loss": 3.3194618225097656, + "step": 3045 + }, + { + "epoch": 0.5859750240153698, + "grad_norm": 0.622552752494812, + "learning_rate": 0.00011998896219630029, + "loss": 3.316600799560547, + "step": 3050 + }, + { + "epoch": 0.5869356388088377, + "grad_norm": 0.5956411957740784, + "learning_rate": 0.0001195221738603899, + "loss": 3.312654495239258, + "step": 3055 + }, + { + "epoch": 0.5878962536023055, + "grad_norm": 0.7186912298202515, + "learning_rate": 0.00011905569318123223, + "loss": 3.316633605957031, + "step": 3060 + }, + { + "epoch": 0.5888568683957733, + "grad_norm": 0.6377474069595337, + "learning_rate": 0.00011858952486769114, + "loss": 3.315812683105469, + "step": 3065 + }, + { + "epoch": 0.5898174831892411, + "grad_norm": 0.6208492517471313, + "learning_rate": 0.00011812367362547716, + "loss": 3.3148754119873045, + "step": 3070 + }, + { + "epoch": 0.590778097982709, + "grad_norm": 0.600408673286438, + "learning_rate": 0.0001176581441571002, + "loss": 3.3176692962646483, + "step": 3075 + }, + { + "epoch": 0.5917387127761767, + "grad_norm": 0.8082125782966614, + "learning_rate": 0.00011719294116182217, + "loss": 3.314889144897461, + "step": 3080 + }, + { + "epoch": 0.5926993275696446, + "grad_norm": 0.6132190823554993, + "learning_rate": 0.00011672806933560925, + "loss": 3.3137298583984376, + "step": 3085 + }, + { + "epoch": 0.5936599423631124, + "grad_norm": 0.6647592782974243, + "learning_rate": 0.00011626353337108476, + "loss": 3.3115211486816407, + "step": 3090 + }, + { + "epoch": 0.5946205571565802, + "grad_norm": 0.5742537379264832, + "learning_rate": 0.00011579933795748164, + "loss": 3.3127456665039063, + "step": 3095 + }, + { + "epoch": 0.595581171950048, + "grad_norm": 0.6123734712600708, + "learning_rate": 0.00011533548778059508, + "loss": 3.315221405029297, + "step": 3100 + }, + { + "epoch": 0.5965417867435159, + "grad_norm": 0.8112967014312744, + "learning_rate": 0.00011487198752273552, + "loss": 3.315155792236328, + "step": 3105 + }, + { + "epoch": 0.5975024015369836, + "grad_norm": 1.0377261638641357, + "learning_rate": 0.0001144088418626809, + "loss": 3.31671142578125, + "step": 3110 + }, + { + "epoch": 0.5984630163304515, + "grad_norm": 0.6357848644256592, + "learning_rate": 0.00011394605547562989, + "loss": 3.314447784423828, + "step": 3115 + }, + { + "epoch": 0.5994236311239193, + "grad_norm": 0.5305024981498718, + "learning_rate": 0.00011348363303315434, + "loss": 3.313646697998047, + "step": 3120 + }, + { + "epoch": 0.6003842459173871, + "grad_norm": 0.6368167400360107, + "learning_rate": 0.00011302157920315244, + "loss": 3.3107112884521483, + "step": 3125 + }, + { + "epoch": 0.6013448607108549, + "grad_norm": 0.6398528218269348, + "learning_rate": 0.00011255989864980133, + "loss": 3.3126060485839846, + "step": 3130 + }, + { + "epoch": 0.6023054755043228, + "grad_norm": 0.7452788352966309, + "learning_rate": 0.00011209859603351015, + "loss": 3.3152156829833985, + "step": 3135 + }, + { + "epoch": 0.6032660902977905, + "grad_norm": 0.6287251710891724, + "learning_rate": 0.00011163767601087301, + "loss": 3.3126220703125, + "step": 3140 + }, + { + "epoch": 0.6042267050912584, + "grad_norm": 0.6889411807060242, + "learning_rate": 0.00011117714323462186, + "loss": 3.3144298553466798, + "step": 3145 + }, + { + "epoch": 0.6051873198847262, + "grad_norm": 0.6560638546943665, + "learning_rate": 0.00011071700235357979, + "loss": 3.311948776245117, + "step": 3150 + }, + { + "epoch": 0.6061479346781941, + "grad_norm": 0.4844714403152466, + "learning_rate": 0.00011025725801261373, + "loss": 3.3093055725097655, + "step": 3155 + }, + { + "epoch": 0.6071085494716618, + "grad_norm": 0.8295413255691528, + "learning_rate": 0.00010979791485258788, + "loss": 3.3112449645996094, + "step": 3160 + }, + { + "epoch": 0.6080691642651297, + "grad_norm": 0.8754414916038513, + "learning_rate": 0.00010933897751031671, + "loss": 3.311637115478516, + "step": 3165 + }, + { + "epoch": 0.6090297790585975, + "grad_norm": 0.5046402812004089, + "learning_rate": 0.00010888045061851807, + "loss": 3.309844207763672, + "step": 3170 + }, + { + "epoch": 0.6099903938520653, + "grad_norm": 0.6311172246932983, + "learning_rate": 0.00010842233880576681, + "loss": 3.3105926513671875, + "step": 3175 + }, + { + "epoch": 0.6109510086455331, + "grad_norm": 0.6714319586753845, + "learning_rate": 0.0001079646466964475, + "loss": 3.308180236816406, + "step": 3180 + }, + { + "epoch": 0.611911623439001, + "grad_norm": 0.6159961223602295, + "learning_rate": 0.00010750737891070824, + "loss": 3.3096988677978514, + "step": 3185 + }, + { + "epoch": 0.6128722382324687, + "grad_norm": 0.5705515742301941, + "learning_rate": 0.00010705054006441371, + "loss": 3.3115421295166017, + "step": 3190 + }, + { + "epoch": 0.6138328530259366, + "grad_norm": 0.49575576186180115, + "learning_rate": 0.00010659413476909865, + "loss": 3.3115379333496096, + "step": 3195 + }, + { + "epoch": 0.6147934678194045, + "grad_norm": 1.0461304187774658, + "learning_rate": 0.00010613816763192152, + "loss": 3.3100265502929687, + "step": 3200 + }, + { + "epoch": 0.6157540826128722, + "grad_norm": 0.6978023648262024, + "learning_rate": 0.00010568264325561763, + "loss": 3.308924102783203, + "step": 3205 + }, + { + "epoch": 0.6167146974063401, + "grad_norm": 0.7631133198738098, + "learning_rate": 0.000105227566238453, + "loss": 3.3145515441894533, + "step": 3210 + }, + { + "epoch": 0.6176753121998079, + "grad_norm": 0.8663102388381958, + "learning_rate": 0.00010477294117417762, + "loss": 3.3145401000976564, + "step": 3215 + }, + { + "epoch": 0.6186359269932757, + "grad_norm": 0.4759737551212311, + "learning_rate": 0.00010431877265197955, + "loss": 3.311273193359375, + "step": 3220 + }, + { + "epoch": 0.6195965417867435, + "grad_norm": 0.47145599126815796, + "learning_rate": 0.00010386506525643808, + "loss": 3.306778335571289, + "step": 3225 + }, + { + "epoch": 0.6205571565802114, + "grad_norm": 0.6582973599433899, + "learning_rate": 0.00010341182356747771, + "loss": 3.3100128173828125, + "step": 3230 + }, + { + "epoch": 0.6215177713736791, + "grad_norm": 0.6123160123825073, + "learning_rate": 0.00010295905216032203, + "loss": 3.3093284606933593, + "step": 3235 + }, + { + "epoch": 0.622478386167147, + "grad_norm": 0.5266756415367126, + "learning_rate": 0.00010250675560544717, + "loss": 3.3095718383789063, + "step": 3240 + }, + { + "epoch": 0.6234390009606148, + "grad_norm": 0.7521538138389587, + "learning_rate": 0.00010205493846853618, + "loss": 3.307404327392578, + "step": 3245 + }, + { + "epoch": 0.6243996157540826, + "grad_norm": 0.5699673891067505, + "learning_rate": 0.00010160360531043239, + "loss": 3.3089645385742186, + "step": 3250 + }, + { + "epoch": 0.6253602305475504, + "grad_norm": 0.7614957690238953, + "learning_rate": 0.00010115276068709377, + "loss": 3.3093006134033205, + "step": 3255 + }, + { + "epoch": 0.6263208453410183, + "grad_norm": 0.652992844581604, + "learning_rate": 0.00010070240914954676, + "loss": 3.306406784057617, + "step": 3260 + }, + { + "epoch": 0.627281460134486, + "grad_norm": 0.5600461363792419, + "learning_rate": 0.00010025255524384033, + "loss": 3.3075355529785155, + "step": 3265 + }, + { + "epoch": 0.6282420749279539, + "grad_norm": 0.7442777156829834, + "learning_rate": 9.980320351100028e-05, + "loss": 3.308829498291016, + "step": 3270 + }, + { + "epoch": 0.6292026897214217, + "grad_norm": 0.8243304491043091, + "learning_rate": 9.935435848698307e-05, + "loss": 3.307141876220703, + "step": 3275 + }, + { + "epoch": 0.6301633045148896, + "grad_norm": 0.6337246298789978, + "learning_rate": 9.890602470263037e-05, + "loss": 3.311864471435547, + "step": 3280 + }, + { + "epoch": 0.6311239193083573, + "grad_norm": 0.6521264910697937, + "learning_rate": 9.845820668362308e-05, + "loss": 3.308788299560547, + "step": 3285 + }, + { + "epoch": 0.6320845341018252, + "grad_norm": 0.6866946220397949, + "learning_rate": 9.801090895043566e-05, + "loss": 3.307585906982422, + "step": 3290 + }, + { + "epoch": 0.633045148895293, + "grad_norm": 0.6082757711410522, + "learning_rate": 9.756413601829083e-05, + "loss": 3.307001495361328, + "step": 3295 + }, + { + "epoch": 0.6340057636887608, + "grad_norm": 0.515135645866394, + "learning_rate": 9.711789239711344e-05, + "loss": 3.308676528930664, + "step": 3300 + }, + { + "epoch": 0.6349663784822286, + "grad_norm": 0.503957211971283, + "learning_rate": 9.667218259148547e-05, + "loss": 3.3088623046875, + "step": 3305 + }, + { + "epoch": 0.6359269932756965, + "grad_norm": 0.663313627243042, + "learning_rate": 9.62270111006001e-05, + "loss": 3.305764007568359, + "step": 3310 + }, + { + "epoch": 0.6368876080691642, + "grad_norm": 0.7052382826805115, + "learning_rate": 9.57823824182168e-05, + "loss": 3.305963134765625, + "step": 3315 + }, + { + "epoch": 0.6378482228626321, + "grad_norm": 0.5659816861152649, + "learning_rate": 9.53383010326155e-05, + "loss": 3.306070327758789, + "step": 3320 + }, + { + "epoch": 0.6388088376560999, + "grad_norm": 0.46071678400039673, + "learning_rate": 9.489477142655147e-05, + "loss": 3.3049732208251954, + "step": 3325 + }, + { + "epoch": 0.6397694524495677, + "grad_norm": 0.9557482004165649, + "learning_rate": 9.445179807721012e-05, + "loss": 3.3062675476074217, + "step": 3330 + }, + { + "epoch": 0.6407300672430355, + "grad_norm": 0.8419716954231262, + "learning_rate": 9.400938545616173e-05, + "loss": 3.308432769775391, + "step": 3335 + }, + { + "epoch": 0.6416906820365034, + "grad_norm": 0.5912371873855591, + "learning_rate": 9.356753802931646e-05, + "loss": 3.3038196563720703, + "step": 3340 + }, + { + "epoch": 0.6426512968299711, + "grad_norm": 0.4956425726413727, + "learning_rate": 9.312626025687897e-05, + "loss": 3.3066734313964843, + "step": 3345 + }, + { + "epoch": 0.643611911623439, + "grad_norm": 0.5250487923622131, + "learning_rate": 9.268555659330364e-05, + "loss": 3.307097625732422, + "step": 3350 + }, + { + "epoch": 0.6445725264169068, + "grad_norm": 0.5282003283500671, + "learning_rate": 9.22454314872496e-05, + "loss": 3.305289459228516, + "step": 3355 + }, + { + "epoch": 0.6455331412103746, + "grad_norm": 0.4647374749183655, + "learning_rate": 9.180588938153566e-05, + "loss": 3.3030609130859374, + "step": 3360 + }, + { + "epoch": 0.6464937560038425, + "grad_norm": 1.0786409378051758, + "learning_rate": 9.136693471309568e-05, + "loss": 3.311345672607422, + "step": 3365 + }, + { + "epoch": 0.6474543707973103, + "grad_norm": 0.6998493075370789, + "learning_rate": 9.092857191293356e-05, + "loss": 3.3091506958007812, + "step": 3370 + }, + { + "epoch": 0.6484149855907781, + "grad_norm": 0.446397066116333, + "learning_rate": 9.049080540607875e-05, + "loss": 3.3034263610839845, + "step": 3375 + }, + { + "epoch": 0.6493756003842459, + "grad_norm": 0.6110076308250427, + "learning_rate": 9.005363961154126e-05, + "loss": 3.3031074523925783, + "step": 3380 + }, + { + "epoch": 0.6503362151777138, + "grad_norm": 0.4611775279045105, + "learning_rate": 8.961707894226735e-05, + "loss": 3.3057632446289062, + "step": 3385 + }, + { + "epoch": 0.6512968299711815, + "grad_norm": 0.8725507855415344, + "learning_rate": 8.918112780509494e-05, + "loss": 3.3007755279541016, + "step": 3390 + }, + { + "epoch": 0.6522574447646494, + "grad_norm": 0.8378633856773376, + "learning_rate": 8.874579060070894e-05, + "loss": 3.3028465270996095, + "step": 3395 + }, + { + "epoch": 0.6532180595581172, + "grad_norm": 0.6364403367042542, + "learning_rate": 8.831107172359707e-05, + "loss": 3.3054805755615235, + "step": 3400 + }, + { + "epoch": 0.654178674351585, + "grad_norm": 0.46224355697631836, + "learning_rate": 8.787697556200519e-05, + "loss": 3.304944610595703, + "step": 3405 + }, + { + "epoch": 0.6551392891450528, + "grad_norm": 0.5552151799201965, + "learning_rate": 8.744350649789347e-05, + "loss": 3.301805114746094, + "step": 3410 + }, + { + "epoch": 0.6560999039385207, + "grad_norm": 0.6075546741485596, + "learning_rate": 8.701066890689166e-05, + "loss": 3.300829315185547, + "step": 3415 + }, + { + "epoch": 0.6570605187319885, + "grad_norm": 0.46151185035705566, + "learning_rate": 8.657846715825508e-05, + "loss": 3.30389289855957, + "step": 3420 + }, + { + "epoch": 0.6580211335254563, + "grad_norm": 0.5615620017051697, + "learning_rate": 8.61469056148209e-05, + "loss": 3.302476501464844, + "step": 3425 + }, + { + "epoch": 0.6589817483189241, + "grad_norm": 0.47145524621009827, + "learning_rate": 8.571598863296342e-05, + "loss": 3.303362274169922, + "step": 3430 + }, + { + "epoch": 0.659942363112392, + "grad_norm": 0.4290817081928253, + "learning_rate": 8.528572056255065e-05, + "loss": 3.3021663665771483, + "step": 3435 + }, + { + "epoch": 0.6609029779058597, + "grad_norm": 0.5213423371315002, + "learning_rate": 8.485610574690021e-05, + "loss": 3.3011238098144533, + "step": 3440 + }, + { + "epoch": 0.6618635926993276, + "grad_norm": 0.661228358745575, + "learning_rate": 8.442714852273523e-05, + "loss": 3.3015186309814455, + "step": 3445 + }, + { + "epoch": 0.6628242074927954, + "grad_norm": 0.9645004868507385, + "learning_rate": 8.399885322014123e-05, + "loss": 3.3042228698730467, + "step": 3450 + }, + { + "epoch": 0.6637848222862632, + "grad_norm": 0.5251511335372925, + "learning_rate": 8.35712241625216e-05, + "loss": 3.3040817260742186, + "step": 3455 + }, + { + "epoch": 0.664745437079731, + "grad_norm": 0.45383018255233765, + "learning_rate": 8.314426566655458e-05, + "loss": 3.3012584686279296, + "step": 3460 + }, + { + "epoch": 0.6657060518731989, + "grad_norm": 0.3945951461791992, + "learning_rate": 8.271798204214942e-05, + "loss": 3.3004776000976563, + "step": 3465 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 0.4676642119884491, + "learning_rate": 8.229237759240289e-05, + "loss": 3.2975730895996094, + "step": 3470 + }, + { + "epoch": 0.6676272814601345, + "grad_norm": 0.5338938236236572, + "learning_rate": 8.186745661355595e-05, + "loss": 3.297288513183594, + "step": 3475 + }, + { + "epoch": 0.6685878962536023, + "grad_norm": 0.6830979585647583, + "learning_rate": 8.144322339495012e-05, + "loss": 3.3035240173339844, + "step": 3480 + }, + { + "epoch": 0.6695485110470701, + "grad_norm": 0.4912271201610565, + "learning_rate": 8.101968221898453e-05, + "loss": 3.300017547607422, + "step": 3485 + }, + { + "epoch": 0.6705091258405379, + "grad_norm": 0.5433835387229919, + "learning_rate": 8.059683736107245e-05, + "loss": 3.304372787475586, + "step": 3490 + }, + { + "epoch": 0.6714697406340058, + "grad_norm": 0.43095162510871887, + "learning_rate": 8.017469308959823e-05, + "loss": 3.302097702026367, + "step": 3495 + }, + { + "epoch": 0.6724303554274735, + "grad_norm": 0.5721524953842163, + "learning_rate": 7.97532536658742e-05, + "loss": 3.302288055419922, + "step": 3500 + }, + { + "epoch": 0.6733909702209414, + "grad_norm": 0.8145022392272949, + "learning_rate": 7.933252334409766e-05, + "loss": 3.302367401123047, + "step": 3505 + }, + { + "epoch": 0.6743515850144092, + "grad_norm": 0.41017457842826843, + "learning_rate": 7.891250637130779e-05, + "loss": 3.3008522033691405, + "step": 3510 + }, + { + "epoch": 0.675312199807877, + "grad_norm": 0.4966427981853485, + "learning_rate": 7.849320698734306e-05, + "loss": 3.305078887939453, + "step": 3515 + }, + { + "epoch": 0.6762728146013448, + "grad_norm": 0.44679656624794006, + "learning_rate": 7.80746294247982e-05, + "loss": 3.297787094116211, + "step": 3520 + }, + { + "epoch": 0.6772334293948127, + "grad_norm": 0.5052957534790039, + "learning_rate": 7.765677790898155e-05, + "loss": 3.2971946716308596, + "step": 3525 + }, + { + "epoch": 0.6781940441882806, + "grad_norm": 0.5869050025939941, + "learning_rate": 7.723965665787255e-05, + "loss": 3.3033409118652344, + "step": 3530 + }, + { + "epoch": 0.6791546589817483, + "grad_norm": 0.5643454194068909, + "learning_rate": 7.682326988207877e-05, + "loss": 3.3006324768066406, + "step": 3535 + }, + { + "epoch": 0.6801152737752162, + "grad_norm": 0.45685872435569763, + "learning_rate": 7.640762178479382e-05, + "loss": 3.2982528686523436, + "step": 3540 + }, + { + "epoch": 0.681075888568684, + "grad_norm": 0.42118874192237854, + "learning_rate": 7.599271656175476e-05, + "loss": 3.303014373779297, + "step": 3545 + }, + { + "epoch": 0.6820365033621518, + "grad_norm": 0.482282817363739, + "learning_rate": 7.557855840119976e-05, + "loss": 3.2992012023925783, + "step": 3550 + }, + { + "epoch": 0.6829971181556196, + "grad_norm": 0.4499205946922302, + "learning_rate": 7.516515148382576e-05, + "loss": 3.2981277465820313, + "step": 3555 + }, + { + "epoch": 0.6839577329490875, + "grad_norm": 0.6683313846588135, + "learning_rate": 7.475249998274621e-05, + "loss": 3.298491668701172, + "step": 3560 + }, + { + "epoch": 0.6849183477425552, + "grad_norm": 0.5663706064224243, + "learning_rate": 7.434060806344936e-05, + "loss": 3.301411819458008, + "step": 3565 + }, + { + "epoch": 0.6858789625360231, + "grad_norm": 0.43267935514450073, + "learning_rate": 7.392947988375555e-05, + "loss": 3.29864616394043, + "step": 3570 + }, + { + "epoch": 0.6868395773294909, + "grad_norm": 0.43747061491012573, + "learning_rate": 7.351911959377585e-05, + "loss": 3.298960876464844, + "step": 3575 + }, + { + "epoch": 0.6878001921229587, + "grad_norm": 0.5621949434280396, + "learning_rate": 7.310953133586981e-05, + "loss": 3.2972442626953127, + "step": 3580 + }, + { + "epoch": 0.6887608069164265, + "grad_norm": 0.4505578279495239, + "learning_rate": 7.27007192446036e-05, + "loss": 3.2979637145996095, + "step": 3585 + }, + { + "epoch": 0.6897214217098944, + "grad_norm": 0.7796880602836609, + "learning_rate": 7.229268744670883e-05, + "loss": 3.297809600830078, + "step": 3590 + }, + { + "epoch": 0.6906820365033621, + "grad_norm": 0.6508268713951111, + "learning_rate": 7.188544006104e-05, + "loss": 3.2982086181640624, + "step": 3595 + }, + { + "epoch": 0.69164265129683, + "grad_norm": 0.44308528304100037, + "learning_rate": 7.147898119853367e-05, + "loss": 3.2976150512695312, + "step": 3600 + }, + { + "epoch": 0.6926032660902978, + "grad_norm": 0.4720350205898285, + "learning_rate": 7.107331496216676e-05, + "loss": 3.2995452880859375, + "step": 3605 + }, + { + "epoch": 0.6935638808837656, + "grad_norm": 0.45068585872650146, + "learning_rate": 7.066844544691474e-05, + "loss": 3.301127243041992, + "step": 3610 + }, + { + "epoch": 0.6945244956772334, + "grad_norm": 0.42395853996276855, + "learning_rate": 7.026437673971107e-05, + "loss": 3.2964126586914064, + "step": 3615 + }, + { + "epoch": 0.6954851104707013, + "grad_norm": 0.4400024712085724, + "learning_rate": 6.986111291940511e-05, + "loss": 3.2967502593994142, + "step": 3620 + }, + { + "epoch": 0.696445725264169, + "grad_norm": 0.410159170627594, + "learning_rate": 6.945865805672154e-05, + "loss": 3.2917686462402345, + "step": 3625 + }, + { + "epoch": 0.6974063400576369, + "grad_norm": 0.6144173741340637, + "learning_rate": 6.905701621421904e-05, + "loss": 3.300373077392578, + "step": 3630 + }, + { + "epoch": 0.6983669548511047, + "grad_norm": 0.47794288396835327, + "learning_rate": 6.865619144624914e-05, + "loss": 3.297341156005859, + "step": 3635 + }, + { + "epoch": 0.6993275696445725, + "grad_norm": 0.520939826965332, + "learning_rate": 6.825618779891577e-05, + "loss": 3.298029327392578, + "step": 3640 + }, + { + "epoch": 0.7002881844380403, + "grad_norm": 0.5281940698623657, + "learning_rate": 6.785700931003381e-05, + "loss": 3.291861724853516, + "step": 3645 + }, + { + "epoch": 0.7012487992315082, + "grad_norm": 0.3904149532318115, + "learning_rate": 6.745866000908874e-05, + "loss": 3.296683502197266, + "step": 3650 + }, + { + "epoch": 0.7022094140249759, + "grad_norm": 0.47106507420539856, + "learning_rate": 6.706114391719586e-05, + "loss": 3.29759521484375, + "step": 3655 + }, + { + "epoch": 0.7031700288184438, + "grad_norm": 0.4817742705345154, + "learning_rate": 6.66644650470597e-05, + "loss": 3.299455261230469, + "step": 3660 + }, + { + "epoch": 0.7041306436119116, + "grad_norm": 0.473783940076828, + "learning_rate": 6.626862740293338e-05, + "loss": 3.2957839965820312, + "step": 3665 + }, + { + "epoch": 0.7050912584053795, + "grad_norm": 0.5512199401855469, + "learning_rate": 6.587363498057845e-05, + "loss": 3.297314453125, + "step": 3670 + }, + { + "epoch": 0.7060518731988472, + "grad_norm": 0.6732028722763062, + "learning_rate": 6.547949176722437e-05, + "loss": 3.2939273834228517, + "step": 3675 + }, + { + "epoch": 0.7070124879923151, + "grad_norm": 0.5238297581672668, + "learning_rate": 6.508620174152826e-05, + "loss": 3.298822784423828, + "step": 3680 + }, + { + "epoch": 0.7079731027857828, + "grad_norm": 0.40396222472190857, + "learning_rate": 6.469376887353491e-05, + "loss": 3.29783935546875, + "step": 3685 + }, + { + "epoch": 0.7089337175792507, + "grad_norm": 0.4132029116153717, + "learning_rate": 6.430219712463634e-05, + "loss": 3.2966102600097655, + "step": 3690 + }, + { + "epoch": 0.7098943323727186, + "grad_norm": 0.4176599085330963, + "learning_rate": 6.391149044753238e-05, + "loss": 3.2970817565917967, + "step": 3695 + }, + { + "epoch": 0.7108549471661864, + "grad_norm": 0.3097979426383972, + "learning_rate": 6.352165278619012e-05, + "loss": 3.294615936279297, + "step": 3700 + }, + { + "epoch": 0.7118155619596542, + "grad_norm": 0.4419673979282379, + "learning_rate": 6.313268807580462e-05, + "loss": 3.2989078521728517, + "step": 3705 + }, + { + "epoch": 0.712776176753122, + "grad_norm": 0.5254682302474976, + "learning_rate": 6.274460024275895e-05, + "loss": 3.2966861724853516, + "step": 3710 + }, + { + "epoch": 0.7137367915465899, + "grad_norm": 0.5654774308204651, + "learning_rate": 6.235739320458442e-05, + "loss": 3.293225860595703, + "step": 3715 + }, + { + "epoch": 0.7146974063400576, + "grad_norm": 0.35445114970207214, + "learning_rate": 6.197107086992156e-05, + "loss": 3.2962112426757812, + "step": 3720 + }, + { + "epoch": 0.7156580211335255, + "grad_norm": 0.43310287594795227, + "learning_rate": 6.158563713847994e-05, + "loss": 3.2953845977783205, + "step": 3725 + }, + { + "epoch": 0.7166186359269933, + "grad_norm": 0.4213480055332184, + "learning_rate": 6.12010959009994e-05, + "loss": 3.2970962524414062, + "step": 3730 + }, + { + "epoch": 0.7175792507204611, + "grad_norm": 0.3784043490886688, + "learning_rate": 6.081745103921047e-05, + "loss": 3.2936870574951174, + "step": 3735 + }, + { + "epoch": 0.7185398655139289, + "grad_norm": 0.6028347611427307, + "learning_rate": 6.043470642579516e-05, + "loss": 3.2977680206298827, + "step": 3740 + }, + { + "epoch": 0.7195004803073968, + "grad_norm": 0.44696855545043945, + "learning_rate": 6.005286592434828e-05, + "loss": 3.293193817138672, + "step": 3745 + }, + { + "epoch": 0.7204610951008645, + "grad_norm": 0.4338856339454651, + "learning_rate": 5.967193338933778e-05, + "loss": 3.294358825683594, + "step": 3750 + }, + { + "epoch": 0.7214217098943324, + "grad_norm": 0.4633798599243164, + "learning_rate": 5.9291912666066405e-05, + "loss": 3.2951465606689454, + "step": 3755 + }, + { + "epoch": 0.7223823246878002, + "grad_norm": 0.42019450664520264, + "learning_rate": 5.891280759063265e-05, + "loss": 3.292225646972656, + "step": 3760 + }, + { + "epoch": 0.723342939481268, + "grad_norm": 0.418599396944046, + "learning_rate": 5.853462198989184e-05, + "loss": 3.2958885192871095, + "step": 3765 + }, + { + "epoch": 0.7243035542747358, + "grad_norm": 0.35753360390663147, + "learning_rate": 5.815735968141813e-05, + "loss": 3.2968238830566405, + "step": 3770 + }, + { + "epoch": 0.7252641690682037, + "grad_norm": 0.4825795888900757, + "learning_rate": 5.778102447346514e-05, + "loss": 3.2957897186279297, + "step": 3775 + }, + { + "epoch": 0.7262247838616714, + "grad_norm": 0.41493552923202515, + "learning_rate": 5.740562016492811e-05, + "loss": 3.2921653747558595, + "step": 3780 + }, + { + "epoch": 0.7271853986551393, + "grad_norm": 0.5309422612190247, + "learning_rate": 5.703115054530537e-05, + "loss": 3.293100357055664, + "step": 3785 + }, + { + "epoch": 0.7281460134486071, + "grad_norm": 0.36943376064300537, + "learning_rate": 5.665761939466008e-05, + "loss": 3.294559860229492, + "step": 3790 + }, + { + "epoch": 0.729106628242075, + "grad_norm": 0.5610998272895813, + "learning_rate": 5.628503048358207e-05, + "loss": 3.292363739013672, + "step": 3795 + }, + { + "epoch": 0.7300672430355427, + "grad_norm": 0.5705908536911011, + "learning_rate": 5.591338757314968e-05, + "loss": 3.2965774536132812, + "step": 3800 + }, + { + "epoch": 0.7310278578290106, + "grad_norm": 0.5056052803993225, + "learning_rate": 5.554269441489204e-05, + "loss": 3.295191192626953, + "step": 3805 + }, + { + "epoch": 0.7319884726224783, + "grad_norm": 0.4820818603038788, + "learning_rate": 5.517295475075102e-05, + "loss": 3.2937217712402345, + "step": 3810 + }, + { + "epoch": 0.7329490874159462, + "grad_norm": 0.3636917471885681, + "learning_rate": 5.4804172313043465e-05, + "loss": 3.291709136962891, + "step": 3815 + }, + { + "epoch": 0.733909702209414, + "grad_norm": 0.3684285283088684, + "learning_rate": 5.443635082442363e-05, + "loss": 3.293587493896484, + "step": 3820 + }, + { + "epoch": 0.7348703170028819, + "grad_norm": 0.741382896900177, + "learning_rate": 5.4069493997845356e-05, + "loss": 3.2945404052734375, + "step": 3825 + }, + { + "epoch": 0.7358309317963496, + "grad_norm": 0.41890212893486023, + "learning_rate": 5.3703605536524905e-05, + "loss": 3.2955181121826174, + "step": 3830 + }, + { + "epoch": 0.7367915465898175, + "grad_norm": 0.3295837938785553, + "learning_rate": 5.333868913390338e-05, + "loss": 3.290005111694336, + "step": 3835 + }, + { + "epoch": 0.7377521613832853, + "grad_norm": 0.3231881558895111, + "learning_rate": 5.2974748473609505e-05, + "loss": 3.291563034057617, + "step": 3840 + }, + { + "epoch": 0.7387127761767531, + "grad_norm": 0.44054171442985535, + "learning_rate": 5.261178722942242e-05, + "loss": 3.2941848754882814, + "step": 3845 + }, + { + "epoch": 0.7396733909702209, + "grad_norm": 0.3981585204601288, + "learning_rate": 5.224980906523462e-05, + "loss": 3.292535400390625, + "step": 3850 + }, + { + "epoch": 0.7406340057636888, + "grad_norm": 0.36969125270843506, + "learning_rate": 5.188881763501486e-05, + "loss": 3.2916770935058595, + "step": 3855 + }, + { + "epoch": 0.7415946205571565, + "grad_norm": 0.414196252822876, + "learning_rate": 5.152881658277147e-05, + "loss": 3.293859100341797, + "step": 3860 + }, + { + "epoch": 0.7425552353506244, + "grad_norm": 0.3872639536857605, + "learning_rate": 5.1169809542515404e-05, + "loss": 3.296031188964844, + "step": 3865 + }, + { + "epoch": 0.7435158501440923, + "grad_norm": 0.4767828583717346, + "learning_rate": 5.081180013822368e-05, + "loss": 3.292156219482422, + "step": 3870 + }, + { + "epoch": 0.74447646493756, + "grad_norm": 0.41197365522384644, + "learning_rate": 5.045479198380272e-05, + "loss": 3.292400360107422, + "step": 3875 + }, + { + "epoch": 0.7454370797310279, + "grad_norm": 0.40532881021499634, + "learning_rate": 5.009878868305171e-05, + "loss": 3.2940216064453125, + "step": 3880 + }, + { + "epoch": 0.7463976945244957, + "grad_norm": 0.3654780685901642, + "learning_rate": 4.9743793829626736e-05, + "loss": 3.293231964111328, + "step": 3885 + }, + { + "epoch": 0.7473583093179635, + "grad_norm": 0.2990121841430664, + "learning_rate": 4.9389811007003834e-05, + "loss": 3.292677307128906, + "step": 3890 + }, + { + "epoch": 0.7483189241114313, + "grad_norm": 0.3309316039085388, + "learning_rate": 4.903684378844333e-05, + "loss": 3.294384765625, + "step": 3895 + }, + { + "epoch": 0.7492795389048992, + "grad_norm": 0.2731437683105469, + "learning_rate": 4.86848957369536e-05, + "loss": 3.292652893066406, + "step": 3900 + }, + { + "epoch": 0.7502401536983669, + "grad_norm": 0.3116244971752167, + "learning_rate": 4.8333970405254904e-05, + "loss": 3.288296127319336, + "step": 3905 + }, + { + "epoch": 0.7512007684918348, + "grad_norm": 0.42631855607032776, + "learning_rate": 4.798407133574405e-05, + "loss": 3.2939361572265624, + "step": 3910 + }, + { + "epoch": 0.7521613832853026, + "grad_norm": 0.35764989256858826, + "learning_rate": 4.7635202060457945e-05, + "loss": 3.2916053771972655, + "step": 3915 + }, + { + "epoch": 0.7531219980787704, + "grad_norm": 0.2865849733352661, + "learning_rate": 4.72873661010385e-05, + "loss": 3.2932079315185545, + "step": 3920 + }, + { + "epoch": 0.7540826128722382, + "grad_norm": 0.3286271095275879, + "learning_rate": 4.694056696869688e-05, + "loss": 3.295194625854492, + "step": 3925 + }, + { + "epoch": 0.7550432276657061, + "grad_norm": 0.346123605966568, + "learning_rate": 4.659480816417785e-05, + "loss": 3.288124847412109, + "step": 3930 + }, + { + "epoch": 0.7560038424591738, + "grad_norm": 0.4188701808452606, + "learning_rate": 4.6250093177725e-05, + "loss": 3.2938987731933596, + "step": 3935 + }, + { + "epoch": 0.7569644572526417, + "grad_norm": 0.44901224970817566, + "learning_rate": 4.590642548904479e-05, + "loss": 3.2915252685546874, + "step": 3940 + }, + { + "epoch": 0.7579250720461095, + "grad_norm": 0.4020468592643738, + "learning_rate": 4.5563808567272e-05, + "loss": 3.290658187866211, + "step": 3945 + }, + { + "epoch": 0.7588856868395774, + "grad_norm": 0.32219162583351135, + "learning_rate": 4.52222458709345e-05, + "loss": 3.2903762817382813, + "step": 3950 + }, + { + "epoch": 0.7598463016330451, + "grad_norm": 0.25904515385627747, + "learning_rate": 4.4881740847918155e-05, + "loss": 3.288920593261719, + "step": 3955 + }, + { + "epoch": 0.760806916426513, + "grad_norm": 0.29081347584724426, + "learning_rate": 4.454229693543251e-05, + "loss": 3.293811798095703, + "step": 3960 + }, + { + "epoch": 0.7617675312199808, + "grad_norm": 0.24549178779125214, + "learning_rate": 4.420391755997548e-05, + "loss": 3.2932552337646483, + "step": 3965 + }, + { + "epoch": 0.7627281460134486, + "grad_norm": 0.459926038980484, + "learning_rate": 4.386660613729925e-05, + "loss": 3.2938629150390626, + "step": 3970 + }, + { + "epoch": 0.7636887608069164, + "grad_norm": 0.4720708429813385, + "learning_rate": 4.35303660723756e-05, + "loss": 3.2878982543945314, + "step": 3975 + }, + { + "epoch": 0.7646493756003843, + "grad_norm": 0.28443172574043274, + "learning_rate": 4.3195200759361455e-05, + "loss": 3.2919075012207033, + "step": 3980 + }, + { + "epoch": 0.765609990393852, + "grad_norm": 0.33102986216545105, + "learning_rate": 4.2861113581564884e-05, + "loss": 3.2893836975097654, + "step": 3985 + }, + { + "epoch": 0.7665706051873199, + "grad_norm": 0.348117858171463, + "learning_rate": 4.252810791141054e-05, + "loss": 3.293404769897461, + "step": 3990 + }, + { + "epoch": 0.7675312199807877, + "grad_norm": 0.3010897636413574, + "learning_rate": 4.2196187110406054e-05, + "loss": 3.2900314331054688, + "step": 3995 + }, + { + "epoch": 0.7684918347742555, + "grad_norm": 0.274684339761734, + "learning_rate": 4.186535452910784e-05, + "loss": 3.2879261016845702, + "step": 4000 + }, + { + "epoch": 0.7694524495677233, + "grad_norm": 0.29900944232940674, + "learning_rate": 4.153561350708732e-05, + "loss": 3.292841339111328, + "step": 4005 + }, + { + "epoch": 0.7704130643611912, + "grad_norm": 0.30298614501953125, + "learning_rate": 4.12069673728973e-05, + "loss": 3.2894565582275392, + "step": 4010 + }, + { + "epoch": 0.7713736791546589, + "grad_norm": 0.31402137875556946, + "learning_rate": 4.087941944403815e-05, + "loss": 3.2861660003662108, + "step": 4015 + }, + { + "epoch": 0.7723342939481268, + "grad_norm": 0.4134189784526825, + "learning_rate": 4.0552973026924625e-05, + "loss": 3.289139175415039, + "step": 4020 + }, + { + "epoch": 0.7732949087415946, + "grad_norm": 0.3398532569408417, + "learning_rate": 4.022763141685226e-05, + "loss": 3.2890396118164062, + "step": 4025 + }, + { + "epoch": 0.7742555235350624, + "grad_norm": 0.33045315742492676, + "learning_rate": 3.990339789796418e-05, + "loss": 3.2925796508789062, + "step": 4030 + }, + { + "epoch": 0.7752161383285303, + "grad_norm": 0.28469014167785645, + "learning_rate": 3.958027574321794e-05, + "loss": 3.2899627685546875, + "step": 4035 + }, + { + "epoch": 0.7761767531219981, + "grad_norm": 0.23531687259674072, + "learning_rate": 3.9258268214352566e-05, + "loss": 3.286402130126953, + "step": 4040 + }, + { + "epoch": 0.777137367915466, + "grad_norm": 0.3813510239124298, + "learning_rate": 3.893737856185538e-05, + "loss": 3.2916938781738283, + "step": 4045 + }, + { + "epoch": 0.7780979827089337, + "grad_norm": 0.3439390957355499, + "learning_rate": 3.861761002492952e-05, + "loss": 3.288800811767578, + "step": 4050 + }, + { + "epoch": 0.7790585975024016, + "grad_norm": 0.2623279392719269, + "learning_rate": 3.8298965831461024e-05, + "loss": 3.288500213623047, + "step": 4055 + }, + { + "epoch": 0.7800192122958693, + "grad_norm": 0.3051926791667938, + "learning_rate": 3.798144919798631e-05, + "loss": 3.2871044158935545, + "step": 4060 + }, + { + "epoch": 0.7809798270893372, + "grad_norm": 0.2598731815814972, + "learning_rate": 3.766506332965976e-05, + "loss": 3.2878665924072266, + "step": 4065 + }, + { + "epoch": 0.781940441882805, + "grad_norm": 0.3285099267959595, + "learning_rate": 3.734981142022117e-05, + "loss": 3.287023162841797, + "step": 4070 + }, + { + "epoch": 0.7829010566762729, + "grad_norm": 0.2698347568511963, + "learning_rate": 3.70356966519638e-05, + "loss": 3.2910301208496096, + "step": 4075 + }, + { + "epoch": 0.7838616714697406, + "grad_norm": 0.2935537099838257, + "learning_rate": 3.672272219570199e-05, + "loss": 3.2882232666015625, + "step": 4080 + }, + { + "epoch": 0.7848222862632085, + "grad_norm": 0.2180212140083313, + "learning_rate": 3.641089121073934e-05, + "loss": 3.288380432128906, + "step": 4085 + }, + { + "epoch": 0.7857829010566763, + "grad_norm": 0.24223865568637848, + "learning_rate": 3.610020684483674e-05, + "loss": 3.2879989624023436, + "step": 4090 + }, + { + "epoch": 0.7867435158501441, + "grad_norm": 0.1990312784910202, + "learning_rate": 3.579067223418046e-05, + "loss": 3.2866798400878907, + "step": 4095 + }, + { + "epoch": 0.7877041306436119, + "grad_norm": 0.4838339388370514, + "learning_rate": 3.548229050335089e-05, + "loss": 3.290237808227539, + "step": 4100 + }, + { + "epoch": 0.7886647454370798, + "grad_norm": 0.27035945653915405, + "learning_rate": 3.517506476529045e-05, + "loss": 3.2898136138916017, + "step": 4105 + }, + { + "epoch": 0.7896253602305475, + "grad_norm": 0.2790147066116333, + "learning_rate": 3.486899812127264e-05, + "loss": 3.2898269653320313, + "step": 4110 + }, + { + "epoch": 0.7905859750240154, + "grad_norm": 0.2595069110393524, + "learning_rate": 3.456409366087054e-05, + "loss": 3.28946533203125, + "step": 4115 + }, + { + "epoch": 0.7915465898174832, + "grad_norm": 0.3146030306816101, + "learning_rate": 3.426035446192546e-05, + "loss": 3.2875953674316407, + "step": 4120 + }, + { + "epoch": 0.792507204610951, + "grad_norm": 0.28389808535575867, + "learning_rate": 3.395778359051634e-05, + "loss": 3.290841293334961, + "step": 4125 + }, + { + "epoch": 0.7934678194044188, + "grad_norm": 0.23182038962841034, + "learning_rate": 3.365638410092819e-05, + "loss": 3.289868927001953, + "step": 4130 + }, + { + "epoch": 0.7944284341978867, + "grad_norm": 0.2726062834262848, + "learning_rate": 3.3356159035621746e-05, + "loss": 3.287432098388672, + "step": 4135 + }, + { + "epoch": 0.7953890489913544, + "grad_norm": 0.21947945654392242, + "learning_rate": 3.3057111425202614e-05, + "loss": 3.286570358276367, + "step": 4140 + }, + { + "epoch": 0.7963496637848223, + "grad_norm": 0.28224751353263855, + "learning_rate": 3.275924428839043e-05, + "loss": 3.2863037109375, + "step": 4145 + }, + { + "epoch": 0.7973102785782901, + "grad_norm": 0.22240881621837616, + "learning_rate": 3.246256063198895e-05, + "loss": 3.2870025634765625, + "step": 4150 + }, + { + "epoch": 0.7982708933717579, + "grad_norm": 0.2880660891532898, + "learning_rate": 3.216706345085499e-05, + "loss": 3.2907535552978517, + "step": 4155 + }, + { + "epoch": 0.7992315081652257, + "grad_norm": 0.2571323812007904, + "learning_rate": 3.187275572786878e-05, + "loss": 3.283245849609375, + "step": 4160 + }, + { + "epoch": 0.8001921229586936, + "grad_norm": 0.3757617771625519, + "learning_rate": 3.15796404339036e-05, + "loss": 3.286081314086914, + "step": 4165 + }, + { + "epoch": 0.8011527377521613, + "grad_norm": 0.34448450803756714, + "learning_rate": 3.128772052779569e-05, + "loss": 3.2928192138671877, + "step": 4170 + }, + { + "epoch": 0.8021133525456292, + "grad_norm": 0.29771631956100464, + "learning_rate": 3.099699895631474e-05, + "loss": 3.2850051879882813, + "step": 4175 + }, + { + "epoch": 0.803073967339097, + "grad_norm": 0.19579406082630157, + "learning_rate": 3.0707478654133706e-05, + "loss": 3.2885662078857423, + "step": 4180 + }, + { + "epoch": 0.8040345821325648, + "grad_norm": 0.21206071972846985, + "learning_rate": 3.041916254379949e-05, + "loss": 3.289264678955078, + "step": 4185 + }, + { + "epoch": 0.8049951969260326, + "grad_norm": 0.2306758463382721, + "learning_rate": 3.0132053535703342e-05, + "loss": 3.289895248413086, + "step": 4190 + }, + { + "epoch": 0.8059558117195005, + "grad_norm": 0.24324500560760498, + "learning_rate": 2.984615452805147e-05, + "loss": 3.289009857177734, + "step": 4195 + }, + { + "epoch": 0.8069164265129684, + "grad_norm": 0.24182389676570892, + "learning_rate": 2.9561468406835865e-05, + "loss": 3.2901374816894533, + "step": 4200 + }, + { + "epoch": 0.8078770413064361, + "grad_norm": 0.19510437548160553, + "learning_rate": 2.927799804580495e-05, + "loss": 3.288174057006836, + "step": 4205 + }, + { + "epoch": 0.808837656099904, + "grad_norm": 0.20101343095302582, + "learning_rate": 2.8995746306434853e-05, + "loss": 3.2845272064208983, + "step": 4210 + }, + { + "epoch": 0.8097982708933718, + "grad_norm": 0.219615176320076, + "learning_rate": 2.871471603790035e-05, + "loss": 3.288011932373047, + "step": 4215 + }, + { + "epoch": 0.8107588856868396, + "grad_norm": 0.2699833810329437, + "learning_rate": 2.8434910077046163e-05, + "loss": 3.2884559631347656, + "step": 4220 + }, + { + "epoch": 0.8117195004803074, + "grad_norm": 0.2704545855522156, + "learning_rate": 2.8156331248358295e-05, + "loss": 3.283679962158203, + "step": 4225 + }, + { + "epoch": 0.8126801152737753, + "grad_norm": 0.22715161740779877, + "learning_rate": 2.787898236393556e-05, + "loss": 3.285501480102539, + "step": 4230 + }, + { + "epoch": 0.813640730067243, + "grad_norm": 0.2033311128616333, + "learning_rate": 2.7602866223461044e-05, + "loss": 3.284902572631836, + "step": 4235 + }, + { + "epoch": 0.8146013448607109, + "grad_norm": 0.2283061444759369, + "learning_rate": 2.7327985614174143e-05, + "loss": 3.2900623321533202, + "step": 4240 + }, + { + "epoch": 0.8155619596541787, + "grad_norm": 0.21356871724128723, + "learning_rate": 2.7054343310842115e-05, + "loss": 3.283802032470703, + "step": 4245 + }, + { + "epoch": 0.8165225744476465, + "grad_norm": 0.21679487824440002, + "learning_rate": 2.6781942075732294e-05, + "loss": 3.284503936767578, + "step": 4250 + }, + { + "epoch": 0.8174831892411143, + "grad_norm": 0.2814090847969055, + "learning_rate": 2.65107846585841e-05, + "loss": 3.2889778137207033, + "step": 4255 + }, + { + "epoch": 0.8184438040345822, + "grad_norm": 0.20454630255699158, + "learning_rate": 2.624087379658123e-05, + "loss": 3.2832550048828124, + "step": 4260 + }, + { + "epoch": 0.8194044188280499, + "grad_norm": 0.25826799869537354, + "learning_rate": 2.5972212214324162e-05, + "loss": 3.2887802124023438, + "step": 4265 + }, + { + "epoch": 0.8203650336215178, + "grad_norm": 0.24519048631191254, + "learning_rate": 2.5704802623802595e-05, + "loss": 3.2866302490234376, + "step": 4270 + }, + { + "epoch": 0.8213256484149856, + "grad_norm": 0.1912376880645752, + "learning_rate": 2.5438647724368054e-05, + "loss": 3.2893089294433593, + "step": 4275 + }, + { + "epoch": 0.8222862632084534, + "grad_norm": 0.18562142550945282, + "learning_rate": 2.5173750202706666e-05, + "loss": 3.2848739624023438, + "step": 4280 + }, + { + "epoch": 0.8232468780019212, + "grad_norm": 0.21541374921798706, + "learning_rate": 2.491011273281189e-05, + "loss": 3.285577392578125, + "step": 4285 + }, + { + "epoch": 0.8242074927953891, + "grad_norm": 0.2195151448249817, + "learning_rate": 2.4647737975957954e-05, + "loss": 3.2902549743652343, + "step": 4290 + }, + { + "epoch": 0.8251681075888568, + "grad_norm": 0.2502996623516083, + "learning_rate": 2.4386628580672396e-05, + "loss": 3.2855270385742186, + "step": 4295 + }, + { + "epoch": 0.8261287223823247, + "grad_norm": 0.168674036860466, + "learning_rate": 2.4126787182709796e-05, + "loss": 3.2874530792236327, + "step": 4300 + }, + { + "epoch": 0.8270893371757925, + "grad_norm": 0.22533266246318817, + "learning_rate": 2.3868216405025002e-05, + "loss": 3.286944580078125, + "step": 4305 + }, + { + "epoch": 0.8280499519692603, + "grad_norm": 0.21998296678066254, + "learning_rate": 2.361091885774652e-05, + "loss": 3.286793518066406, + "step": 4310 + }, + { + "epoch": 0.8290105667627281, + "grad_norm": 0.2057517021894455, + "learning_rate": 2.3354897138150536e-05, + "loss": 3.2883201599121095, + "step": 4315 + }, + { + "epoch": 0.829971181556196, + "grad_norm": 0.2263990193605423, + "learning_rate": 2.3100153830634218e-05, + "loss": 3.2877071380615233, + "step": 4320 + }, + { + "epoch": 0.8309317963496637, + "grad_norm": 0.17990660667419434, + "learning_rate": 2.284669150669001e-05, + "loss": 3.2851654052734376, + "step": 4325 + }, + { + "epoch": 0.8318924111431316, + "grad_norm": 0.19730253517627716, + "learning_rate": 2.259451272487957e-05, + "loss": 3.2849620819091796, + "step": 4330 + }, + { + "epoch": 0.8328530259365994, + "grad_norm": 0.1865607649087906, + "learning_rate": 2.234362003080772e-05, + "loss": 3.2859230041503906, + "step": 4335 + }, + { + "epoch": 0.8338136407300673, + "grad_norm": 0.21946489810943604, + "learning_rate": 2.2094015957097215e-05, + "loss": 3.2830989837646483, + "step": 4340 + }, + { + "epoch": 0.834774255523535, + "grad_norm": 0.22765038907527924, + "learning_rate": 2.1845703023362647e-05, + "loss": 3.2854949951171877, + "step": 4345 + }, + { + "epoch": 0.8357348703170029, + "grad_norm": 0.1913105696439743, + "learning_rate": 2.159868373618544e-05, + "loss": 3.2875335693359373, + "step": 4350 + }, + { + "epoch": 0.8366954851104706, + "grad_norm": 0.21100343763828278, + "learning_rate": 2.13529605890883e-05, + "loss": 3.2848331451416017, + "step": 4355 + }, + { + "epoch": 0.8376560999039385, + "grad_norm": 0.20388315618038177, + "learning_rate": 2.110853606251004e-05, + "loss": 3.283430480957031, + "step": 4360 + }, + { + "epoch": 0.8386167146974063, + "grad_norm": 0.18785250186920166, + "learning_rate": 2.0865412623780858e-05, + "loss": 3.2853065490722657, + "step": 4365 + }, + { + "epoch": 0.8395773294908742, + "grad_norm": 0.19692327082157135, + "learning_rate": 2.0623592727096916e-05, + "loss": 3.2828216552734375, + "step": 4370 + }, + { + "epoch": 0.840537944284342, + "grad_norm": 0.2322104126214981, + "learning_rate": 2.0383078813496e-05, + "loss": 3.2829490661621095, + "step": 4375 + }, + { + "epoch": 0.8414985590778098, + "grad_norm": 0.2523021697998047, + "learning_rate": 2.014387331083268e-05, + "loss": 3.2841728210449217, + "step": 4380 + }, + { + "epoch": 0.8424591738712777, + "grad_norm": 0.2998299300670624, + "learning_rate": 1.990597863375389e-05, + "loss": 3.2823081970214845, + "step": 4385 + }, + { + "epoch": 0.8434197886647454, + "grad_norm": 0.22866974771022797, + "learning_rate": 1.966939718367444e-05, + "loss": 3.2830490112304687, + "step": 4390 + }, + { + "epoch": 0.8443804034582133, + "grad_norm": 0.2374458909034729, + "learning_rate": 1.9434131348752842e-05, + "loss": 3.2861083984375, + "step": 4395 + }, + { + "epoch": 0.8453410182516811, + "grad_norm": 0.271164208650589, + "learning_rate": 1.920018350386725e-05, + "loss": 3.2853363037109373, + "step": 4400 + }, + { + "epoch": 0.8463016330451489, + "grad_norm": 0.26674726605415344, + "learning_rate": 1.8967556010591423e-05, + "loss": 3.2835784912109376, + "step": 4405 + }, + { + "epoch": 0.8472622478386167, + "grad_norm": 0.2277594804763794, + "learning_rate": 1.873625121717089e-05, + "loss": 3.287038040161133, + "step": 4410 + }, + { + "epoch": 0.8482228626320846, + "grad_norm": 0.21174593269824982, + "learning_rate": 1.850627145849926e-05, + "loss": 3.281999206542969, + "step": 4415 + }, + { + "epoch": 0.8491834774255523, + "grad_norm": 0.19385212659835815, + "learning_rate": 1.8277619056094684e-05, + "loss": 3.283008575439453, + "step": 4420 + }, + { + "epoch": 0.8501440922190202, + "grad_norm": 0.22816501557826996, + "learning_rate": 1.805029631807632e-05, + "loss": 3.282820129394531, + "step": 4425 + }, + { + "epoch": 0.851104707012488, + "grad_norm": 0.20349720120429993, + "learning_rate": 1.7824305539141165e-05, + "loss": 3.2845672607421874, + "step": 4430 + }, + { + "epoch": 0.8520653218059558, + "grad_norm": 0.17361263930797577, + "learning_rate": 1.7599649000540828e-05, + "loss": 3.284697341918945, + "step": 4435 + }, + { + "epoch": 0.8530259365994236, + "grad_norm": 0.19036982953548431, + "learning_rate": 1.7376328970058382e-05, + "loss": 3.2811737060546875, + "step": 4440 + }, + { + "epoch": 0.8539865513928915, + "grad_norm": 0.18866802752017975, + "learning_rate": 1.715434770198582e-05, + "loss": 3.2872802734375, + "step": 4445 + }, + { + "epoch": 0.8549471661863592, + "grad_norm": 0.17417368292808533, + "learning_rate": 1.6933707437100852e-05, + "loss": 3.2829563140869142, + "step": 4450 + }, + { + "epoch": 0.8559077809798271, + "grad_norm": 0.1611868292093277, + "learning_rate": 1.67144104026446e-05, + "loss": 3.281053161621094, + "step": 4455 + }, + { + "epoch": 0.8568683957732949, + "grad_norm": 0.18905870616436005, + "learning_rate": 1.6496458812299073e-05, + "loss": 3.281689453125, + "step": 4460 + }, + { + "epoch": 0.8578290105667628, + "grad_norm": 0.1868743747472763, + "learning_rate": 1.6279854866164586e-05, + "loss": 3.2842777252197264, + "step": 4465 + }, + { + "epoch": 0.8587896253602305, + "grad_norm": 0.16863790154457092, + "learning_rate": 1.6064600750737995e-05, + "loss": 3.2836139678955076, + "step": 4470 + }, + { + "epoch": 0.8597502401536984, + "grad_norm": 0.15220773220062256, + "learning_rate": 1.5850698638890093e-05, + "loss": 3.28204460144043, + "step": 4475 + }, + { + "epoch": 0.8607108549471661, + "grad_norm": 0.16620075702667236, + "learning_rate": 1.563815068984418e-05, + "loss": 3.2853729248046877, + "step": 4480 + }, + { + "epoch": 0.861671469740634, + "grad_norm": 0.14935848116874695, + "learning_rate": 1.54269590491539e-05, + "loss": 3.2866542816162108, + "step": 4485 + }, + { + "epoch": 0.8626320845341018, + "grad_norm": 0.1583867073059082, + "learning_rate": 1.521712584868166e-05, + "loss": 3.283392333984375, + "step": 4490 + }, + { + "epoch": 0.8635926993275697, + "grad_norm": 0.1681807041168213, + "learning_rate": 1.5008653206577376e-05, + "loss": 3.281229782104492, + "step": 4495 + }, + { + "epoch": 0.8645533141210374, + "grad_norm": 0.1857963651418686, + "learning_rate": 1.4801543227256685e-05, + "loss": 3.2861068725585936, + "step": 4500 + }, + { + "epoch": 0.8655139289145053, + "grad_norm": 0.16560649871826172, + "learning_rate": 1.4595798001379965e-05, + "loss": 3.2863983154296874, + "step": 4505 + }, + { + "epoch": 0.866474543707973, + "grad_norm": 0.17330175638198853, + "learning_rate": 1.4391419605831218e-05, + "loss": 3.284455490112305, + "step": 4510 + }, + { + "epoch": 0.8674351585014409, + "grad_norm": 0.15583863854408264, + "learning_rate": 1.4188410103696957e-05, + "loss": 3.283338165283203, + "step": 4515 + }, + { + "epoch": 0.8683957732949087, + "grad_norm": 0.1486099660396576, + "learning_rate": 1.398677154424559e-05, + "loss": 3.282981109619141, + "step": 4520 + }, + { + "epoch": 0.8693563880883766, + "grad_norm": 0.16416342556476593, + "learning_rate": 1.3786505962906475e-05, + "loss": 3.281314468383789, + "step": 4525 + }, + { + "epoch": 0.8703170028818443, + "grad_norm": 0.17155486345291138, + "learning_rate": 1.3587615381249622e-05, + "loss": 3.2852855682373048, + "step": 4530 + }, + { + "epoch": 0.8712776176753122, + "grad_norm": 0.17230698466300964, + "learning_rate": 1.3390101806965165e-05, + "loss": 3.2855682373046875, + "step": 4535 + }, + { + "epoch": 0.8722382324687801, + "grad_norm": 0.16426202654838562, + "learning_rate": 1.3193967233843083e-05, + "loss": 3.284606170654297, + "step": 4540 + }, + { + "epoch": 0.8731988472622478, + "grad_norm": 0.18724007904529572, + "learning_rate": 1.2999213641753164e-05, + "loss": 3.281108093261719, + "step": 4545 + }, + { + "epoch": 0.8741594620557157, + "grad_norm": 0.171888530254364, + "learning_rate": 1.280584299662486e-05, + "loss": 3.2837615966796876, + "step": 4550 + }, + { + "epoch": 0.8751200768491835, + "grad_norm": 0.15725582838058472, + "learning_rate": 1.2613857250427622e-05, + "loss": 3.281300354003906, + "step": 4555 + }, + { + "epoch": 0.8760806916426513, + "grad_norm": 0.1605101078748703, + "learning_rate": 1.2423258341151088e-05, + "loss": 3.2830268859863283, + "step": 4560 + }, + { + "epoch": 0.8770413064361191, + "grad_norm": 0.16305667161941528, + "learning_rate": 1.2234048192785539e-05, + "loss": 3.28362922668457, + "step": 4565 + }, + { + "epoch": 0.878001921229587, + "grad_norm": 0.1558937281370163, + "learning_rate": 1.2046228715302509e-05, + "loss": 3.28128662109375, + "step": 4570 + }, + { + "epoch": 0.8789625360230547, + "grad_norm": 0.14114581048488617, + "learning_rate": 1.1859801804635471e-05, + "loss": 3.2840221405029295, + "step": 4575 + }, + { + "epoch": 0.8799231508165226, + "grad_norm": 0.14987899363040924, + "learning_rate": 1.167476934266065e-05, + "loss": 3.2838886260986326, + "step": 4580 + }, + { + "epoch": 0.8808837656099904, + "grad_norm": 0.15901191532611847, + "learning_rate": 1.1491133197178177e-05, + "loss": 3.2861114501953126, + "step": 4585 + }, + { + "epoch": 0.8818443804034583, + "grad_norm": 0.1510273516178131, + "learning_rate": 1.1308895221893088e-05, + "loss": 3.2844474792480467, + "step": 4590 + }, + { + "epoch": 0.882804995196926, + "grad_norm": 0.14691267907619476, + "learning_rate": 1.1128057256396684e-05, + "loss": 3.2848670959472654, + "step": 4595 + }, + { + "epoch": 0.8837656099903939, + "grad_norm": 0.16120979189872742, + "learning_rate": 1.0948621126147978e-05, + "loss": 3.2807106018066405, + "step": 4600 + }, + { + "epoch": 0.8847262247838616, + "grad_norm": 0.161734938621521, + "learning_rate": 1.0770588642455092e-05, + "loss": 3.279821014404297, + "step": 4605 + }, + { + "epoch": 0.8856868395773295, + "grad_norm": 0.17960962653160095, + "learning_rate": 1.0593961602457346e-05, + "loss": 3.2869110107421875, + "step": 4610 + }, + { + "epoch": 0.8866474543707973, + "grad_norm": 0.13539910316467285, + "learning_rate": 1.041874178910666e-05, + "loss": 3.283498001098633, + "step": 4615 + }, + { + "epoch": 0.8876080691642652, + "grad_norm": 0.147013857960701, + "learning_rate": 1.0244930971149918e-05, + "loss": 3.28480224609375, + "step": 4620 + }, + { + "epoch": 0.8885686839577329, + "grad_norm": 0.16810062527656555, + "learning_rate": 1.0072530903110942e-05, + "loss": 3.284687805175781, + "step": 4625 + }, + { + "epoch": 0.8895292987512008, + "grad_norm": 0.16976973414421082, + "learning_rate": 9.901543325272753e-06, + "loss": 3.282553863525391, + "step": 4630 + }, + { + "epoch": 0.8904899135446686, + "grad_norm": 0.16181102395057678, + "learning_rate": 9.731969963660224e-06, + "loss": 3.2854312896728515, + "step": 4635 + }, + { + "epoch": 0.8914505283381364, + "grad_norm": 0.16076242923736572, + "learning_rate": 9.56381253002233e-06, + "loss": 3.2812950134277346, + "step": 4640 + }, + { + "epoch": 0.8924111431316042, + "grad_norm": 0.14188607037067413, + "learning_rate": 9.397072721815113e-06, + "loss": 3.284113311767578, + "step": 4645 + }, + { + "epoch": 0.8933717579250721, + "grad_norm": 0.16452723741531372, + "learning_rate": 9.231752222184496e-06, + "loss": 3.2799339294433594, + "step": 4650 + }, + { + "epoch": 0.8943323727185398, + "grad_norm": 0.13960616290569305, + "learning_rate": 9.067852699949197e-06, + "loss": 3.2871246337890625, + "step": 4655 + }, + { + "epoch": 0.8952929875120077, + "grad_norm": 0.1369549185037613, + "learning_rate": 8.905375809584053e-06, + "loss": 3.283050537109375, + "step": 4660 + }, + { + "epoch": 0.8962536023054755, + "grad_norm": 0.13010026514530182, + "learning_rate": 8.744323191203129e-06, + "loss": 3.2822250366210937, + "step": 4665 + }, + { + "epoch": 0.8972142170989433, + "grad_norm": 0.14469455182552338, + "learning_rate": 8.5846964705433e-06, + "loss": 3.2816314697265625, + "step": 4670 + }, + { + "epoch": 0.8981748318924111, + "grad_norm": 0.15112441778182983, + "learning_rate": 8.426497258947813e-06, + "loss": 3.279371643066406, + "step": 4675 + }, + { + "epoch": 0.899135446685879, + "grad_norm": 0.14177513122558594, + "learning_rate": 8.269727153349915e-06, + "loss": 3.2836181640625, + "step": 4680 + }, + { + "epoch": 0.9000960614793467, + "grad_norm": 0.1383771449327469, + "learning_rate": 8.114387736256966e-06, + "loss": 3.2833229064941407, + "step": 4685 + }, + { + "epoch": 0.9010566762728146, + "grad_norm": 0.1385938823223114, + "learning_rate": 7.960480575734162e-06, + "loss": 3.283488464355469, + "step": 4690 + }, + { + "epoch": 0.9020172910662824, + "grad_norm": 0.1253698468208313, + "learning_rate": 7.808007225388963e-06, + "loss": 3.282206726074219, + "step": 4695 + }, + { + "epoch": 0.9029779058597502, + "grad_norm": 0.14205971360206604, + "learning_rate": 7.656969224355285e-06, + "loss": 3.2821189880371096, + "step": 4700 + }, + { + "epoch": 0.9039385206532181, + "grad_norm": 0.13180512189865112, + "learning_rate": 7.507368097277994e-06, + "loss": 3.278270721435547, + "step": 4705 + }, + { + "epoch": 0.9048991354466859, + "grad_norm": 0.1405053585767746, + "learning_rate": 7.359205354297499e-06, + "loss": 3.279631423950195, + "step": 4710 + }, + { + "epoch": 0.9058597502401537, + "grad_norm": 0.1261477768421173, + "learning_rate": 7.21248249103451e-06, + "loss": 3.2792625427246094, + "step": 4715 + }, + { + "epoch": 0.9068203650336215, + "grad_norm": 0.1316026747226715, + "learning_rate": 7.067200988574983e-06, + "loss": 3.2806537628173826, + "step": 4720 + }, + { + "epoch": 0.9077809798270894, + "grad_norm": 0.1313539296388626, + "learning_rate": 6.923362313455094e-06, + "loss": 3.2771621704101563, + "step": 4725 + }, + { + "epoch": 0.9087415946205571, + "grad_norm": 0.1397256702184677, + "learning_rate": 6.780967917646518e-06, + "loss": 3.279193115234375, + "step": 4730 + }, + { + "epoch": 0.909702209414025, + "grad_norm": 0.13661499321460724, + "learning_rate": 6.640019238541727e-06, + "loss": 3.2864913940429688, + "step": 4735 + }, + { + "epoch": 0.9106628242074928, + "grad_norm": 0.13143934309482574, + "learning_rate": 6.5005176989394335e-06, + "loss": 3.2848949432373047, + "step": 4740 + }, + { + "epoch": 0.9116234390009607, + "grad_norm": 0.12046822905540466, + "learning_rate": 6.362464707030334e-06, + "loss": 3.283034896850586, + "step": 4745 + }, + { + "epoch": 0.9125840537944284, + "grad_norm": 0.12852244079113007, + "learning_rate": 6.225861656382825e-06, + "loss": 3.2837677001953125, + "step": 4750 + }, + { + "epoch": 0.9135446685878963, + "grad_norm": 0.13605345785617828, + "learning_rate": 6.090709925928938e-06, + "loss": 3.2784072875976564, + "step": 4755 + }, + { + "epoch": 0.914505283381364, + "grad_norm": 0.1370062530040741, + "learning_rate": 5.957010879950446e-06, + "loss": 3.279316711425781, + "step": 4760 + }, + { + "epoch": 0.9154658981748319, + "grad_norm": 0.14122441411018372, + "learning_rate": 5.824765868065101e-06, + "loss": 3.2786293029785156, + "step": 4765 + }, + { + "epoch": 0.9164265129682997, + "grad_norm": 0.1247406154870987, + "learning_rate": 5.69397622521291e-06, + "loss": 3.2822948455810548, + "step": 4770 + }, + { + "epoch": 0.9173871277617676, + "grad_norm": 0.13668857514858246, + "learning_rate": 5.564643271642799e-06, + "loss": 3.282684326171875, + "step": 4775 + }, + { + "epoch": 0.9183477425552353, + "grad_norm": 0.14488086104393005, + "learning_rate": 5.436768312899226e-06, + "loss": 3.2865325927734377, + "step": 4780 + }, + { + "epoch": 0.9193083573487032, + "grad_norm": 0.13214252889156342, + "learning_rate": 5.310352639808968e-06, + "loss": 3.2808204650878907, + "step": 4785 + }, + { + "epoch": 0.920268972142171, + "grad_norm": 0.12839952111244202, + "learning_rate": 5.185397528468155e-06, + "loss": 3.2781848907470703, + "step": 4790 + }, + { + "epoch": 0.9212295869356388, + "grad_norm": 0.12066492438316345, + "learning_rate": 5.061904240229309e-06, + "loss": 3.285063934326172, + "step": 4795 + }, + { + "epoch": 0.9221902017291066, + "grad_norm": 0.1341644525527954, + "learning_rate": 4.939874021688739e-06, + "loss": 3.2813148498535156, + "step": 4800 + }, + { + "epoch": 0.9231508165225745, + "grad_norm": 0.11596546322107315, + "learning_rate": 4.819308104673769e-06, + "loss": 3.281114196777344, + "step": 4805 + }, + { + "epoch": 0.9241114313160422, + "grad_norm": 0.1205928698182106, + "learning_rate": 4.700207706230513e-06, + "loss": 3.2828353881835937, + "step": 4810 + }, + { + "epoch": 0.9250720461095101, + "grad_norm": 0.1389501690864563, + "learning_rate": 4.582574028611435e-06, + "loss": 3.2821121215820312, + "step": 4815 + }, + { + "epoch": 0.9260326609029779, + "grad_norm": 0.11629052460193634, + "learning_rate": 4.466408259263243e-06, + "loss": 3.2819660186767576, + "step": 4820 + }, + { + "epoch": 0.9269932756964457, + "grad_norm": 0.13381557166576385, + "learning_rate": 4.351711570815014e-06, + "loss": 3.278516387939453, + "step": 4825 + }, + { + "epoch": 0.9279538904899135, + "grad_norm": 0.12584474682807922, + "learning_rate": 4.238485121066154e-06, + "loss": 3.281688690185547, + "step": 4830 + }, + { + "epoch": 0.9289145052833814, + "grad_norm": 0.12782976031303406, + "learning_rate": 4.126730052974908e-06, + "loss": 3.279840087890625, + "step": 4835 + }, + { + "epoch": 0.9298751200768491, + "grad_norm": 0.1391313374042511, + "learning_rate": 4.016447494646718e-06, + "loss": 3.2828041076660157, + "step": 4840 + }, + { + "epoch": 0.930835734870317, + "grad_norm": 0.11799421906471252, + "learning_rate": 3.907638559322817e-06, + "loss": 3.284708786010742, + "step": 4845 + }, + { + "epoch": 0.9317963496637848, + "grad_norm": 0.14186686277389526, + "learning_rate": 3.8003043453691207e-06, + "loss": 3.2787437438964844, + "step": 4850 + }, + { + "epoch": 0.9327569644572526, + "grad_norm": 0.11602967977523804, + "learning_rate": 3.69444593626495e-06, + "loss": 3.28221435546875, + "step": 4855 + }, + { + "epoch": 0.9337175792507204, + "grad_norm": 0.13371752202510834, + "learning_rate": 3.5900644005922465e-06, + "loss": 3.2848297119140626, + "step": 4860 + }, + { + "epoch": 0.9346781940441883, + "grad_norm": 0.12414630502462387, + "learning_rate": 3.48716079202474e-06, + "loss": 3.2800872802734373, + "step": 4865 + }, + { + "epoch": 0.9356388088376562, + "grad_norm": 0.11889371275901794, + "learning_rate": 3.385736149317264e-06, + "loss": 3.2823089599609374, + "step": 4870 + }, + { + "epoch": 0.9365994236311239, + "grad_norm": 0.1263454407453537, + "learning_rate": 3.2857914962953926e-06, + "loss": 3.2813446044921877, + "step": 4875 + }, + { + "epoch": 0.9375600384245918, + "grad_norm": 0.12067841738462448, + "learning_rate": 3.1873278418449e-06, + "loss": 3.2796573638916016, + "step": 4880 + }, + { + "epoch": 0.9385206532180596, + "grad_norm": 0.137226864695549, + "learning_rate": 3.090346179901837e-06, + "loss": 3.2818023681640627, + "step": 4885 + }, + { + "epoch": 0.9394812680115274, + "grad_norm": 0.11869537830352783, + "learning_rate": 2.99484748944227e-06, + "loss": 3.2809097290039064, + "step": 4890 + }, + { + "epoch": 0.9404418828049952, + "grad_norm": 0.11975245177745819, + "learning_rate": 2.9008327344724913e-06, + "loss": 3.282526397705078, + "step": 4895 + }, + { + "epoch": 0.9414024975984631, + "grad_norm": 0.11280205845832825, + "learning_rate": 2.808302864019374e-06, + "loss": 3.279458999633789, + "step": 4900 + }, + { + "epoch": 0.9423631123919308, + "grad_norm": 0.13222162425518036, + "learning_rate": 2.7172588121205983e-06, + "loss": 3.2817481994628905, + "step": 4905 + }, + { + "epoch": 0.9433237271853987, + "grad_norm": 0.09990071505308151, + "learning_rate": 2.62770149781541e-06, + "loss": 3.281599426269531, + "step": 4910 + }, + { + "epoch": 0.9442843419788665, + "grad_norm": 0.1150190606713295, + "learning_rate": 2.539631825135191e-06, + "loss": 3.2885650634765624, + "step": 4915 + }, + { + "epoch": 0.9452449567723343, + "grad_norm": 0.13486024737358093, + "learning_rate": 2.4530506830944876e-06, + "loss": 3.2844482421875, + "step": 4920 + }, + { + "epoch": 0.9462055715658021, + "grad_norm": 0.11893726140260696, + "learning_rate": 2.367958945681897e-06, + "loss": 3.28011474609375, + "step": 4925 + }, + { + "epoch": 0.94716618635927, + "grad_norm": 0.1119316965341568, + "learning_rate": 2.2843574718513114e-06, + "loss": 3.2823932647705076, + "step": 4930 + }, + { + "epoch": 0.9481268011527377, + "grad_norm": 0.09861624985933304, + "learning_rate": 2.2022471055132553e-06, + "loss": 3.281215286254883, + "step": 4935 + }, + { + "epoch": 0.9490874159462056, + "grad_norm": 0.12267674505710602, + "learning_rate": 2.1216286755263434e-06, + "loss": 3.281680679321289, + "step": 4940 + }, + { + "epoch": 0.9500480307396734, + "grad_norm": 0.11797958612442017, + "learning_rate": 2.042502995688905e-06, + "loss": 3.2819839477539063, + "step": 4945 + }, + { + "epoch": 0.9510086455331412, + "grad_norm": 0.10683488100767136, + "learning_rate": 1.964870864730789e-06, + "loss": 3.2817310333251952, + "step": 4950 + }, + { + "epoch": 0.951969260326609, + "grad_norm": 0.12260305136442184, + "learning_rate": 1.8887330663053536e-06, + "loss": 3.2814239501953124, + "step": 4955 + }, + { + "epoch": 0.9529298751200769, + "grad_norm": 0.12215402722358704, + "learning_rate": 1.8140903689814079e-06, + "loss": 3.283163070678711, + "step": 4960 + }, + { + "epoch": 0.9538904899135446, + "grad_norm": 0.1066419780254364, + "learning_rate": 1.740943526235583e-06, + "loss": 3.2810131072998048, + "step": 4965 + }, + { + "epoch": 0.9548511047070125, + "grad_norm": 0.10814165323972702, + "learning_rate": 1.6692932764447054e-06, + "loss": 3.281442642211914, + "step": 4970 + }, + { + "epoch": 0.9558117195004803, + "grad_norm": 0.10270337015390396, + "learning_rate": 1.5991403428783188e-06, + "loss": 3.283908462524414, + "step": 4975 + }, + { + "epoch": 0.9567723342939481, + "grad_norm": 0.10882619768381119, + "learning_rate": 1.5304854336913752e-06, + "loss": 3.2817230224609375, + "step": 4980 + }, + { + "epoch": 0.9577329490874159, + "grad_norm": 0.10903850197792053, + "learning_rate": 1.463329241917105e-06, + "loss": 3.282952880859375, + "step": 4985 + }, + { + "epoch": 0.9586935638808838, + "grad_norm": 0.11845772713422775, + "learning_rate": 1.397672445460024e-06, + "loss": 3.2793148040771483, + "step": 4990 + }, + { + "epoch": 0.9596541786743515, + "grad_norm": 0.10239589214324951, + "learning_rate": 1.333515707089089e-06, + "loss": 3.28063850402832, + "step": 4995 + }, + { + "epoch": 0.9606147934678194, + "grad_norm": 0.10683233290910721, + "learning_rate": 1.2708596744309685e-06, + "loss": 3.2818748474121096, + "step": 5000 + }, + { + "epoch": 0.9615754082612872, + "grad_norm": 0.1100321039557457, + "learning_rate": 1.209704979963616e-06, + "loss": 3.287214660644531, + "step": 5005 + }, + { + "epoch": 0.962536023054755, + "grad_norm": 0.10722316801548004, + "learning_rate": 1.1500522410096912e-06, + "loss": 3.2820858001708983, + "step": 5010 + }, + { + "epoch": 0.9634966378482228, + "grad_norm": 0.09480059891939163, + "learning_rate": 1.0919020597305649e-06, + "loss": 3.281991195678711, + "step": 5015 + }, + { + "epoch": 0.9644572526416907, + "grad_norm": 0.09626810252666473, + "learning_rate": 1.0352550231200407e-06, + "loss": 3.282924270629883, + "step": 5020 + }, + { + "epoch": 0.9654178674351584, + "grad_norm": 0.09960578382015228, + "learning_rate": 9.801117029985596e-07, + "loss": 3.280469512939453, + "step": 5025 + }, + { + "epoch": 0.9663784822286263, + "grad_norm": 0.10461635142564774, + "learning_rate": 9.264726560073377e-07, + "loss": 3.282293701171875, + "step": 5030 + }, + { + "epoch": 0.9673390970220941, + "grad_norm": 0.09554687887430191, + "learning_rate": 8.743384236028051e-07, + "loss": 3.2805267333984376, + "step": 5035 + }, + { + "epoch": 0.968299711815562, + "grad_norm": 0.09053874015808105, + "learning_rate": 8.237095320511589e-07, + "loss": 3.282093048095703, + "step": 5040 + }, + { + "epoch": 0.9692603266090298, + "grad_norm": 0.09056618809700012, + "learning_rate": 7.745864924229351e-07, + "loss": 3.2803783416748047, + "step": 5045 + }, + { + "epoch": 0.9702209414024976, + "grad_norm": 0.09100056439638138, + "learning_rate": 7.26969800588012e-07, + "loss": 3.279576873779297, + "step": 5050 + }, + { + "epoch": 0.9711815561959655, + "grad_norm": 0.10496609658002853, + "learning_rate": 6.808599372104817e-07, + "loss": 3.2778125762939454, + "step": 5055 + }, + { + "epoch": 0.9721421709894332, + "grad_norm": 0.09851839393377304, + "learning_rate": 6.362573677438199e-07, + "loss": 3.2856407165527344, + "step": 5060 + }, + { + "epoch": 0.9731027857829011, + "grad_norm": 0.1031249463558197, + "learning_rate": 5.931625424262731e-07, + "loss": 3.281835174560547, + "step": 5065 + }, + { + "epoch": 0.9740634005763689, + "grad_norm": 0.0869988277554512, + "learning_rate": 5.51575896276163e-07, + "loss": 3.2821426391601562, + "step": 5070 + }, + { + "epoch": 0.9750240153698367, + "grad_norm": 0.10024792701005936, + "learning_rate": 5.11497849087622e-07, + "loss": 3.278700256347656, + "step": 5075 + }, + { + "epoch": 0.9759846301633045, + "grad_norm": 0.08857379108667374, + "learning_rate": 4.7292880542634805e-07, + "loss": 3.2780426025390623, + "step": 5080 + }, + { + "epoch": 0.9769452449567724, + "grad_norm": 0.08493303507566452, + "learning_rate": 4.358691546254067e-07, + "loss": 3.279518890380859, + "step": 5085 + }, + { + "epoch": 0.9779058597502401, + "grad_norm": 0.09684525430202484, + "learning_rate": 4.0031927078145176e-07, + "loss": 3.281083679199219, + "step": 5090 + }, + { + "epoch": 0.978866474543708, + "grad_norm": 0.09520915895700455, + "learning_rate": 3.662795127508111e-07, + "loss": 3.2831382751464844, + "step": 5095 + }, + { + "epoch": 0.9798270893371758, + "grad_norm": 0.09378495812416077, + "learning_rate": 3.3375022414598994e-07, + "loss": 3.280741882324219, + "step": 5100 + }, + { + "epoch": 0.9807877041306436, + "grad_norm": 0.0972544252872467, + "learning_rate": 3.027317333321233e-07, + "loss": 3.282391357421875, + "step": 5105 + }, + { + "epoch": 0.9817483189241114, + "grad_norm": 0.09050152450799942, + "learning_rate": 2.7322435342364556e-07, + "loss": 3.2794769287109373, + "step": 5110 + }, + { + "epoch": 0.9827089337175793, + "grad_norm": 0.098331019282341, + "learning_rate": 2.452283822812262e-07, + "loss": 3.283687210083008, + "step": 5115 + }, + { + "epoch": 0.983669548511047, + "grad_norm": 0.09614613652229309, + "learning_rate": 2.1874410250863893e-07, + "loss": 3.2843692779541014, + "step": 5120 + }, + { + "epoch": 0.9846301633045149, + "grad_norm": 0.08628969639539719, + "learning_rate": 1.9377178145003059e-07, + "loss": 3.2782211303710938, + "step": 5125 + }, + { + "epoch": 0.9855907780979827, + "grad_norm": 0.09115748107433319, + "learning_rate": 1.7031167118708998e-07, + "loss": 3.282467269897461, + "step": 5130 + }, + { + "epoch": 0.9865513928914506, + "grad_norm": 0.08609933406114578, + "learning_rate": 1.4836400853666662e-07, + "loss": 3.280436706542969, + "step": 5135 + }, + { + "epoch": 0.9875120076849183, + "grad_norm": 0.08899597823619843, + "learning_rate": 1.2792901504820595e-07, + "loss": 3.2786022186279298, + "step": 5140 + }, + { + "epoch": 0.9884726224783862, + "grad_norm": 0.0838359072804451, + "learning_rate": 1.0900689700166776e-07, + "loss": 3.283920669555664, + "step": 5145 + }, + { + "epoch": 0.989433237271854, + "grad_norm": 0.08638288825750351, + "learning_rate": 9.159784540531124e-08, + "loss": 3.2787384033203124, + "step": 5150 + }, + { + "epoch": 0.9903938520653218, + "grad_norm": 0.08724058419466019, + "learning_rate": 7.57020359938798e-08, + "loss": 3.2797054290771483, + "step": 5155 + }, + { + "epoch": 0.9913544668587896, + "grad_norm": 0.08608481287956238, + "learning_rate": 6.131962922673595e-08, + "loss": 3.2824764251708984, + "step": 5160 + }, + { + "epoch": 0.9923150816522575, + "grad_norm": 0.08680781722068787, + "learning_rate": 4.845077028631239e-08, + "loss": 3.2829864501953123, + "step": 5165 + }, + { + "epoch": 0.9932756964457252, + "grad_norm": 0.08766987174749374, + "learning_rate": 3.709558907659671e-08, + "loss": 3.279438018798828, + "step": 5170 + }, + { + "epoch": 0.9942363112391931, + "grad_norm": 0.08375875651836395, + "learning_rate": 2.7254200221848988e-08, + "loss": 3.28199462890625, + "step": 5175 + }, + { + "epoch": 0.9951969260326609, + "grad_norm": 0.08243842422962189, + "learning_rate": 1.8926703065436087e-08, + "loss": 3.281136322021484, + "step": 5180 + }, + { + "epoch": 0.9961575408261287, + "grad_norm": 0.08867152035236359, + "learning_rate": 1.2113181668815808e-08, + "loss": 3.2808319091796876, + "step": 5185 + }, + { + "epoch": 0.9971181556195965, + "grad_norm": 0.08233336359262466, + "learning_rate": 6.813704810704201e-09, + "loss": 3.28259162902832, + "step": 5190 + }, + { + "epoch": 0.9980787704130644, + "grad_norm": 0.08597059547901154, + "learning_rate": 3.028325986392799e-09, + "loss": 3.281261444091797, + "step": 5195 + }, + { + "epoch": 0.9990393852065321, + "grad_norm": 0.08740798383951187, + "learning_rate": 7.570834071823905e-10, + "loss": 3.2809608459472654, + "step": 5200 + }, + { + "epoch": 1.0, + "grad_norm": 0.08548780530691147, + "learning_rate": 0.0, + "loss": 3.2842376708984373, + "step": 5205 + } + ], + "logging_steps": 5, + "max_steps": 5205, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 1000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 9.770678188442963e+19, + "train_batch_size": 64, + "trial_name": null, + "trial_params": null +}