{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 1004, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00099601593625498, "grad_norm": 0.4463295638561249, "learning_rate": 9.900990099009901e-08, "loss": 2.6197, "step": 1 }, { "epoch": 0.00199203187250996, "grad_norm": 0.8444207906723022, "learning_rate": 1.9801980198019803e-07, "loss": 2.7588, "step": 2 }, { "epoch": 0.00298804780876494, "grad_norm": 0.6591606140136719, "learning_rate": 2.9702970297029703e-07, "loss": 2.9263, "step": 3 }, { "epoch": 0.00398406374501992, "grad_norm": 0.5656299591064453, "learning_rate": 3.9603960396039606e-07, "loss": 2.4296, "step": 4 }, { "epoch": 0.0049800796812749, "grad_norm": 0.5051721930503845, "learning_rate": 4.950495049504951e-07, "loss": 2.3939, "step": 5 }, { "epoch": 0.00597609561752988, "grad_norm": 0.7762399911880493, "learning_rate": 5.940594059405941e-07, "loss": 2.6638, "step": 6 }, { "epoch": 0.0069721115537848604, "grad_norm": 0.5301679968833923, "learning_rate": 6.930693069306931e-07, "loss": 2.2199, "step": 7 }, { "epoch": 0.00796812749003984, "grad_norm": 0.6617525219917297, "learning_rate": 7.920792079207921e-07, "loss": 2.8019, "step": 8 }, { "epoch": 0.008964143426294821, "grad_norm": 0.7944237589836121, "learning_rate": 8.910891089108911e-07, "loss": 2.7146, "step": 9 }, { "epoch": 0.0099601593625498, "grad_norm": 0.6918312907218933, "learning_rate": 9.900990099009902e-07, "loss": 2.6128, "step": 10 }, { "epoch": 0.010956175298804782, "grad_norm": 0.550072193145752, "learning_rate": 1.0891089108910893e-06, "loss": 2.5252, "step": 11 }, { "epoch": 0.01195219123505976, "grad_norm": 0.4420550763607025, "learning_rate": 1.1881188118811881e-06, "loss": 2.6964, "step": 12 }, { "epoch": 0.012948207171314742, "grad_norm": 0.4683515429496765, "learning_rate": 1.2871287128712872e-06, "loss": 2.6433, "step": 13 }, { "epoch": 0.013944223107569721, "grad_norm": 0.5689812898635864, "learning_rate": 1.3861386138613863e-06, "loss": 2.3309, "step": 14 }, { "epoch": 0.014940239043824702, "grad_norm": 0.5711223483085632, "learning_rate": 1.4851485148514852e-06, "loss": 2.4396, "step": 15 }, { "epoch": 0.01593625498007968, "grad_norm": 0.4562544822692871, "learning_rate": 1.5841584158415842e-06, "loss": 2.234, "step": 16 }, { "epoch": 0.01693227091633466, "grad_norm": 0.33882570266723633, "learning_rate": 1.6831683168316833e-06, "loss": 2.5468, "step": 17 }, { "epoch": 0.017928286852589643, "grad_norm": 0.46446338295936584, "learning_rate": 1.7821782178217822e-06, "loss": 2.6143, "step": 18 }, { "epoch": 0.018924302788844622, "grad_norm": 0.625619649887085, "learning_rate": 1.8811881188118813e-06, "loss": 2.5565, "step": 19 }, { "epoch": 0.0199203187250996, "grad_norm": 0.5139931440353394, "learning_rate": 1.9801980198019803e-06, "loss": 2.8371, "step": 20 }, { "epoch": 0.02091633466135458, "grad_norm": 0.45826011896133423, "learning_rate": 2.0792079207920794e-06, "loss": 2.5883, "step": 21 }, { "epoch": 0.021912350597609563, "grad_norm": 0.5945838093757629, "learning_rate": 2.1782178217821785e-06, "loss": 2.6956, "step": 22 }, { "epoch": 0.022908366533864542, "grad_norm": 0.6705940961837769, "learning_rate": 2.2772277227722776e-06, "loss": 3.1889, "step": 23 }, { "epoch": 0.02390438247011952, "grad_norm": 0.489014208316803, "learning_rate": 2.3762376237623762e-06, "loss": 2.646, "step": 24 }, { "epoch": 0.0249003984063745, "grad_norm": 0.5540168285369873, "learning_rate": 2.4752475247524753e-06, "loss": 2.6909, "step": 25 }, { "epoch": 0.025896414342629483, "grad_norm": 0.34993091225624084, "learning_rate": 2.5742574257425744e-06, "loss": 2.5703, "step": 26 }, { "epoch": 0.026892430278884463, "grad_norm": 0.6204649209976196, "learning_rate": 2.6732673267326735e-06, "loss": 2.6973, "step": 27 }, { "epoch": 0.027888446215139442, "grad_norm": 0.6948006749153137, "learning_rate": 2.7722772277227726e-06, "loss": 2.8826, "step": 28 }, { "epoch": 0.02888446215139442, "grad_norm": 0.4865665137767792, "learning_rate": 2.8712871287128712e-06, "loss": 2.6566, "step": 29 }, { "epoch": 0.029880478087649404, "grad_norm": 0.7654755711555481, "learning_rate": 2.9702970297029703e-06, "loss": 2.9627, "step": 30 }, { "epoch": 0.030876494023904383, "grad_norm": 0.636715829372406, "learning_rate": 3.0693069306930694e-06, "loss": 2.3846, "step": 31 }, { "epoch": 0.03187250996015936, "grad_norm": 0.3698335289955139, "learning_rate": 3.1683168316831685e-06, "loss": 2.6724, "step": 32 }, { "epoch": 0.03286852589641434, "grad_norm": 0.7592146396636963, "learning_rate": 3.2673267326732676e-06, "loss": 2.9239, "step": 33 }, { "epoch": 0.03386454183266932, "grad_norm": 0.7194887399673462, "learning_rate": 3.3663366336633666e-06, "loss": 2.4727, "step": 34 }, { "epoch": 0.0348605577689243, "grad_norm": 0.4150688350200653, "learning_rate": 3.4653465346534653e-06, "loss": 2.6407, "step": 35 }, { "epoch": 0.035856573705179286, "grad_norm": 0.4450097382068634, "learning_rate": 3.5643564356435644e-06, "loss": 2.6709, "step": 36 }, { "epoch": 0.036852589641434265, "grad_norm": 0.4754781424999237, "learning_rate": 3.6633663366336635e-06, "loss": 2.6848, "step": 37 }, { "epoch": 0.037848605577689244, "grad_norm": 0.41208407282829285, "learning_rate": 3.7623762376237625e-06, "loss": 2.5176, "step": 38 }, { "epoch": 0.03884462151394422, "grad_norm": 0.41328731179237366, "learning_rate": 3.861386138613862e-06, "loss": 2.3275, "step": 39 }, { "epoch": 0.0398406374501992, "grad_norm": 0.5368106365203857, "learning_rate": 3.960396039603961e-06, "loss": 2.4877, "step": 40 }, { "epoch": 0.04083665338645418, "grad_norm": 0.37100547552108765, "learning_rate": 4.05940594059406e-06, "loss": 2.5933, "step": 41 }, { "epoch": 0.04183266932270916, "grad_norm": 0.4816776216030121, "learning_rate": 4.158415841584159e-06, "loss": 2.844, "step": 42 }, { "epoch": 0.04282868525896414, "grad_norm": 0.4209342896938324, "learning_rate": 4.2574257425742575e-06, "loss": 2.5422, "step": 43 }, { "epoch": 0.043824701195219126, "grad_norm": 0.6714078783988953, "learning_rate": 4.356435643564357e-06, "loss": 2.7081, "step": 44 }, { "epoch": 0.044820717131474105, "grad_norm": 0.26568883657455444, "learning_rate": 4.455445544554456e-06, "loss": 2.9217, "step": 45 }, { "epoch": 0.045816733067729085, "grad_norm": 0.37946802377700806, "learning_rate": 4.554455445544555e-06, "loss": 2.4118, "step": 46 }, { "epoch": 0.046812749003984064, "grad_norm": 0.6484291553497314, "learning_rate": 4.653465346534654e-06, "loss": 2.547, "step": 47 }, { "epoch": 0.04780876494023904, "grad_norm": 0.29198533296585083, "learning_rate": 4.7524752475247525e-06, "loss": 2.519, "step": 48 }, { "epoch": 0.04880478087649402, "grad_norm": 0.7583147883415222, "learning_rate": 4.851485148514852e-06, "loss": 2.9767, "step": 49 }, { "epoch": 0.049800796812749, "grad_norm": 0.5130609273910522, "learning_rate": 4.950495049504951e-06, "loss": 2.7004, "step": 50 }, { "epoch": 0.05079681274900399, "grad_norm": 0.31620916724205017, "learning_rate": 5.04950495049505e-06, "loss": 2.4462, "step": 51 }, { "epoch": 0.05179282868525897, "grad_norm": 0.8395189046859741, "learning_rate": 5.148514851485149e-06, "loss": 2.5588, "step": 52 }, { "epoch": 0.052788844621513946, "grad_norm": 0.387138307094574, "learning_rate": 5.247524752475248e-06, "loss": 2.1448, "step": 53 }, { "epoch": 0.053784860557768925, "grad_norm": 1.2158163785934448, "learning_rate": 5.346534653465347e-06, "loss": 2.3136, "step": 54 }, { "epoch": 0.054780876494023904, "grad_norm": 0.2387009710073471, "learning_rate": 5.4455445544554465e-06, "loss": 2.1985, "step": 55 }, { "epoch": 0.055776892430278883, "grad_norm": 0.3181290924549103, "learning_rate": 5.544554455445545e-06, "loss": 2.3419, "step": 56 }, { "epoch": 0.05677290836653386, "grad_norm": 0.37027299404144287, "learning_rate": 5.643564356435644e-06, "loss": 2.6245, "step": 57 }, { "epoch": 0.05776892430278884, "grad_norm": 1.2025309801101685, "learning_rate": 5.7425742574257425e-06, "loss": 3.9658, "step": 58 }, { "epoch": 0.05876494023904383, "grad_norm": 0.2420024573802948, "learning_rate": 5.841584158415842e-06, "loss": 2.6491, "step": 59 }, { "epoch": 0.05976095617529881, "grad_norm": 0.5019764304161072, "learning_rate": 5.940594059405941e-06, "loss": 2.5349, "step": 60 }, { "epoch": 0.060756972111553786, "grad_norm": 0.42150792479515076, "learning_rate": 6.03960396039604e-06, "loss": 3.0192, "step": 61 }, { "epoch": 0.061752988047808766, "grad_norm": 0.21461670100688934, "learning_rate": 6.138613861386139e-06, "loss": 2.4131, "step": 62 }, { "epoch": 0.06274900398406374, "grad_norm": 0.27715393900871277, "learning_rate": 6.237623762376238e-06, "loss": 2.0699, "step": 63 }, { "epoch": 0.06374501992031872, "grad_norm": 0.23175019025802612, "learning_rate": 6.336633663366337e-06, "loss": 2.454, "step": 64 }, { "epoch": 0.0647410358565737, "grad_norm": 1.4909917116165161, "learning_rate": 6.4356435643564364e-06, "loss": 2.394, "step": 65 }, { "epoch": 0.06573705179282868, "grad_norm": 0.25836002826690674, "learning_rate": 6.534653465346535e-06, "loss": 2.1991, "step": 66 }, { "epoch": 0.06673306772908366, "grad_norm": 0.24367666244506836, "learning_rate": 6.633663366336635e-06, "loss": 2.1038, "step": 67 }, { "epoch": 0.06772908366533864, "grad_norm": 0.2366018295288086, "learning_rate": 6.732673267326733e-06, "loss": 2.4102, "step": 68 }, { "epoch": 0.06872509960159362, "grad_norm": 0.2741665542125702, "learning_rate": 6.831683168316833e-06, "loss": 2.4163, "step": 69 }, { "epoch": 0.0697211155378486, "grad_norm": 1.1350017786026, "learning_rate": 6.930693069306931e-06, "loss": 3.1909, "step": 70 }, { "epoch": 0.07071713147410359, "grad_norm": 0.7466657161712646, "learning_rate": 7.02970297029703e-06, "loss": 3.0505, "step": 71 }, { "epoch": 0.07171314741035857, "grad_norm": 0.6016573309898376, "learning_rate": 7.128712871287129e-06, "loss": 2.523, "step": 72 }, { "epoch": 0.07270916334661355, "grad_norm": 0.29950985312461853, "learning_rate": 7.227722772277228e-06, "loss": 2.2338, "step": 73 }, { "epoch": 0.07370517928286853, "grad_norm": 0.3367365598678589, "learning_rate": 7.326732673267327e-06, "loss": 2.3668, "step": 74 }, { "epoch": 0.07470119521912351, "grad_norm": 0.23957543075084686, "learning_rate": 7.425742574257426e-06, "loss": 2.3929, "step": 75 }, { "epoch": 0.07569721115537849, "grad_norm": 0.2996574342250824, "learning_rate": 7.524752475247525e-06, "loss": 2.4929, "step": 76 }, { "epoch": 0.07669322709163347, "grad_norm": 1.1166422367095947, "learning_rate": 7.6237623762376246e-06, "loss": 2.2256, "step": 77 }, { "epoch": 0.07768924302788845, "grad_norm": 0.3733150362968445, "learning_rate": 7.722772277227724e-06, "loss": 2.4912, "step": 78 }, { "epoch": 0.07868525896414343, "grad_norm": 0.3624296486377716, "learning_rate": 7.821782178217822e-06, "loss": 2.7605, "step": 79 }, { "epoch": 0.0796812749003984, "grad_norm": 0.47846829891204834, "learning_rate": 7.920792079207921e-06, "loss": 2.7636, "step": 80 }, { "epoch": 0.08067729083665338, "grad_norm": 0.3782709836959839, "learning_rate": 8.019801980198021e-06, "loss": 2.4481, "step": 81 }, { "epoch": 0.08167330677290836, "grad_norm": 0.5004844665527344, "learning_rate": 8.11881188118812e-06, "loss": 2.4166, "step": 82 }, { "epoch": 0.08266932270916334, "grad_norm": 0.20830737054347992, "learning_rate": 8.217821782178218e-06, "loss": 2.4728, "step": 83 }, { "epoch": 0.08366533864541832, "grad_norm": 0.2479114830493927, "learning_rate": 8.316831683168318e-06, "loss": 2.2449, "step": 84 }, { "epoch": 0.0846613545816733, "grad_norm": 0.42911332845687866, "learning_rate": 8.415841584158416e-06, "loss": 2.2295, "step": 85 }, { "epoch": 0.08565737051792828, "grad_norm": 0.44220131635665894, "learning_rate": 8.514851485148515e-06, "loss": 2.4495, "step": 86 }, { "epoch": 0.08665338645418327, "grad_norm": 0.23947738111019135, "learning_rate": 8.613861386138615e-06, "loss": 2.1415, "step": 87 }, { "epoch": 0.08764940239043825, "grad_norm": 0.42801541090011597, "learning_rate": 8.712871287128714e-06, "loss": 2.3226, "step": 88 }, { "epoch": 0.08864541832669323, "grad_norm": 0.39098042249679565, "learning_rate": 8.811881188118812e-06, "loss": 2.3063, "step": 89 }, { "epoch": 0.08964143426294821, "grad_norm": 0.29923197627067566, "learning_rate": 8.910891089108911e-06, "loss": 2.6163, "step": 90 }, { "epoch": 0.09063745019920319, "grad_norm": 0.2684191167354584, "learning_rate": 9.009900990099011e-06, "loss": 2.6845, "step": 91 }, { "epoch": 0.09163346613545817, "grad_norm": 0.27097082138061523, "learning_rate": 9.10891089108911e-06, "loss": 2.598, "step": 92 }, { "epoch": 0.09262948207171315, "grad_norm": 0.2647894024848938, "learning_rate": 9.20792079207921e-06, "loss": 2.2659, "step": 93 }, { "epoch": 0.09362549800796813, "grad_norm": 0.38580745458602905, "learning_rate": 9.306930693069308e-06, "loss": 2.5818, "step": 94 }, { "epoch": 0.0946215139442231, "grad_norm": 0.23376502096652985, "learning_rate": 9.405940594059405e-06, "loss": 2.207, "step": 95 }, { "epoch": 0.09561752988047809, "grad_norm": 0.7030872702598572, "learning_rate": 9.504950495049505e-06, "loss": 2.1312, "step": 96 }, { "epoch": 0.09661354581673307, "grad_norm": 0.23579809069633484, "learning_rate": 9.603960396039604e-06, "loss": 2.0685, "step": 97 }, { "epoch": 0.09760956175298804, "grad_norm": 0.3901154100894928, "learning_rate": 9.702970297029704e-06, "loss": 2.662, "step": 98 }, { "epoch": 0.09860557768924302, "grad_norm": 0.2687411606311798, "learning_rate": 9.801980198019802e-06, "loss": 2.9062, "step": 99 }, { "epoch": 0.099601593625498, "grad_norm": 0.1916651427745819, "learning_rate": 9.900990099009901e-06, "loss": 2.1874, "step": 100 }, { "epoch": 0.10059760956175298, "grad_norm": 0.3915342688560486, "learning_rate": 1e-05, "loss": 2.4196, "step": 101 }, { "epoch": 0.10159362549800798, "grad_norm": 0.4256736636161804, "learning_rate": 9.999969740355927e-06, "loss": 2.3229, "step": 102 }, { "epoch": 0.10258964143426295, "grad_norm": 0.24454592168331146, "learning_rate": 9.999878961789962e-06, "loss": 2.4725, "step": 103 }, { "epoch": 0.10358565737051793, "grad_norm": 0.4549209475517273, "learning_rate": 9.999727665400876e-06, "loss": 2.6802, "step": 104 }, { "epoch": 0.10458167330677291, "grad_norm": 0.27968448400497437, "learning_rate": 9.999515853019941e-06, "loss": 2.154, "step": 105 }, { "epoch": 0.10557768924302789, "grad_norm": 0.30781856179237366, "learning_rate": 9.999243527210904e-06, "loss": 2.1358, "step": 106 }, { "epoch": 0.10657370517928287, "grad_norm": 0.22190262377262115, "learning_rate": 9.998910691269957e-06, "loss": 2.1951, "step": 107 }, { "epoch": 0.10756972111553785, "grad_norm": 0.36984801292419434, "learning_rate": 9.998517349225698e-06, "loss": 2.2283, "step": 108 }, { "epoch": 0.10856573705179283, "grad_norm": 0.24837668240070343, "learning_rate": 9.998063505839084e-06, "loss": 2.4556, "step": 109 }, { "epoch": 0.10956175298804781, "grad_norm": 0.17527616024017334, "learning_rate": 9.99754916660337e-06, "loss": 2.2414, "step": 110 }, { "epoch": 0.11055776892430279, "grad_norm": 0.2681446969509125, "learning_rate": 9.996974337744047e-06, "loss": 2.6504, "step": 111 }, { "epoch": 0.11155378486055777, "grad_norm": 0.41997164487838745, "learning_rate": 9.99633902621876e-06, "loss": 2.4704, "step": 112 }, { "epoch": 0.11254980079681275, "grad_norm": 0.43319636583328247, "learning_rate": 9.995643239717228e-06, "loss": 2.4391, "step": 113 }, { "epoch": 0.11354581673306773, "grad_norm": 0.3344462811946869, "learning_rate": 9.994886986661155e-06, "loss": 2.4113, "step": 114 }, { "epoch": 0.1145418326693227, "grad_norm": 0.2086816132068634, "learning_rate": 9.994070276204115e-06, "loss": 2.1469, "step": 115 }, { "epoch": 0.11553784860557768, "grad_norm": 0.35499969124794006, "learning_rate": 9.993193118231463e-06, "loss": 2.6212, "step": 116 }, { "epoch": 0.11653386454183266, "grad_norm": 0.4640713036060333, "learning_rate": 9.992255523360187e-06, "loss": 2.5926, "step": 117 }, { "epoch": 0.11752988047808766, "grad_norm": 0.3302168548107147, "learning_rate": 9.991257502938805e-06, "loss": 2.0769, "step": 118 }, { "epoch": 0.11852589641434264, "grad_norm": 0.4918990135192871, "learning_rate": 9.990199069047216e-06, "loss": 2.4818, "step": 119 }, { "epoch": 0.11952191235059761, "grad_norm": 0.25973260402679443, "learning_rate": 9.989080234496548e-06, "loss": 2.3236, "step": 120 }, { "epoch": 0.1205179282868526, "grad_norm": 0.26280921697616577, "learning_rate": 9.98790101282902e-06, "loss": 2.4276, "step": 121 }, { "epoch": 0.12151394422310757, "grad_norm": 0.21018964052200317, "learning_rate": 9.986661418317759e-06, "loss": 2.3006, "step": 122 }, { "epoch": 0.12250996015936255, "grad_norm": 0.30688363313674927, "learning_rate": 9.985361465966644e-06, "loss": 2.3402, "step": 123 }, { "epoch": 0.12350597609561753, "grad_norm": 0.27469494938850403, "learning_rate": 9.984001171510112e-06, "loss": 2.7987, "step": 124 }, { "epoch": 0.12450199203187251, "grad_norm": 0.2825527489185333, "learning_rate": 9.982580551412972e-06, "loss": 2.4642, "step": 125 }, { "epoch": 0.1254980079681275, "grad_norm": 0.6109227538108826, "learning_rate": 9.98109962287021e-06, "loss": 2.6041, "step": 126 }, { "epoch": 0.12649402390438247, "grad_norm": 0.3983345031738281, "learning_rate": 9.979558403806773e-06, "loss": 2.4599, "step": 127 }, { "epoch": 0.12749003984063745, "grad_norm": 0.3466341495513916, "learning_rate": 9.977956912877356e-06, "loss": 2.1902, "step": 128 }, { "epoch": 0.12848605577689243, "grad_norm": 0.3762282729148865, "learning_rate": 9.97629516946618e-06, "loss": 2.2643, "step": 129 }, { "epoch": 0.1294820717131474, "grad_norm": 0.523991048336029, "learning_rate": 9.974573193686747e-06, "loss": 2.196, "step": 130 }, { "epoch": 0.13047808764940239, "grad_norm": 0.23254041373729706, "learning_rate": 9.97279100638161e-06, "loss": 2.4402, "step": 131 }, { "epoch": 0.13147410358565736, "grad_norm": 0.3137255311012268, "learning_rate": 9.970948629122108e-06, "loss": 2.5905, "step": 132 }, { "epoch": 0.13247011952191234, "grad_norm": 0.21106691658496857, "learning_rate": 9.969046084208116e-06, "loss": 2.3683, "step": 133 }, { "epoch": 0.13346613545816732, "grad_norm": 0.4183836877346039, "learning_rate": 9.967083394667763e-06, "loss": 2.0614, "step": 134 }, { "epoch": 0.1344621513944223, "grad_norm": 0.4468408226966858, "learning_rate": 9.965060584257165e-06, "loss": 2.4639, "step": 135 }, { "epoch": 0.13545816733067728, "grad_norm": 0.22207094728946686, "learning_rate": 9.962977677460132e-06, "loss": 2.2261, "step": 136 }, { "epoch": 0.13645418326693226, "grad_norm": 0.2465856373310089, "learning_rate": 9.960834699487873e-06, "loss": 2.2444, "step": 137 }, { "epoch": 0.13745019920318724, "grad_norm": 0.3648821711540222, "learning_rate": 9.958631676278686e-06, "loss": 2.6132, "step": 138 }, { "epoch": 0.13844621513944222, "grad_norm": 0.21419532597064972, "learning_rate": 9.956368634497648e-06, "loss": 2.4863, "step": 139 }, { "epoch": 0.1394422310756972, "grad_norm": 0.5358874797821045, "learning_rate": 9.9540456015363e-06, "loss": 2.401, "step": 140 }, { "epoch": 0.14043824701195218, "grad_norm": 0.25208160281181335, "learning_rate": 9.951662605512298e-06, "loss": 2.5901, "step": 141 }, { "epoch": 0.14143426294820718, "grad_norm": 0.2659305930137634, "learning_rate": 9.949219675269089e-06, "loss": 2.3229, "step": 142 }, { "epoch": 0.14243027888446216, "grad_norm": 0.5667638778686523, "learning_rate": 9.946716840375552e-06, "loss": 2.4998, "step": 143 }, { "epoch": 0.14342629482071714, "grad_norm": 0.315893292427063, "learning_rate": 9.944154131125643e-06, "loss": 2.4017, "step": 144 }, { "epoch": 0.14442231075697212, "grad_norm": 0.43832895159721375, "learning_rate": 9.941531578538032e-06, "loss": 2.473, "step": 145 }, { "epoch": 0.1454183266932271, "grad_norm": 0.2750052511692047, "learning_rate": 9.938849214355722e-06, "loss": 2.3208, "step": 146 }, { "epoch": 0.14641434262948208, "grad_norm": 0.26100143790245056, "learning_rate": 9.936107071045665e-06, "loss": 2.2151, "step": 147 }, { "epoch": 0.14741035856573706, "grad_norm": 0.3080121576786041, "learning_rate": 9.933305181798374e-06, "loss": 2.0868, "step": 148 }, { "epoch": 0.14840637450199204, "grad_norm": 0.24721984565258026, "learning_rate": 9.93044358052752e-06, "loss": 2.3312, "step": 149 }, { "epoch": 0.14940239043824702, "grad_norm": 0.5916289687156677, "learning_rate": 9.927522301869515e-06, "loss": 2.6542, "step": 150 }, { "epoch": 0.150398406374502, "grad_norm": 0.22790588438510895, "learning_rate": 9.924541381183099e-06, "loss": 2.3121, "step": 151 }, { "epoch": 0.15139442231075698, "grad_norm": 0.3201110064983368, "learning_rate": 9.921500854548916e-06, "loss": 2.8776, "step": 152 }, { "epoch": 0.15239043824701196, "grad_norm": 0.3063032627105713, "learning_rate": 9.918400758769063e-06, "loss": 2.3604, "step": 153 }, { "epoch": 0.15338645418326693, "grad_norm": 0.41891732811927795, "learning_rate": 9.915241131366657e-06, "loss": 2.1036, "step": 154 }, { "epoch": 0.15438247011952191, "grad_norm": 1.0909143686294556, "learning_rate": 9.912022010585385e-06, "loss": 2.5247, "step": 155 }, { "epoch": 0.1553784860557769, "grad_norm": 0.26673582196235657, "learning_rate": 9.90874343538902e-06, "loss": 2.159, "step": 156 }, { "epoch": 0.15637450199203187, "grad_norm": 0.3614170551300049, "learning_rate": 9.905405445460972e-06, "loss": 2.4383, "step": 157 }, { "epoch": 0.15737051792828685, "grad_norm": 0.5078898668289185, "learning_rate": 9.902008081203796e-06, "loss": 2.2543, "step": 158 }, { "epoch": 0.15836653386454183, "grad_norm": 0.3936934769153595, "learning_rate": 9.898551383738707e-06, "loss": 2.538, "step": 159 }, { "epoch": 0.1593625498007968, "grad_norm": 0.6516975164413452, "learning_rate": 9.895035394905073e-06, "loss": 2.5301, "step": 160 }, { "epoch": 0.1603585657370518, "grad_norm": 0.26518794894218445, "learning_rate": 9.89146015725993e-06, "loss": 2.2118, "step": 161 }, { "epoch": 0.16135458167330677, "grad_norm": 0.41366127133369446, "learning_rate": 9.887825714077439e-06, "loss": 2.4799, "step": 162 }, { "epoch": 0.16235059760956175, "grad_norm": 0.5400006771087646, "learning_rate": 9.884132109348386e-06, "loss": 2.5019, "step": 163 }, { "epoch": 0.16334661354581673, "grad_norm": 0.36508408188819885, "learning_rate": 9.880379387779637e-06, "loss": 2.7938, "step": 164 }, { "epoch": 0.1643426294820717, "grad_norm": 0.4240388572216034, "learning_rate": 9.876567594793597e-06, "loss": 2.5002, "step": 165 }, { "epoch": 0.16533864541832669, "grad_norm": 0.277864933013916, "learning_rate": 9.87269677652767e-06, "loss": 2.5436, "step": 166 }, { "epoch": 0.16633466135458166, "grad_norm": 0.25840163230895996, "learning_rate": 9.868766979833686e-06, "loss": 2.2811, "step": 167 }, { "epoch": 0.16733067729083664, "grad_norm": 0.32198214530944824, "learning_rate": 9.864778252277344e-06, "loss": 2.3215, "step": 168 }, { "epoch": 0.16832669322709162, "grad_norm": 0.613046407699585, "learning_rate": 9.86073064213764e-06, "loss": 2.8133, "step": 169 }, { "epoch": 0.1693227091633466, "grad_norm": 0.38470038771629333, "learning_rate": 9.856624198406262e-06, "loss": 2.4133, "step": 170 }, { "epoch": 0.17031872509960158, "grad_norm": 0.3675747811794281, "learning_rate": 9.852458970787027e-06, "loss": 2.0617, "step": 171 }, { "epoch": 0.17131474103585656, "grad_norm": 0.26074767112731934, "learning_rate": 9.848235009695255e-06, "loss": 2.132, "step": 172 }, { "epoch": 0.17231075697211157, "grad_norm": 0.5974801778793335, "learning_rate": 9.84395236625717e-06, "loss": 2.3888, "step": 173 }, { "epoch": 0.17330677290836655, "grad_norm": 0.2652048170566559, "learning_rate": 9.839611092309278e-06, "loss": 2.4468, "step": 174 }, { "epoch": 0.17430278884462153, "grad_norm": 0.6336271166801453, "learning_rate": 9.835211240397743e-06, "loss": 2.3256, "step": 175 }, { "epoch": 0.1752988047808765, "grad_norm": 0.3853505849838257, "learning_rate": 9.830752863777741e-06, "loss": 2.5527, "step": 176 }, { "epoch": 0.17629482071713148, "grad_norm": 0.25374558568000793, "learning_rate": 9.826236016412833e-06, "loss": 2.5593, "step": 177 }, { "epoch": 0.17729083665338646, "grad_norm": 0.29101264476776123, "learning_rate": 9.821660752974294e-06, "loss": 2.6399, "step": 178 }, { "epoch": 0.17828685258964144, "grad_norm": 0.7464101910591125, "learning_rate": 9.817027128840462e-06, "loss": 2.1674, "step": 179 }, { "epoch": 0.17928286852589642, "grad_norm": 0.28557366132736206, "learning_rate": 9.812335200096064e-06, "loss": 2.7127, "step": 180 }, { "epoch": 0.1802788844621514, "grad_norm": 0.5655897259712219, "learning_rate": 9.807585023531536e-06, "loss": 2.0397, "step": 181 }, { "epoch": 0.18127490039840638, "grad_norm": 0.2831386625766754, "learning_rate": 9.802776656642341e-06, "loss": 2.1947, "step": 182 }, { "epoch": 0.18227091633466136, "grad_norm": 0.30917420983314514, "learning_rate": 9.797910157628265e-06, "loss": 2.3951, "step": 183 }, { "epoch": 0.18326693227091634, "grad_norm": 0.3886703550815582, "learning_rate": 9.792985585392722e-06, "loss": 2.4107, "step": 184 }, { "epoch": 0.18426294820717132, "grad_norm": 0.51981121301651, "learning_rate": 9.78800299954203e-06, "loss": 2.3655, "step": 185 }, { "epoch": 0.1852589641434263, "grad_norm": 0.31090375781059265, "learning_rate": 9.782962460384701e-06, "loss": 1.984, "step": 186 }, { "epoch": 0.18625498007968128, "grad_norm": 0.4561314880847931, "learning_rate": 9.777864028930705e-06, "loss": 3.0161, "step": 187 }, { "epoch": 0.18725099601593626, "grad_norm": 0.3265978693962097, "learning_rate": 9.772707766890726e-06, "loss": 2.6738, "step": 188 }, { "epoch": 0.18824701195219123, "grad_norm": 0.5627899765968323, "learning_rate": 9.767493736675429e-06, "loss": 2.4544, "step": 189 }, { "epoch": 0.1892430278884462, "grad_norm": 0.3551636338233948, "learning_rate": 9.762222001394692e-06, "loss": 2.4696, "step": 190 }, { "epoch": 0.1902390438247012, "grad_norm": 0.27445298433303833, "learning_rate": 9.756892624856848e-06, "loss": 2.5626, "step": 191 }, { "epoch": 0.19123505976095617, "grad_norm": 0.4346907436847687, "learning_rate": 9.751505671567914e-06, "loss": 2.6588, "step": 192 }, { "epoch": 0.19223107569721115, "grad_norm": 1.7177170515060425, "learning_rate": 9.746061206730801e-06, "loss": 3.3538, "step": 193 }, { "epoch": 0.19322709163346613, "grad_norm": 0.294007807970047, "learning_rate": 9.740559296244543e-06, "loss": 2.7963, "step": 194 }, { "epoch": 0.1942231075697211, "grad_norm": 0.3322044014930725, "learning_rate": 9.735000006703475e-06, "loss": 2.1763, "step": 195 }, { "epoch": 0.1952191235059761, "grad_norm": 0.2852723002433777, "learning_rate": 9.72938340539645e-06, "loss": 2.2182, "step": 196 }, { "epoch": 0.19621513944223107, "grad_norm": 0.2600834369659424, "learning_rate": 9.723709560306009e-06, "loss": 2.7632, "step": 197 }, { "epoch": 0.19721115537848605, "grad_norm": 0.27677562832832336, "learning_rate": 9.717978540107566e-06, "loss": 2.3831, "step": 198 }, { "epoch": 0.19820717131474103, "grad_norm": 0.4312080144882202, "learning_rate": 9.712190414168573e-06, "loss": 2.4096, "step": 199 }, { "epoch": 0.199203187250996, "grad_norm": 0.7516922950744629, "learning_rate": 9.706345252547681e-06, "loss": 3.0072, "step": 200 }, { "epoch": 0.20019920318725098, "grad_norm": 0.47257497906684875, "learning_rate": 9.700443125993897e-06, "loss": 2.4537, "step": 201 }, { "epoch": 0.20119521912350596, "grad_norm": 0.5163850784301758, "learning_rate": 9.694484105945719e-06, "loss": 2.4488, "step": 202 }, { "epoch": 0.20219123505976094, "grad_norm": 0.2632780373096466, "learning_rate": 9.688468264530278e-06, "loss": 2.5477, "step": 203 }, { "epoch": 0.20318725099601595, "grad_norm": 1.0932762622833252, "learning_rate": 9.682395674562459e-06, "loss": 2.8381, "step": 204 }, { "epoch": 0.20418326693227093, "grad_norm": 0.568217396736145, "learning_rate": 9.676266409544031e-06, "loss": 2.2398, "step": 205 }, { "epoch": 0.2051792828685259, "grad_norm": 0.5864899754524231, "learning_rate": 9.670080543662742e-06, "loss": 2.5067, "step": 206 }, { "epoch": 0.2061752988047809, "grad_norm": 0.38742220401763916, "learning_rate": 9.663838151791431e-06, "loss": 2.3831, "step": 207 }, { "epoch": 0.20717131474103587, "grad_norm": 0.441034197807312, "learning_rate": 9.657539309487123e-06, "loss": 2.3785, "step": 208 }, { "epoch": 0.20816733067729085, "grad_norm": 1.6438182592391968, "learning_rate": 9.651184092990109e-06, "loss": 3.6952, "step": 209 }, { "epoch": 0.20916334661354583, "grad_norm": 0.35267430543899536, "learning_rate": 9.644772579223023e-06, "loss": 2.4354, "step": 210 }, { "epoch": 0.2101593625498008, "grad_norm": 0.27790936827659607, "learning_rate": 9.638304845789916e-06, "loss": 2.3407, "step": 211 }, { "epoch": 0.21115537848605578, "grad_norm": 0.4317843019962311, "learning_rate": 9.631780970975311e-06, "loss": 2.2805, "step": 212 }, { "epoch": 0.21215139442231076, "grad_norm": 0.35801681876182556, "learning_rate": 9.625201033743262e-06, "loss": 2.3219, "step": 213 }, { "epoch": 0.21314741035856574, "grad_norm": 0.3666556477546692, "learning_rate": 9.618565113736388e-06, "loss": 2.1962, "step": 214 }, { "epoch": 0.21414342629482072, "grad_norm": 0.36347630620002747, "learning_rate": 9.611873291274927e-06, "loss": 2.1945, "step": 215 }, { "epoch": 0.2151394422310757, "grad_norm": 0.47142109274864197, "learning_rate": 9.60512564735574e-06, "loss": 2.1907, "step": 216 }, { "epoch": 0.21613545816733068, "grad_norm": 0.3300761580467224, "learning_rate": 9.598322263651352e-06, "loss": 2.0638, "step": 217 }, { "epoch": 0.21713147410358566, "grad_norm": 0.3918429911136627, "learning_rate": 9.591463222508947e-06, "loss": 2.4349, "step": 218 }, { "epoch": 0.21812749003984064, "grad_norm": 0.3837280869483948, "learning_rate": 9.584548606949384e-06, "loss": 2.2359, "step": 219 }, { "epoch": 0.21912350597609562, "grad_norm": 0.47225990891456604, "learning_rate": 9.577578500666187e-06, "loss": 2.4696, "step": 220 }, { "epoch": 0.2201195219123506, "grad_norm": 0.3448033630847931, "learning_rate": 9.570552988024527e-06, "loss": 2.3639, "step": 221 }, { "epoch": 0.22111553784860558, "grad_norm": 0.39937150478363037, "learning_rate": 9.563472154060212e-06, "loss": 2.3513, "step": 222 }, { "epoch": 0.22211155378486055, "grad_norm": 0.3486849367618561, "learning_rate": 9.556336084478645e-06, "loss": 2.3674, "step": 223 }, { "epoch": 0.22310756972111553, "grad_norm": 0.4388813376426697, "learning_rate": 9.5491448656538e-06, "loss": 2.4748, "step": 224 }, { "epoch": 0.2241035856573705, "grad_norm": 0.4307428300380707, "learning_rate": 9.541898584627164e-06, "loss": 2.1206, "step": 225 }, { "epoch": 0.2250996015936255, "grad_norm": 0.5265683531761169, "learning_rate": 9.534597329106688e-06, "loss": 2.589, "step": 226 }, { "epoch": 0.22609561752988047, "grad_norm": 0.5943540930747986, "learning_rate": 9.527241187465735e-06, "loss": 2.8641, "step": 227 }, { "epoch": 0.22709163346613545, "grad_norm": 0.3554113209247589, "learning_rate": 9.519830248741991e-06, "loss": 2.0978, "step": 228 }, { "epoch": 0.22808764940239043, "grad_norm": 0.43764352798461914, "learning_rate": 9.512364602636405e-06, "loss": 2.2777, "step": 229 }, { "epoch": 0.2290836653386454, "grad_norm": 0.27372264862060547, "learning_rate": 9.504844339512096e-06, "loss": 2.3366, "step": 230 }, { "epoch": 0.2300796812749004, "grad_norm": 0.5419708490371704, "learning_rate": 9.497269550393257e-06, "loss": 2.5115, "step": 231 }, { "epoch": 0.23107569721115537, "grad_norm": 0.3294195234775543, "learning_rate": 9.489640326964058e-06, "loss": 2.3812, "step": 232 }, { "epoch": 0.23207171314741035, "grad_norm": 0.3676604926586151, "learning_rate": 9.481956761567531e-06, "loss": 2.1645, "step": 233 }, { "epoch": 0.23306772908366533, "grad_norm": 0.24499647319316864, "learning_rate": 9.47421894720446e-06, "loss": 2.436, "step": 234 }, { "epoch": 0.2340637450199203, "grad_norm": 0.4562065601348877, "learning_rate": 9.466426977532246e-06, "loss": 2.4614, "step": 235 }, { "epoch": 0.2350597609561753, "grad_norm": 0.4152824580669403, "learning_rate": 9.458580946863784e-06, "loss": 2.6406, "step": 236 }, { "epoch": 0.2360557768924303, "grad_norm": 0.2812240421772003, "learning_rate": 9.45068095016631e-06, "loss": 2.4129, "step": 237 }, { "epoch": 0.23705179282868527, "grad_norm": 0.37695300579071045, "learning_rate": 9.442727083060258e-06, "loss": 2.8288, "step": 238 }, { "epoch": 0.23804780876494025, "grad_norm": 0.29094114899635315, "learning_rate": 9.434719441818106e-06, "loss": 2.2392, "step": 239 }, { "epoch": 0.23904382470119523, "grad_norm": 0.6004308462142944, "learning_rate": 9.426658123363202e-06, "loss": 2.6978, "step": 240 }, { "epoch": 0.2400398406374502, "grad_norm": 0.824376106262207, "learning_rate": 9.418543225268598e-06, "loss": 2.8384, "step": 241 }, { "epoch": 0.2410358565737052, "grad_norm": 0.37361350655555725, "learning_rate": 9.410374845755862e-06, "loss": 2.7737, "step": 242 }, { "epoch": 0.24203187250996017, "grad_norm": 0.3311799168586731, "learning_rate": 9.402153083693898e-06, "loss": 2.6569, "step": 243 }, { "epoch": 0.24302788844621515, "grad_norm": 0.31674501299858093, "learning_rate": 9.393878038597748e-06, "loss": 2.4791, "step": 244 }, { "epoch": 0.24402390438247012, "grad_norm": 0.5512855052947998, "learning_rate": 9.385549810627374e-06, "loss": 2.0412, "step": 245 }, { "epoch": 0.2450199203187251, "grad_norm": 0.48202139139175415, "learning_rate": 9.377168500586465e-06, "loss": 2.4472, "step": 246 }, { "epoch": 0.24601593625498008, "grad_norm": 0.29134997725486755, "learning_rate": 9.3687342099212e-06, "loss": 2.1911, "step": 247 }, { "epoch": 0.24701195219123506, "grad_norm": 0.27262917160987854, "learning_rate": 9.36024704071904e-06, "loss": 2.3633, "step": 248 }, { "epoch": 0.24800796812749004, "grad_norm": 0.2618001699447632, "learning_rate": 9.351707095707465e-06, "loss": 2.3556, "step": 249 }, { "epoch": 0.24900398406374502, "grad_norm": 0.4861814081668854, "learning_rate": 9.343114478252758e-06, "loss": 2.2809, "step": 250 }, { "epoch": 0.25, "grad_norm": 0.3331791162490845, "learning_rate": 9.334469292358736e-06, "loss": 2.2747, "step": 251 }, { "epoch": 0.250996015936255, "grad_norm": 0.3599317669868469, "learning_rate": 9.3257716426655e-06, "loss": 2.3204, "step": 252 }, { "epoch": 0.25199203187250996, "grad_norm": 0.28026479482650757, "learning_rate": 9.317021634448162e-06, "loss": 2.5997, "step": 253 }, { "epoch": 0.25298804780876494, "grad_norm": 0.4968087077140808, "learning_rate": 9.308219373615574e-06, "loss": 2.4623, "step": 254 }, { "epoch": 0.2539840637450199, "grad_norm": 0.5899234414100647, "learning_rate": 9.299364966709051e-06, "loss": 2.4678, "step": 255 }, { "epoch": 0.2549800796812749, "grad_norm": 0.27741050720214844, "learning_rate": 9.290458520901072e-06, "loss": 2.4373, "step": 256 }, { "epoch": 0.2559760956175299, "grad_norm": 0.44141483306884766, "learning_rate": 9.28150014399399e-06, "loss": 2.3013, "step": 257 }, { "epoch": 0.25697211155378485, "grad_norm": 0.4108343720436096, "learning_rate": 9.272489944418724e-06, "loss": 2.6281, "step": 258 }, { "epoch": 0.25796812749003983, "grad_norm": 0.4309611916542053, "learning_rate": 9.263428031233444e-06, "loss": 2.6192, "step": 259 }, { "epoch": 0.2589641434262948, "grad_norm": 0.3191240727901459, "learning_rate": 9.25431451412226e-06, "loss": 2.3667, "step": 260 }, { "epoch": 0.2599601593625498, "grad_norm": 0.4311404824256897, "learning_rate": 9.245149503393884e-06, "loss": 2.5286, "step": 261 }, { "epoch": 0.26095617529880477, "grad_norm": 0.8753085136413574, "learning_rate": 9.235933109980302e-06, "loss": 2.6609, "step": 262 }, { "epoch": 0.26195219123505975, "grad_norm": 0.679023265838623, "learning_rate": 9.226665445435428e-06, "loss": 2.4715, "step": 263 }, { "epoch": 0.26294820717131473, "grad_norm": 0.4910929501056671, "learning_rate": 9.217346621933753e-06, "loss": 2.0939, "step": 264 }, { "epoch": 0.2639442231075697, "grad_norm": 1.3370636701583862, "learning_rate": 9.207976752268992e-06, "loss": 2.4367, "step": 265 }, { "epoch": 0.2649402390438247, "grad_norm": 0.4148995578289032, "learning_rate": 9.19855594985271e-06, "loss": 2.5403, "step": 266 }, { "epoch": 0.26593625498007967, "grad_norm": 0.5130553841590881, "learning_rate": 9.189084328712961e-06, "loss": 2.654, "step": 267 }, { "epoch": 0.26693227091633465, "grad_norm": 0.5000612139701843, "learning_rate": 9.179562003492898e-06, "loss": 2.2451, "step": 268 }, { "epoch": 0.2679282868525896, "grad_norm": 1.3651481866836548, "learning_rate": 9.16998908944939e-06, "loss": 2.3566, "step": 269 }, { "epoch": 0.2689243027888446, "grad_norm": 0.4688972532749176, "learning_rate": 9.160365702451625e-06, "loss": 2.4274, "step": 270 }, { "epoch": 0.2699203187250996, "grad_norm": 0.44729602336883545, "learning_rate": 9.150691958979712e-06, "loss": 2.3431, "step": 271 }, { "epoch": 0.27091633466135456, "grad_norm": 0.4126404821872711, "learning_rate": 9.14096797612326e-06, "loss": 2.4358, "step": 272 }, { "epoch": 0.27191235059760954, "grad_norm": 0.8661454319953918, "learning_rate": 9.131193871579975e-06, "loss": 2.6125, "step": 273 }, { "epoch": 0.2729083665338645, "grad_norm": 0.35947325825691223, "learning_rate": 9.121369763654228e-06, "loss": 2.2618, "step": 274 }, { "epoch": 0.2739043824701195, "grad_norm": 0.3399883210659027, "learning_rate": 9.111495771255623e-06, "loss": 2.1576, "step": 275 }, { "epoch": 0.2749003984063745, "grad_norm": 0.4308667778968811, "learning_rate": 9.101572013897555e-06, "loss": 2.0927, "step": 276 }, { "epoch": 0.27589641434262946, "grad_norm": 0.3370983302593231, "learning_rate": 9.091598611695774e-06, "loss": 2.3073, "step": 277 }, { "epoch": 0.27689243027888444, "grad_norm": 0.30387991666793823, "learning_rate": 9.081575685366919e-06, "loss": 2.5888, "step": 278 }, { "epoch": 0.2778884462151394, "grad_norm": 0.6190817952156067, "learning_rate": 9.071503356227063e-06, "loss": 2.5076, "step": 279 }, { "epoch": 0.2788844621513944, "grad_norm": 0.43932202458381653, "learning_rate": 9.061381746190243e-06, "loss": 2.3828, "step": 280 }, { "epoch": 0.2798804780876494, "grad_norm": 0.4108044505119324, "learning_rate": 9.051210977766987e-06, "loss": 2.4859, "step": 281 }, { "epoch": 0.28087649402390436, "grad_norm": 0.7741344571113586, "learning_rate": 9.040991174062827e-06, "loss": 2.332, "step": 282 }, { "epoch": 0.2818725099601594, "grad_norm": 0.42518022656440735, "learning_rate": 9.030722458776815e-06, "loss": 2.5427, "step": 283 }, { "epoch": 0.28286852589641437, "grad_norm": 0.3040229082107544, "learning_rate": 9.020404956200016e-06, "loss": 2.5173, "step": 284 }, { "epoch": 0.28386454183266935, "grad_norm": 0.4257875680923462, "learning_rate": 9.010038791214012e-06, "loss": 2.3872, "step": 285 }, { "epoch": 0.2848605577689243, "grad_norm": 0.6529532074928284, "learning_rate": 8.999624089289389e-06, "loss": 2.2936, "step": 286 }, { "epoch": 0.2858565737051793, "grad_norm": 0.2723180055618286, "learning_rate": 8.989160976484218e-06, "loss": 2.5736, "step": 287 }, { "epoch": 0.2868525896414343, "grad_norm": 0.3371571898460388, "learning_rate": 8.978649579442525e-06, "loss": 2.3213, "step": 288 }, { "epoch": 0.28784860557768926, "grad_norm": 0.4722624719142914, "learning_rate": 8.968090025392763e-06, "loss": 2.4039, "step": 289 }, { "epoch": 0.28884462151394424, "grad_norm": 0.6963698863983154, "learning_rate": 8.957482442146271e-06, "loss": 2.4849, "step": 290 }, { "epoch": 0.2898406374501992, "grad_norm": 0.41670724749565125, "learning_rate": 8.946826958095726e-06, "loss": 2.4297, "step": 291 }, { "epoch": 0.2908366533864542, "grad_norm": 0.44924449920654297, "learning_rate": 8.936123702213593e-06, "loss": 2.29, "step": 292 }, { "epoch": 0.2918326693227092, "grad_norm": 0.5405289530754089, "learning_rate": 8.925372804050554e-06, "loss": 2.4732, "step": 293 }, { "epoch": 0.29282868525896416, "grad_norm": 0.5333283543586731, "learning_rate": 8.914574393733953e-06, "loss": 2.3553, "step": 294 }, { "epoch": 0.29382470119521914, "grad_norm": 0.4173821806907654, "learning_rate": 8.903728601966206e-06, "loss": 2.4729, "step": 295 }, { "epoch": 0.2948207171314741, "grad_norm": 0.6668480634689331, "learning_rate": 8.892835560023236e-06, "loss": 2.7302, "step": 296 }, { "epoch": 0.2958167330677291, "grad_norm": 0.5601832270622253, "learning_rate": 8.881895399752873e-06, "loss": 2.8201, "step": 297 }, { "epoch": 0.2968127490039841, "grad_norm": 0.7715175151824951, "learning_rate": 8.870908253573255e-06, "loss": 2.439, "step": 298 }, { "epoch": 0.29780876494023906, "grad_norm": 0.6411163210868835, "learning_rate": 8.85987425447124e-06, "loss": 2.2098, "step": 299 }, { "epoch": 0.29880478087649404, "grad_norm": 1.8174595832824707, "learning_rate": 8.848793536000779e-06, "loss": 2.862, "step": 300 }, { "epoch": 0.299800796812749, "grad_norm": 0.4861983060836792, "learning_rate": 8.837666232281312e-06, "loss": 1.964, "step": 301 }, { "epoch": 0.300796812749004, "grad_norm": 0.629531979560852, "learning_rate": 8.826492477996138e-06, "loss": 2.4866, "step": 302 }, { "epoch": 0.301792828685259, "grad_norm": 0.33506232500076294, "learning_rate": 8.81527240839079e-06, "loss": 2.1813, "step": 303 }, { "epoch": 0.30278884462151395, "grad_norm": 0.5065098404884338, "learning_rate": 8.80400615927139e-06, "loss": 2.2313, "step": 304 }, { "epoch": 0.30378486055776893, "grad_norm": 0.31633898615837097, "learning_rate": 8.792693867003017e-06, "loss": 2.5764, "step": 305 }, { "epoch": 0.3047808764940239, "grad_norm": 0.6082801818847656, "learning_rate": 8.781335668508044e-06, "loss": 2.1408, "step": 306 }, { "epoch": 0.3057768924302789, "grad_norm": 0.3309324383735657, "learning_rate": 8.76993170126449e-06, "loss": 2.5198, "step": 307 }, { "epoch": 0.30677290836653387, "grad_norm": 0.38401421904563904, "learning_rate": 8.758482103304348e-06, "loss": 2.2784, "step": 308 }, { "epoch": 0.30776892430278885, "grad_norm": 0.4021207392215729, "learning_rate": 8.746987013211924e-06, "loss": 2.1789, "step": 309 }, { "epoch": 0.30876494023904383, "grad_norm": 1.5585757493972778, "learning_rate": 8.735446570122151e-06, "loss": 2.429, "step": 310 }, { "epoch": 0.3097609561752988, "grad_norm": 0.5734106302261353, "learning_rate": 8.72386091371891e-06, "loss": 2.3943, "step": 311 }, { "epoch": 0.3107569721115538, "grad_norm": 0.5181722044944763, "learning_rate": 8.712230184233337e-06, "loss": 2.4501, "step": 312 }, { "epoch": 0.31175298804780877, "grad_norm": 0.42989951372146606, "learning_rate": 8.700554522442124e-06, "loss": 2.3612, "step": 313 }, { "epoch": 0.31274900398406374, "grad_norm": 0.8372073769569397, "learning_rate": 8.688834069665819e-06, "loss": 2.6138, "step": 314 }, { "epoch": 0.3137450199203187, "grad_norm": 0.5447811484336853, "learning_rate": 8.677068967767117e-06, "loss": 2.4036, "step": 315 }, { "epoch": 0.3147410358565737, "grad_norm": 0.28892412781715393, "learning_rate": 8.665259359149132e-06, "loss": 2.4249, "step": 316 }, { "epoch": 0.3157370517928287, "grad_norm": 0.33981916308403015, "learning_rate": 8.653405386753688e-06, "loss": 2.8255, "step": 317 }, { "epoch": 0.31673306772908366, "grad_norm": 0.37266361713409424, "learning_rate": 8.64150719405958e-06, "loss": 2.4619, "step": 318 }, { "epoch": 0.31772908366533864, "grad_norm": 0.611991286277771, "learning_rate": 8.629564925080838e-06, "loss": 2.6266, "step": 319 }, { "epoch": 0.3187250996015936, "grad_norm": 0.7753398418426514, "learning_rate": 8.617578724364984e-06, "loss": 2.5113, "step": 320 }, { "epoch": 0.3197211155378486, "grad_norm": 0.7812793850898743, "learning_rate": 8.605548736991284e-06, "loss": 2.1736, "step": 321 }, { "epoch": 0.3207171314741036, "grad_norm": 0.5114774703979492, "learning_rate": 8.593475108568995e-06, "loss": 2.7169, "step": 322 }, { "epoch": 0.32171314741035856, "grad_norm": 0.3701231777667999, "learning_rate": 8.581357985235595e-06, "loss": 2.2407, "step": 323 }, { "epoch": 0.32270916334661354, "grad_norm": 1.135130763053894, "learning_rate": 8.569197513655022e-06, "loss": 2.0901, "step": 324 }, { "epoch": 0.3237051792828685, "grad_norm": 0.2848869562149048, "learning_rate": 8.55699384101589e-06, "loss": 2.592, "step": 325 }, { "epoch": 0.3247011952191235, "grad_norm": 0.5609592795372009, "learning_rate": 8.544747115029717e-06, "loss": 2.3673, "step": 326 }, { "epoch": 0.3256972111553785, "grad_norm": 0.3470471203327179, "learning_rate": 8.53245748392913e-06, "loss": 2.377, "step": 327 }, { "epoch": 0.32669322709163345, "grad_norm": 0.9178757667541504, "learning_rate": 8.520125096466072e-06, "loss": 2.7617, "step": 328 }, { "epoch": 0.32768924302788843, "grad_norm": 0.46402791142463684, "learning_rate": 8.50775010191001e-06, "loss": 2.2848, "step": 329 }, { "epoch": 0.3286852589641434, "grad_norm": 0.4459151327610016, "learning_rate": 8.495332650046112e-06, "loss": 2.1105, "step": 330 }, { "epoch": 0.3296812749003984, "grad_norm": 0.7026370763778687, "learning_rate": 8.48287289117345e-06, "loss": 2.6823, "step": 331 }, { "epoch": 0.33067729083665337, "grad_norm": 0.7429327964782715, "learning_rate": 8.470370976103171e-06, "loss": 2.2689, "step": 332 }, { "epoch": 0.33167330677290835, "grad_norm": 0.37948814034461975, "learning_rate": 8.457827056156673e-06, "loss": 2.628, "step": 333 }, { "epoch": 0.33266932270916333, "grad_norm": 0.3749179244041443, "learning_rate": 8.44524128316378e-06, "loss": 2.4552, "step": 334 }, { "epoch": 0.3336653386454183, "grad_norm": 0.421303391456604, "learning_rate": 8.432613809460895e-06, "loss": 2.1797, "step": 335 }, { "epoch": 0.3346613545816733, "grad_norm": 0.5645405054092407, "learning_rate": 8.419944787889162e-06, "loss": 2.4308, "step": 336 }, { "epoch": 0.33565737051792827, "grad_norm": 0.7171806693077087, "learning_rate": 8.407234371792614e-06, "loss": 2.3868, "step": 337 }, { "epoch": 0.33665338645418325, "grad_norm": 0.7937443852424622, "learning_rate": 8.394482715016318e-06, "loss": 2.433, "step": 338 }, { "epoch": 0.3376494023904382, "grad_norm": 0.5325895547866821, "learning_rate": 8.381689971904514e-06, "loss": 2.5116, "step": 339 }, { "epoch": 0.3386454183266932, "grad_norm": 0.3550787568092346, "learning_rate": 8.368856297298742e-06, "loss": 2.4187, "step": 340 }, { "epoch": 0.3396414342629482, "grad_norm": 0.5118217468261719, "learning_rate": 8.355981846535972e-06, "loss": 2.1325, "step": 341 }, { "epoch": 0.34063745019920316, "grad_norm": 0.35231295228004456, "learning_rate": 8.343066775446724e-06, "loss": 2.3751, "step": 342 }, { "epoch": 0.34163346613545814, "grad_norm": 0.6949347853660583, "learning_rate": 8.330111240353178e-06, "loss": 2.5615, "step": 343 }, { "epoch": 0.3426294820717131, "grad_norm": 0.5718231797218323, "learning_rate": 8.317115398067289e-06, "loss": 2.2858, "step": 344 }, { "epoch": 0.3436254980079681, "grad_norm": 0.6337103843688965, "learning_rate": 8.30407940588888e-06, "loss": 2.5088, "step": 345 }, { "epoch": 0.34462151394422313, "grad_norm": 0.4129788875579834, "learning_rate": 8.29100342160374e-06, "loss": 2.2748, "step": 346 }, { "epoch": 0.3456175298804781, "grad_norm": 0.3462570607662201, "learning_rate": 8.27788760348173e-06, "loss": 2.3205, "step": 347 }, { "epoch": 0.3466135458167331, "grad_norm": 0.4417884349822998, "learning_rate": 8.26473211027484e-06, "loss": 2.3901, "step": 348 }, { "epoch": 0.34760956175298807, "grad_norm": 0.45579978823661804, "learning_rate": 8.251537101215287e-06, "loss": 2.2336, "step": 349 }, { "epoch": 0.34860557768924305, "grad_norm": 0.6957226991653442, "learning_rate": 8.238302736013587e-06, "loss": 2.7518, "step": 350 }, { "epoch": 0.34960159362549803, "grad_norm": 0.696114718914032, "learning_rate": 8.225029174856602e-06, "loss": 2.0373, "step": 351 }, { "epoch": 0.350597609561753, "grad_norm": 0.40747031569480896, "learning_rate": 8.211716578405635e-06, "loss": 2.4626, "step": 352 }, { "epoch": 0.351593625498008, "grad_norm": 0.47290411591529846, "learning_rate": 8.198365107794457e-06, "loss": 2.5871, "step": 353 }, { "epoch": 0.35258964143426297, "grad_norm": 0.592217206954956, "learning_rate": 8.184974924627365e-06, "loss": 2.3886, "step": 354 }, { "epoch": 0.35358565737051795, "grad_norm": 0.8985310196876526, "learning_rate": 8.171546190977231e-06, "loss": 2.6021, "step": 355 }, { "epoch": 0.3545816733067729, "grad_norm": 0.7999231815338135, "learning_rate": 8.158079069383535e-06, "loss": 2.2757, "step": 356 }, { "epoch": 0.3555776892430279, "grad_norm": 0.6230331063270569, "learning_rate": 8.1445737228504e-06, "loss": 2.4343, "step": 357 }, { "epoch": 0.3565737051792829, "grad_norm": 0.43640607595443726, "learning_rate": 8.131030314844617e-06, "loss": 1.9181, "step": 358 }, { "epoch": 0.35756972111553786, "grad_norm": 0.5745819211006165, "learning_rate": 8.117449009293668e-06, "loss": 2.9202, "step": 359 }, { "epoch": 0.35856573705179284, "grad_norm": 0.46643760800361633, "learning_rate": 8.103829970583742e-06, "loss": 2.5197, "step": 360 }, { "epoch": 0.3595617529880478, "grad_norm": 0.5368001461029053, "learning_rate": 8.090173363557748e-06, "loss": 2.3562, "step": 361 }, { "epoch": 0.3605577689243028, "grad_norm": 0.5117561221122742, "learning_rate": 8.076479353513308e-06, "loss": 2.2398, "step": 362 }, { "epoch": 0.3615537848605578, "grad_norm": 1.2853957414627075, "learning_rate": 8.06274810620077e-06, "loss": 2.041, "step": 363 }, { "epoch": 0.36254980079681276, "grad_norm": 0.5592566132545471, "learning_rate": 8.0489797878212e-06, "loss": 2.614, "step": 364 }, { "epoch": 0.36354581673306774, "grad_norm": 0.3889990448951721, "learning_rate": 8.035174565024362e-06, "loss": 2.4095, "step": 365 }, { "epoch": 0.3645418326693227, "grad_norm": 0.39642640948295593, "learning_rate": 8.021332604906709e-06, "loss": 1.9971, "step": 366 }, { "epoch": 0.3655378486055777, "grad_norm": 0.5725635290145874, "learning_rate": 8.007454075009352e-06, "loss": 2.4322, "step": 367 }, { "epoch": 0.3665338645418327, "grad_norm": 0.538329541683197, "learning_rate": 7.993539143316044e-06, "loss": 2.3927, "step": 368 }, { "epoch": 0.36752988047808766, "grad_norm": 0.8502817153930664, "learning_rate": 7.979587978251136e-06, "loss": 2.6487, "step": 369 }, { "epoch": 0.36852589641434264, "grad_norm": 0.5096033811569214, "learning_rate": 7.965600748677545e-06, "loss": 2.4393, "step": 370 }, { "epoch": 0.3695219123505976, "grad_norm": 0.39716872572898865, "learning_rate": 7.951577623894701e-06, "loss": 2.9555, "step": 371 }, { "epoch": 0.3705179282868526, "grad_norm": 0.6120476126670837, "learning_rate": 7.937518773636518e-06, "loss": 2.1758, "step": 372 }, { "epoch": 0.3715139442231076, "grad_norm": 0.4519326984882355, "learning_rate": 7.923424368069312e-06, "loss": 2.4508, "step": 373 }, { "epoch": 0.37250996015936255, "grad_norm": 0.4252610504627228, "learning_rate": 7.909294577789765e-06, "loss": 2.134, "step": 374 }, { "epoch": 0.37350597609561753, "grad_norm": 0.5511481761932373, "learning_rate": 7.895129573822844e-06, "loss": 2.4735, "step": 375 }, { "epoch": 0.3745019920318725, "grad_norm": 1.1340324878692627, "learning_rate": 7.880929527619742e-06, "loss": 2.7602, "step": 376 }, { "epoch": 0.3754980079681275, "grad_norm": 0.45862501859664917, "learning_rate": 7.866694611055796e-06, "loss": 2.5242, "step": 377 }, { "epoch": 0.37649402390438247, "grad_norm": 0.48843666911125183, "learning_rate": 7.852424996428412e-06, "loss": 2.3878, "step": 378 }, { "epoch": 0.37749003984063745, "grad_norm": 1.07990562915802, "learning_rate": 7.838120856454967e-06, "loss": 2.2745, "step": 379 }, { "epoch": 0.3784860557768924, "grad_norm": 0.466766893863678, "learning_rate": 7.823782364270743e-06, "loss": 2.5844, "step": 380 }, { "epoch": 0.3794820717131474, "grad_norm": 0.6437628865242004, "learning_rate": 7.809409693426803e-06, "loss": 2.381, "step": 381 }, { "epoch": 0.3804780876494024, "grad_norm": 0.6378084421157837, "learning_rate": 7.79500301788791e-06, "loss": 2.4076, "step": 382 }, { "epoch": 0.38147410358565736, "grad_norm": 0.6559402346611023, "learning_rate": 7.780562512030414e-06, "loss": 2.2115, "step": 383 }, { "epoch": 0.38247011952191234, "grad_norm": 0.4882892668247223, "learning_rate": 7.766088350640141e-06, "loss": 2.179, "step": 384 }, { "epoch": 0.3834661354581673, "grad_norm": 0.37981244921684265, "learning_rate": 7.75158070891028e-06, "loss": 2.3806, "step": 385 }, { "epoch": 0.3844621513944223, "grad_norm": 0.6670547723770142, "learning_rate": 7.737039762439263e-06, "loss": 2.0258, "step": 386 }, { "epoch": 0.3854581673306773, "grad_norm": 0.5870895981788635, "learning_rate": 7.722465687228634e-06, "loss": 2.4609, "step": 387 }, { "epoch": 0.38645418326693226, "grad_norm": 0.5210617184638977, "learning_rate": 7.707858659680924e-06, "loss": 2.0152, "step": 388 }, { "epoch": 0.38745019920318724, "grad_norm": 0.8872121572494507, "learning_rate": 7.693218856597515e-06, "loss": 2.4107, "step": 389 }, { "epoch": 0.3884462151394422, "grad_norm": 0.408750057220459, "learning_rate": 7.6785464551765e-06, "loss": 2.4381, "step": 390 }, { "epoch": 0.3894422310756972, "grad_norm": 0.4789107143878937, "learning_rate": 7.663841633010539e-06, "loss": 2.3057, "step": 391 }, { "epoch": 0.3904382470119522, "grad_norm": 0.37599480152130127, "learning_rate": 7.649104568084701e-06, "loss": 2.4101, "step": 392 }, { "epoch": 0.39143426294820716, "grad_norm": 0.6352246403694153, "learning_rate": 7.634335438774325e-06, "loss": 2.2202, "step": 393 }, { "epoch": 0.39243027888446214, "grad_norm": 0.5684521198272705, "learning_rate": 7.619534423842852e-06, "loss": 2.7473, "step": 394 }, { "epoch": 0.3934262948207171, "grad_norm": 0.5998817682266235, "learning_rate": 7.604701702439652e-06, "loss": 2.109, "step": 395 }, { "epoch": 0.3944223107569721, "grad_norm": 0.69579017162323, "learning_rate": 7.589837454097879e-06, "loss": 2.6706, "step": 396 }, { "epoch": 0.3954183266932271, "grad_norm": 0.5617753267288208, "learning_rate": 7.574941858732279e-06, "loss": 2.4175, "step": 397 }, { "epoch": 0.39641434262948205, "grad_norm": 0.509094774723053, "learning_rate": 7.560015096637015e-06, "loss": 2.4367, "step": 398 }, { "epoch": 0.39741035856573703, "grad_norm": 0.5854381322860718, "learning_rate": 7.54505734848349e-06, "loss": 2.3366, "step": 399 }, { "epoch": 0.398406374501992, "grad_norm": 0.7329273819923401, "learning_rate": 7.53006879531816e-06, "loss": 2.2252, "step": 400 }, { "epoch": 0.399402390438247, "grad_norm": 0.7855085730552673, "learning_rate": 7.515049618560337e-06, "loss": 2.2244, "step": 401 }, { "epoch": 0.40039840637450197, "grad_norm": 0.608400821685791, "learning_rate": 7.500000000000001e-06, "loss": 2.2232, "step": 402 }, { "epoch": 0.40139442231075695, "grad_norm": 0.5910929441452026, "learning_rate": 7.484920121795589e-06, "loss": 2.4034, "step": 403 }, { "epoch": 0.40239043824701193, "grad_norm": 0.5254145264625549, "learning_rate": 7.469810166471802e-06, "loss": 2.3053, "step": 404 }, { "epoch": 0.4033864541832669, "grad_norm": 0.5197448134422302, "learning_rate": 7.454670316917387e-06, "loss": 2.2416, "step": 405 }, { "epoch": 0.4043824701195219, "grad_norm": 0.759675145149231, "learning_rate": 7.4395007563829295e-06, "loss": 2.4197, "step": 406 }, { "epoch": 0.40537848605577687, "grad_norm": 0.4646972417831421, "learning_rate": 7.424301668478626e-06, "loss": 2.5638, "step": 407 }, { "epoch": 0.4063745019920319, "grad_norm": 0.5564824938774109, "learning_rate": 7.4090732371720774e-06, "loss": 2.5038, "step": 408 }, { "epoch": 0.4073705179282869, "grad_norm": 0.8897591829299927, "learning_rate": 7.393815646786047e-06, "loss": 2.6751, "step": 409 }, { "epoch": 0.40836653386454186, "grad_norm": 0.8237727284431458, "learning_rate": 7.378529081996233e-06, "loss": 2.2649, "step": 410 }, { "epoch": 0.40936254980079684, "grad_norm": 0.4922022521495819, "learning_rate": 7.363213727829045e-06, "loss": 2.2894, "step": 411 }, { "epoch": 0.4103585657370518, "grad_norm": 0.48476412892341614, "learning_rate": 7.347869769659346e-06, "loss": 2.4317, "step": 412 }, { "epoch": 0.4113545816733068, "grad_norm": 0.7627730965614319, "learning_rate": 7.332497393208221e-06, "loss": 3.0171, "step": 413 }, { "epoch": 0.4123505976095618, "grad_norm": 1.4489892721176147, "learning_rate": 7.317096784540728e-06, "loss": 3.0745, "step": 414 }, { "epoch": 0.41334661354581675, "grad_norm": 0.44949018955230713, "learning_rate": 7.301668130063639e-06, "loss": 2.4086, "step": 415 }, { "epoch": 0.41434262948207173, "grad_norm": 0.9026828408241272, "learning_rate": 7.286211616523193e-06, "loss": 2.4219, "step": 416 }, { "epoch": 0.4153386454183267, "grad_norm": 0.5528742074966431, "learning_rate": 7.2707274310028306e-06, "loss": 2.069, "step": 417 }, { "epoch": 0.4163346613545817, "grad_norm": 0.8069695830345154, "learning_rate": 7.255215760920925e-06, "loss": 2.2901, "step": 418 }, { "epoch": 0.41733067729083667, "grad_norm": 0.9854758977890015, "learning_rate": 7.239676794028526e-06, "loss": 2.533, "step": 419 }, { "epoch": 0.41832669322709165, "grad_norm": 0.8304996490478516, "learning_rate": 7.224110718407075e-06, "loss": 2.4076, "step": 420 }, { "epoch": 0.41932270916334663, "grad_norm": 0.5824740529060364, "learning_rate": 7.208517722466135e-06, "loss": 2.2298, "step": 421 }, { "epoch": 0.4203187250996016, "grad_norm": 0.740998387336731, "learning_rate": 7.192897994941111e-06, "loss": 2.2991, "step": 422 }, { "epoch": 0.4213147410358566, "grad_norm": 0.9516714215278625, "learning_rate": 7.177251724890957e-06, "loss": 2.2348, "step": 423 }, { "epoch": 0.42231075697211157, "grad_norm": 1.330517053604126, "learning_rate": 7.1615791016959024e-06, "loss": 2.5969, "step": 424 }, { "epoch": 0.42330677290836655, "grad_norm": 0.8636577129364014, "learning_rate": 7.145880315055145e-06, "loss": 2.4203, "step": 425 }, { "epoch": 0.4243027888446215, "grad_norm": 0.40839532017707825, "learning_rate": 7.1301555549845634e-06, "loss": 2.428, "step": 426 }, { "epoch": 0.4252988047808765, "grad_norm": 0.6337350010871887, "learning_rate": 7.114405011814415e-06, "loss": 2.3018, "step": 427 }, { "epoch": 0.4262948207171315, "grad_norm": 0.8653415441513062, "learning_rate": 7.098628876187031e-06, "loss": 2.8947, "step": 428 }, { "epoch": 0.42729083665338646, "grad_norm": 0.7558097839355469, "learning_rate": 7.082827339054513e-06, "loss": 2.2061, "step": 429 }, { "epoch": 0.42828685258964144, "grad_norm": 0.8000844717025757, "learning_rate": 7.067000591676416e-06, "loss": 2.4076, "step": 430 }, { "epoch": 0.4292828685258964, "grad_norm": 0.7074631452560425, "learning_rate": 7.051148825617435e-06, "loss": 2.7278, "step": 431 }, { "epoch": 0.4302788844621514, "grad_norm": 1.1026921272277832, "learning_rate": 7.035272232745093e-06, "loss": 2.5602, "step": 432 }, { "epoch": 0.4312749003984064, "grad_norm": 0.8589175939559937, "learning_rate": 7.019371005227407e-06, "loss": 2.3789, "step": 433 }, { "epoch": 0.43227091633466136, "grad_norm": 0.8680172562599182, "learning_rate": 7.003445335530572e-06, "loss": 2.6989, "step": 434 }, { "epoch": 0.43326693227091634, "grad_norm": 0.7565051317214966, "learning_rate": 6.987495416416627e-06, "loss": 2.3688, "step": 435 }, { "epoch": 0.4342629482071713, "grad_norm": 1.050288438796997, "learning_rate": 6.9715214409411204e-06, "loss": 2.1948, "step": 436 }, { "epoch": 0.4352589641434263, "grad_norm": 0.5728120803833008, "learning_rate": 6.95552360245078e-06, "loss": 2.2716, "step": 437 }, { "epoch": 0.4362549800796813, "grad_norm": 1.0542654991149902, "learning_rate": 6.939502094581164e-06, "loss": 2.5845, "step": 438 }, { "epoch": 0.43725099601593626, "grad_norm": 0.7242105603218079, "learning_rate": 6.923457111254322e-06, "loss": 2.2048, "step": 439 }, { "epoch": 0.43824701195219123, "grad_norm": 1.0688732862472534, "learning_rate": 6.9073888466764495e-06, "loss": 2.1883, "step": 440 }, { "epoch": 0.4392430278884462, "grad_norm": 0.8276563286781311, "learning_rate": 6.891297495335531e-06, "loss": 2.2532, "step": 441 }, { "epoch": 0.4402390438247012, "grad_norm": 0.7350841760635376, "learning_rate": 6.875183251998993e-06, "loss": 2.2517, "step": 442 }, { "epoch": 0.44123505976095617, "grad_norm": 0.7074620127677917, "learning_rate": 6.859046311711344e-06, "loss": 2.1943, "step": 443 }, { "epoch": 0.44223107569721115, "grad_norm": 0.9273977279663086, "learning_rate": 6.84288686979181e-06, "loss": 2.4566, "step": 444 }, { "epoch": 0.44322709163346613, "grad_norm": 0.54817795753479, "learning_rate": 6.8267051218319766e-06, "loss": 2.1578, "step": 445 }, { "epoch": 0.4442231075697211, "grad_norm": 0.8614233136177063, "learning_rate": 6.810501263693416e-06, "loss": 2.3546, "step": 446 }, { "epoch": 0.4452191235059761, "grad_norm": 0.5919457077980042, "learning_rate": 6.7942754915053225e-06, "loss": 2.4907, "step": 447 }, { "epoch": 0.44621513944223107, "grad_norm": 0.7229816317558289, "learning_rate": 6.77802800166213e-06, "loss": 2.2884, "step": 448 }, { "epoch": 0.44721115537848605, "grad_norm": 1.1251389980316162, "learning_rate": 6.761758990821143e-06, "loss": 2.508, "step": 449 }, { "epoch": 0.448207171314741, "grad_norm": 0.7033310532569885, "learning_rate": 6.745468655900156e-06, "loss": 2.4315, "step": 450 }, { "epoch": 0.449203187250996, "grad_norm": 1.2203772068023682, "learning_rate": 6.7291571940750575e-06, "loss": 2.6156, "step": 451 }, { "epoch": 0.450199203187251, "grad_norm": 0.7150283455848694, "learning_rate": 6.712824802777465e-06, "loss": 2.2121, "step": 452 }, { "epoch": 0.45119521912350596, "grad_norm": 0.9422833323478699, "learning_rate": 6.696471679692313e-06, "loss": 2.2294, "step": 453 }, { "epoch": 0.45219123505976094, "grad_norm": 0.6846040487289429, "learning_rate": 6.680098022755478e-06, "loss": 2.5567, "step": 454 }, { "epoch": 0.4531872509960159, "grad_norm": 0.5450727343559265, "learning_rate": 6.66370403015137e-06, "loss": 2.4599, "step": 455 }, { "epoch": 0.4541832669322709, "grad_norm": 0.5914618968963623, "learning_rate": 6.647289900310545e-06, "loss": 2.5134, "step": 456 }, { "epoch": 0.4551792828685259, "grad_norm": 0.8993861079216003, "learning_rate": 6.63085583190729e-06, "loss": 2.5729, "step": 457 }, { "epoch": 0.45617529880478086, "grad_norm": 0.5783509016036987, "learning_rate": 6.614402023857231e-06, "loss": 2.5881, "step": 458 }, { "epoch": 0.45717131474103584, "grad_norm": 0.6340298652648926, "learning_rate": 6.597928675314918e-06, "loss": 2.4509, "step": 459 }, { "epoch": 0.4581673306772908, "grad_norm": 0.7813217043876648, "learning_rate": 6.581435985671418e-06, "loss": 2.242, "step": 460 }, { "epoch": 0.4591633466135458, "grad_norm": 0.8051680326461792, "learning_rate": 6.564924154551895e-06, "loss": 2.398, "step": 461 }, { "epoch": 0.4601593625498008, "grad_norm": 0.6447633504867554, "learning_rate": 6.548393381813205e-06, "loss": 2.9214, "step": 462 }, { "epoch": 0.46115537848605576, "grad_norm": 0.5684821605682373, "learning_rate": 6.5318438675414665e-06, "loss": 2.2545, "step": 463 }, { "epoch": 0.46215139442231074, "grad_norm": 0.6067225337028503, "learning_rate": 6.515275812049644e-06, "loss": 2.5541, "step": 464 }, { "epoch": 0.4631474103585657, "grad_norm": 0.5634474754333496, "learning_rate": 6.498689415875121e-06, "loss": 2.581, "step": 465 }, { "epoch": 0.4641434262948207, "grad_norm": 0.4764470160007477, "learning_rate": 6.48208487977728e-06, "loss": 2.2492, "step": 466 }, { "epoch": 0.4651394422310757, "grad_norm": 0.6636649370193481, "learning_rate": 6.4654624047350575e-06, "loss": 2.418, "step": 467 }, { "epoch": 0.46613545816733065, "grad_norm": 0.6568376421928406, "learning_rate": 6.448822191944526e-06, "loss": 2.4608, "step": 468 }, { "epoch": 0.46713147410358563, "grad_norm": 0.5867657661437988, "learning_rate": 6.432164442816452e-06, "loss": 2.7101, "step": 469 }, { "epoch": 0.4681274900398406, "grad_norm": 0.5845500826835632, "learning_rate": 6.41548935897386e-06, "loss": 2.1822, "step": 470 }, { "epoch": 0.46912350597609564, "grad_norm": 0.9894917011260986, "learning_rate": 6.398797142249591e-06, "loss": 2.4232, "step": 471 }, { "epoch": 0.4701195219123506, "grad_norm": 0.5987226366996765, "learning_rate": 6.3820879946838585e-06, "loss": 2.6506, "step": 472 }, { "epoch": 0.4711155378486056, "grad_norm": 1.552355408668518, "learning_rate": 6.365362118521807e-06, "loss": 3.0233, "step": 473 }, { "epoch": 0.4721115537848606, "grad_norm": 0.6667497158050537, "learning_rate": 6.348619716211058e-06, "loss": 2.6748, "step": 474 }, { "epoch": 0.47310756972111556, "grad_norm": 0.970600962638855, "learning_rate": 6.33186099039927e-06, "loss": 2.4542, "step": 475 }, { "epoch": 0.47410358565737054, "grad_norm": 0.6683152914047241, "learning_rate": 6.31508614393167e-06, "loss": 2.4034, "step": 476 }, { "epoch": 0.4750996015936255, "grad_norm": 0.9013263583183289, "learning_rate": 6.2982953798486124e-06, "loss": 2.4747, "step": 477 }, { "epoch": 0.4760956175298805, "grad_norm": 0.8827518820762634, "learning_rate": 6.2814889013831174e-06, "loss": 2.5649, "step": 478 }, { "epoch": 0.4770916334661355, "grad_norm": 0.8040870428085327, "learning_rate": 6.264666911958404e-06, "loss": 2.2855, "step": 479 }, { "epoch": 0.47808764940239046, "grad_norm": 0.9028819799423218, "learning_rate": 6.247829615185441e-06, "loss": 2.3607, "step": 480 }, { "epoch": 0.47908366533864544, "grad_norm": 0.8722829818725586, "learning_rate": 6.230977214860468e-06, "loss": 2.2346, "step": 481 }, { "epoch": 0.4800796812749004, "grad_norm": 0.6933993697166443, "learning_rate": 6.214109914962542e-06, "loss": 2.7604, "step": 482 }, { "epoch": 0.4810756972111554, "grad_norm": 0.6176011562347412, "learning_rate": 6.1972279196510565e-06, "loss": 2.8278, "step": 483 }, { "epoch": 0.4820717131474104, "grad_norm": 0.7204033732414246, "learning_rate": 6.180331433263283e-06, "loss": 2.4275, "step": 484 }, { "epoch": 0.48306772908366535, "grad_norm": 1.1777395009994507, "learning_rate": 6.1634206603118844e-06, "loss": 2.501, "step": 485 }, { "epoch": 0.48406374501992033, "grad_norm": 1.3905079364776611, "learning_rate": 6.146495805482451e-06, "loss": 2.9668, "step": 486 }, { "epoch": 0.4850597609561753, "grad_norm": 1.8294525146484375, "learning_rate": 6.129557073631013e-06, "loss": 2.0932, "step": 487 }, { "epoch": 0.4860557768924303, "grad_norm": 1.0543792247772217, "learning_rate": 6.112604669781572e-06, "loss": 2.1609, "step": 488 }, { "epoch": 0.48705179282868527, "grad_norm": 0.6805559992790222, "learning_rate": 6.09563879912361e-06, "loss": 2.5514, "step": 489 }, { "epoch": 0.48804780876494025, "grad_norm": 0.913993239402771, "learning_rate": 6.07865966700961e-06, "loss": 2.1365, "step": 490 }, { "epoch": 0.48904382470119523, "grad_norm": 1.127585768699646, "learning_rate": 6.06166747895257e-06, "loss": 2.5846, "step": 491 }, { "epoch": 0.4900398406374502, "grad_norm": 0.5136232376098633, "learning_rate": 6.044662440623512e-06, "loss": 2.3032, "step": 492 }, { "epoch": 0.4910358565737052, "grad_norm": 1.1162834167480469, "learning_rate": 6.027644757849004e-06, "loss": 2.4506, "step": 493 }, { "epoch": 0.49203187250996017, "grad_norm": 0.8343062996864319, "learning_rate": 6.0106146366086514e-06, "loss": 2.5259, "step": 494 }, { "epoch": 0.49302788844621515, "grad_norm": 1.2255266904830933, "learning_rate": 5.99357228303262e-06, "loss": 2.4964, "step": 495 }, { "epoch": 0.4940239043824701, "grad_norm": 0.8280947804450989, "learning_rate": 5.976517903399128e-06, "loss": 2.3576, "step": 496 }, { "epoch": 0.4950199203187251, "grad_norm": 0.5358011722564697, "learning_rate": 5.959451704131962e-06, "loss": 2.4103, "step": 497 }, { "epoch": 0.4960159362549801, "grad_norm": 0.7300974130630493, "learning_rate": 5.9423738917979655e-06, "loss": 2.4792, "step": 498 }, { "epoch": 0.49701195219123506, "grad_norm": 1.1368849277496338, "learning_rate": 5.92528467310455e-06, "loss": 2.7343, "step": 499 }, { "epoch": 0.49800796812749004, "grad_norm": 1.1836261749267578, "learning_rate": 5.908184254897183e-06, "loss": 2.263, "step": 500 }, { "epoch": 0.499003984063745, "grad_norm": 0.7548766136169434, "learning_rate": 5.891072844156895e-06, "loss": 2.5232, "step": 501 }, { "epoch": 0.5, "grad_norm": 0.5676743984222412, "learning_rate": 5.87395064799776e-06, "loss": 2.532, "step": 502 }, { "epoch": 0.500996015936255, "grad_norm": 0.6243408918380737, "learning_rate": 5.856817873664409e-06, "loss": 2.6337, "step": 503 }, { "epoch": 0.50199203187251, "grad_norm": 0.6920816898345947, "learning_rate": 5.839674728529499e-06, "loss": 2.3761, "step": 504 }, { "epoch": 0.5029880478087649, "grad_norm": 0.773199737071991, "learning_rate": 5.8225214200912195e-06, "loss": 2.3334, "step": 505 }, { "epoch": 0.5039840637450199, "grad_norm": 0.8243865370750427, "learning_rate": 5.8053581559707754e-06, "loss": 2.2201, "step": 506 }, { "epoch": 0.5049800796812749, "grad_norm": 0.6483944654464722, "learning_rate": 5.788185143909868e-06, "loss": 2.2019, "step": 507 }, { "epoch": 0.5059760956175299, "grad_norm": 1.1562165021896362, "learning_rate": 5.7710025917681954e-06, "loss": 2.6738, "step": 508 }, { "epoch": 0.5069721115537849, "grad_norm": 0.7714682221412659, "learning_rate": 5.753810707520918e-06, "loss": 2.5662, "step": 509 }, { "epoch": 0.5079681274900398, "grad_norm": 1.8939898014068604, "learning_rate": 5.736609699256158e-06, "loss": 2.4841, "step": 510 }, { "epoch": 0.5089641434262948, "grad_norm": 1.0769683122634888, "learning_rate": 5.719399775172468e-06, "loss": 2.2339, "step": 511 }, { "epoch": 0.5099601593625498, "grad_norm": 0.7231705188751221, "learning_rate": 5.702181143576323e-06, "loss": 2.2754, "step": 512 }, { "epoch": 0.5109561752988048, "grad_norm": 1.0154212713241577, "learning_rate": 5.684954012879583e-06, "loss": 2.5029, "step": 513 }, { "epoch": 0.5119521912350598, "grad_norm": 1.130210041999817, "learning_rate": 5.66771859159699e-06, "loss": 2.5272, "step": 514 }, { "epoch": 0.5129482071713147, "grad_norm": 0.6206554770469666, "learning_rate": 5.6504750883436275e-06, "loss": 2.3235, "step": 515 }, { "epoch": 0.5139442231075697, "grad_norm": 1.0481131076812744, "learning_rate": 5.633223711832403e-06, "loss": 2.2866, "step": 516 }, { "epoch": 0.5149402390438247, "grad_norm": 0.7321045398712158, "learning_rate": 5.615964670871524e-06, "loss": 2.4402, "step": 517 }, { "epoch": 0.5159362549800797, "grad_norm": 0.7684382796287537, "learning_rate": 5.5986981743619615e-06, "loss": 2.3404, "step": 518 }, { "epoch": 0.5169322709163346, "grad_norm": 1.1246601343154907, "learning_rate": 5.581424431294936e-06, "loss": 2.2032, "step": 519 }, { "epoch": 0.5179282868525896, "grad_norm": 0.617168128490448, "learning_rate": 5.56414365074937e-06, "loss": 2.5379, "step": 520 }, { "epoch": 0.5189243027888446, "grad_norm": 0.7718055248260498, "learning_rate": 5.546856041889374e-06, "loss": 2.4803, "step": 521 }, { "epoch": 0.5199203187250996, "grad_norm": 1.5422130823135376, "learning_rate": 5.5295618139617e-06, "loss": 2.3014, "step": 522 }, { "epoch": 0.5209163346613546, "grad_norm": 0.5689607262611389, "learning_rate": 5.512261176293226e-06, "loss": 2.1094, "step": 523 }, { "epoch": 0.5219123505976095, "grad_norm": 0.6268129944801331, "learning_rate": 5.494954338288404e-06, "loss": 2.4562, "step": 524 }, { "epoch": 0.5229083665338645, "grad_norm": 1.1070692539215088, "learning_rate": 5.477641509426739e-06, "loss": 2.0365, "step": 525 }, { "epoch": 0.5239043824701195, "grad_norm": 1.2830649614334106, "learning_rate": 5.460322899260245e-06, "loss": 2.2455, "step": 526 }, { "epoch": 0.5249003984063745, "grad_norm": 0.6706056594848633, "learning_rate": 5.442998717410916e-06, "loss": 2.3095, "step": 527 }, { "epoch": 0.5258964143426295, "grad_norm": 0.6565206050872803, "learning_rate": 5.425669173568179e-06, "loss": 2.5873, "step": 528 }, { "epoch": 0.5268924302788844, "grad_norm": 0.8396487236022949, "learning_rate": 5.408334477486369e-06, "loss": 2.4121, "step": 529 }, { "epoch": 0.5278884462151394, "grad_norm": 0.8442867398262024, "learning_rate": 5.390994838982178e-06, "loss": 2.1112, "step": 530 }, { "epoch": 0.5288844621513944, "grad_norm": 0.6712931990623474, "learning_rate": 5.373650467932122e-06, "loss": 2.5735, "step": 531 }, { "epoch": 0.5298804780876494, "grad_norm": 0.8547630906105042, "learning_rate": 5.356301574269998e-06, "loss": 2.5899, "step": 532 }, { "epoch": 0.5308764940239044, "grad_norm": 1.1971904039382935, "learning_rate": 5.338948367984347e-06, "loss": 2.7327, "step": 533 }, { "epoch": 0.5318725099601593, "grad_norm": 0.8025546669960022, "learning_rate": 5.321591059115906e-06, "loss": 2.3269, "step": 534 }, { "epoch": 0.5328685258964143, "grad_norm": 0.9815241098403931, "learning_rate": 5.30422985775507e-06, "loss": 2.4555, "step": 535 }, { "epoch": 0.5338645418326693, "grad_norm": 0.9032710194587708, "learning_rate": 5.286864974039349e-06, "loss": 2.4246, "step": 536 }, { "epoch": 0.5348605577689243, "grad_norm": 1.508058786392212, "learning_rate": 5.269496618150823e-06, "loss": 2.1206, "step": 537 }, { "epoch": 0.5358565737051793, "grad_norm": 1.0455362796783447, "learning_rate": 5.2521250003136005e-06, "loss": 2.4406, "step": 538 }, { "epoch": 0.5368525896414342, "grad_norm": 0.8184682726860046, "learning_rate": 5.234750330791268e-06, "loss": 2.1303, "step": 539 }, { "epoch": 0.5378486055776892, "grad_norm": 0.8355326056480408, "learning_rate": 5.217372819884353e-06, "loss": 2.3468, "step": 540 }, { "epoch": 0.5388446215139442, "grad_norm": 1.0594871044158936, "learning_rate": 5.199992677927775e-06, "loss": 2.2558, "step": 541 }, { "epoch": 0.5398406374501992, "grad_norm": 1.3626909255981445, "learning_rate": 5.182610115288296e-06, "loss": 2.4741, "step": 542 }, { "epoch": 0.5408366533864541, "grad_norm": 0.6529922485351562, "learning_rate": 5.165225342361978e-06, "loss": 2.3131, "step": 543 }, { "epoch": 0.5418326693227091, "grad_norm": 0.8616756796836853, "learning_rate": 5.147838569571642e-06, "loss": 2.2786, "step": 544 }, { "epoch": 0.5428286852589641, "grad_norm": 1.9104212522506714, "learning_rate": 5.1304500073643045e-06, "loss": 2.2784, "step": 545 }, { "epoch": 0.5438247011952191, "grad_norm": 0.7414434552192688, "learning_rate": 5.1130598662086514e-06, "loss": 2.1551, "step": 546 }, { "epoch": 0.5448207171314741, "grad_norm": 0.889681339263916, "learning_rate": 5.095668356592474e-06, "loss": 1.9545, "step": 547 }, { "epoch": 0.545816733067729, "grad_norm": 0.8562681674957275, "learning_rate": 5.078275689020129e-06, "loss": 2.1665, "step": 548 }, { "epoch": 0.546812749003984, "grad_norm": 0.810234010219574, "learning_rate": 5.060882074009988e-06, "loss": 2.5324, "step": 549 }, { "epoch": 0.547808764940239, "grad_norm": 1.867493748664856, "learning_rate": 5.043487722091891e-06, "loss": 2.6972, "step": 550 }, { "epoch": 0.548804780876494, "grad_norm": 0.8987257480621338, "learning_rate": 5.026092843804599e-06, "loss": 2.3632, "step": 551 }, { "epoch": 0.549800796812749, "grad_norm": 0.9021519422531128, "learning_rate": 5.0086976496932374e-06, "loss": 2.2825, "step": 552 }, { "epoch": 0.5507968127490039, "grad_norm": 1.0589499473571777, "learning_rate": 4.991302350306764e-06, "loss": 2.6046, "step": 553 }, { "epoch": 0.5517928286852589, "grad_norm": 0.8575243949890137, "learning_rate": 4.973907156195405e-06, "loss": 2.6962, "step": 554 }, { "epoch": 0.5527888446215139, "grad_norm": 1.7027395963668823, "learning_rate": 4.956512277908109e-06, "loss": 2.4405, "step": 555 }, { "epoch": 0.5537848605577689, "grad_norm": 0.8842843770980835, "learning_rate": 4.939117925990013e-06, "loss": 2.4694, "step": 556 }, { "epoch": 0.5547808764940239, "grad_norm": 0.9840981364250183, "learning_rate": 4.921724310979872e-06, "loss": 2.5034, "step": 557 }, { "epoch": 0.5557768924302788, "grad_norm": 0.8198688626289368, "learning_rate": 4.904331643407527e-06, "loss": 2.5664, "step": 558 }, { "epoch": 0.5567729083665338, "grad_norm": 1.1393426656723022, "learning_rate": 4.886940133791349e-06, "loss": 2.4415, "step": 559 }, { "epoch": 0.5577689243027888, "grad_norm": 0.835932731628418, "learning_rate": 4.869549992635697e-06, "loss": 2.3719, "step": 560 }, { "epoch": 0.5587649402390438, "grad_norm": 0.952167272567749, "learning_rate": 4.8521614304283615e-06, "loss": 2.3092, "step": 561 }, { "epoch": 0.5597609561752988, "grad_norm": 0.8146136999130249, "learning_rate": 4.834774657638023e-06, "loss": 2.5503, "step": 562 }, { "epoch": 0.5607569721115537, "grad_norm": 0.7990023493766785, "learning_rate": 4.817389884711706e-06, "loss": 2.3319, "step": 563 }, { "epoch": 0.5617529880478087, "grad_norm": 0.5947994589805603, "learning_rate": 4.800007322072226e-06, "loss": 2.1591, "step": 564 }, { "epoch": 0.5627490039840638, "grad_norm": 0.8183361291885376, "learning_rate": 4.7826271801156485e-06, "loss": 2.431, "step": 565 }, { "epoch": 0.5637450199203188, "grad_norm": 0.896101713180542, "learning_rate": 4.765249669208733e-06, "loss": 2.3313, "step": 566 }, { "epoch": 0.5647410358565738, "grad_norm": 1.7690149545669556, "learning_rate": 4.747874999686401e-06, "loss": 2.1365, "step": 567 }, { "epoch": 0.5657370517928287, "grad_norm": 0.8712881803512573, "learning_rate": 4.730503381849179e-06, "loss": 2.4139, "step": 568 }, { "epoch": 0.5667330677290837, "grad_norm": 1.0946391820907593, "learning_rate": 4.713135025960652e-06, "loss": 2.1844, "step": 569 }, { "epoch": 0.5677290836653387, "grad_norm": 0.9877662658691406, "learning_rate": 4.695770142244931e-06, "loss": 2.4189, "step": 570 }, { "epoch": 0.5687250996015937, "grad_norm": 3.4181249141693115, "learning_rate": 4.6784089408840955e-06, "loss": 2.7926, "step": 571 }, { "epoch": 0.5697211155378487, "grad_norm": 0.7407424449920654, "learning_rate": 4.661051632015655e-06, "loss": 2.4036, "step": 572 }, { "epoch": 0.5707171314741036, "grad_norm": 0.8603093028068542, "learning_rate": 4.643698425730004e-06, "loss": 2.0895, "step": 573 }, { "epoch": 0.5717131474103586, "grad_norm": 1.5615819692611694, "learning_rate": 4.626349532067879e-06, "loss": 2.6668, "step": 574 }, { "epoch": 0.5727091633466136, "grad_norm": 0.9211438298225403, "learning_rate": 4.609005161017824e-06, "loss": 2.7534, "step": 575 }, { "epoch": 0.5737051792828686, "grad_norm": 0.6863355040550232, "learning_rate": 4.591665522513633e-06, "loss": 2.4823, "step": 576 }, { "epoch": 0.5747011952191236, "grad_norm": 1.1497968435287476, "learning_rate": 4.574330826431822e-06, "loss": 2.7213, "step": 577 }, { "epoch": 0.5756972111553785, "grad_norm": 0.6396723389625549, "learning_rate": 4.557001282589086e-06, "loss": 2.342, "step": 578 }, { "epoch": 0.5766932270916335, "grad_norm": 0.7893930673599243, "learning_rate": 4.5396771007397565e-06, "loss": 2.426, "step": 579 }, { "epoch": 0.5776892430278885, "grad_norm": 1.0288350582122803, "learning_rate": 4.5223584905732635e-06, "loss": 2.3415, "step": 580 }, { "epoch": 0.5786852589641435, "grad_norm": 1.215003252029419, "learning_rate": 4.505045661711596e-06, "loss": 2.2311, "step": 581 }, { "epoch": 0.5796812749003984, "grad_norm": 1.2418211698532104, "learning_rate": 4.487738823706775e-06, "loss": 2.6043, "step": 582 }, { "epoch": 0.5806772908366534, "grad_norm": 0.658145546913147, "learning_rate": 4.470438186038301e-06, "loss": 2.5859, "step": 583 }, { "epoch": 0.5816733067729084, "grad_norm": 0.9392750859260559, "learning_rate": 4.4531439581106295e-06, "loss": 2.4185, "step": 584 }, { "epoch": 0.5826693227091634, "grad_norm": 0.726355254650116, "learning_rate": 4.43585634925063e-06, "loss": 2.656, "step": 585 }, { "epoch": 0.5836653386454184, "grad_norm": 0.8283027410507202, "learning_rate": 4.418575568705066e-06, "loss": 2.6963, "step": 586 }, { "epoch": 0.5846613545816733, "grad_norm": 2.0478458404541016, "learning_rate": 4.401301825638039e-06, "loss": 2.8958, "step": 587 }, { "epoch": 0.5856573705179283, "grad_norm": 0.9227387309074402, "learning_rate": 4.3840353291284776e-06, "loss": 2.5498, "step": 588 }, { "epoch": 0.5866533864541833, "grad_norm": 1.2917126417160034, "learning_rate": 4.366776288167598e-06, "loss": 2.4996, "step": 589 }, { "epoch": 0.5876494023904383, "grad_norm": 0.9034551978111267, "learning_rate": 4.349524911656373e-06, "loss": 2.4281, "step": 590 }, { "epoch": 0.5886454183266933, "grad_norm": 1.010878324508667, "learning_rate": 4.332281408403011e-06, "loss": 2.2912, "step": 591 }, { "epoch": 0.5896414342629482, "grad_norm": 1.7873375415802002, "learning_rate": 4.315045987120417e-06, "loss": 2.5171, "step": 592 }, { "epoch": 0.5906374501992032, "grad_norm": 0.8005262613296509, "learning_rate": 4.297818856423679e-06, "loss": 2.33, "step": 593 }, { "epoch": 0.5916334661354582, "grad_norm": 0.6728765368461609, "learning_rate": 4.280600224827533e-06, "loss": 2.1524, "step": 594 }, { "epoch": 0.5926294820717132, "grad_norm": 0.8610662221908569, "learning_rate": 4.2633903007438445e-06, "loss": 2.7524, "step": 595 }, { "epoch": 0.5936254980079682, "grad_norm": 1.0232973098754883, "learning_rate": 4.2461892924790825e-06, "loss": 2.6512, "step": 596 }, { "epoch": 0.5946215139442231, "grad_norm": 1.0610368251800537, "learning_rate": 4.228997408231806e-06, "loss": 2.4996, "step": 597 }, { "epoch": 0.5956175298804781, "grad_norm": 1.2796133756637573, "learning_rate": 4.2118148560901325e-06, "loss": 2.2488, "step": 598 }, { "epoch": 0.5966135458167331, "grad_norm": 1.5423349142074585, "learning_rate": 4.194641844029227e-06, "loss": 2.6293, "step": 599 }, { "epoch": 0.5976095617529881, "grad_norm": 1.5228114128112793, "learning_rate": 4.1774785799087805e-06, "loss": 2.3751, "step": 600 }, { "epoch": 0.598605577689243, "grad_norm": 0.9803175330162048, "learning_rate": 4.160325271470502e-06, "loss": 2.5003, "step": 601 }, { "epoch": 0.599601593625498, "grad_norm": 1.0139139890670776, "learning_rate": 4.143182126335594e-06, "loss": 2.5435, "step": 602 }, { "epoch": 0.600597609561753, "grad_norm": 0.8577011227607727, "learning_rate": 4.12604935200224e-06, "loss": 2.6227, "step": 603 }, { "epoch": 0.601593625498008, "grad_norm": 1.015549659729004, "learning_rate": 4.108927155843108e-06, "loss": 2.3803, "step": 604 }, { "epoch": 0.602589641434263, "grad_norm": 1.5675932168960571, "learning_rate": 4.091815745102818e-06, "loss": 2.364, "step": 605 }, { "epoch": 0.603585657370518, "grad_norm": 0.7266266345977783, "learning_rate": 4.074715326895453e-06, "loss": 2.6198, "step": 606 }, { "epoch": 0.6045816733067729, "grad_norm": 0.617874026298523, "learning_rate": 4.0576261082020345e-06, "loss": 2.6864, "step": 607 }, { "epoch": 0.6055776892430279, "grad_norm": 1.3024420738220215, "learning_rate": 4.040548295868039e-06, "loss": 2.4742, "step": 608 }, { "epoch": 0.6065737051792829, "grad_norm": 0.7109612822532654, "learning_rate": 4.023482096600873e-06, "loss": 2.6456, "step": 609 }, { "epoch": 0.6075697211155379, "grad_norm": 1.2119102478027344, "learning_rate": 4.006427716967382e-06, "loss": 2.3745, "step": 610 }, { "epoch": 0.6085657370517928, "grad_norm": 1.3014880418777466, "learning_rate": 3.9893853633913485e-06, "loss": 2.6845, "step": 611 }, { "epoch": 0.6095617529880478, "grad_norm": 1.3628534078598022, "learning_rate": 3.9723552421509975e-06, "loss": 2.6973, "step": 612 }, { "epoch": 0.6105577689243028, "grad_norm": 0.9777284860610962, "learning_rate": 3.955337559376489e-06, "loss": 2.4989, "step": 613 }, { "epoch": 0.6115537848605578, "grad_norm": 0.6070024371147156, "learning_rate": 3.938332521047434e-06, "loss": 2.0082, "step": 614 }, { "epoch": 0.6125498007968128, "grad_norm": 0.6223677396774292, "learning_rate": 3.921340332990392e-06, "loss": 2.2016, "step": 615 }, { "epoch": 0.6135458167330677, "grad_norm": 1.2076197862625122, "learning_rate": 3.904361200876391e-06, "loss": 2.7328, "step": 616 }, { "epoch": 0.6145418326693227, "grad_norm": 0.7502063512802124, "learning_rate": 3.887395330218429e-06, "loss": 2.1634, "step": 617 }, { "epoch": 0.6155378486055777, "grad_norm": 1.090084195137024, "learning_rate": 3.8704429263689865e-06, "loss": 2.2409, "step": 618 }, { "epoch": 0.6165338645418327, "grad_norm": 1.7830555438995361, "learning_rate": 3.853504194517551e-06, "loss": 2.5541, "step": 619 }, { "epoch": 0.6175298804780877, "grad_norm": 1.0715655088424683, "learning_rate": 3.836579339688116e-06, "loss": 2.7304, "step": 620 }, { "epoch": 0.6185258964143426, "grad_norm": 0.7255896925926208, "learning_rate": 3.819668566736719e-06, "loss": 2.5671, "step": 621 }, { "epoch": 0.6195219123505976, "grad_norm": 1.475665807723999, "learning_rate": 3.802772080348943e-06, "loss": 2.2374, "step": 622 }, { "epoch": 0.6205179282868526, "grad_norm": 1.1244341135025024, "learning_rate": 3.7858900850374596e-06, "loss": 2.2705, "step": 623 }, { "epoch": 0.6215139442231076, "grad_norm": 1.270950436592102, "learning_rate": 3.769022785139534e-06, "loss": 2.427, "step": 624 }, { "epoch": 0.6225099601593626, "grad_norm": 0.9996942281723022, "learning_rate": 3.752170384814562e-06, "loss": 2.3181, "step": 625 }, { "epoch": 0.6235059760956175, "grad_norm": 0.9702761173248291, "learning_rate": 3.7353330880415963e-06, "loss": 2.4871, "step": 626 }, { "epoch": 0.6245019920318725, "grad_norm": 0.7174897193908691, "learning_rate": 3.7185110986168842e-06, "loss": 2.6481, "step": 627 }, { "epoch": 0.6254980079681275, "grad_norm": 1.0198302268981934, "learning_rate": 3.701704620151389e-06, "loss": 2.4368, "step": 628 }, { "epoch": 0.6264940239043825, "grad_norm": 0.6317278742790222, "learning_rate": 3.6849138560683305e-06, "loss": 2.2506, "step": 629 }, { "epoch": 0.6274900398406374, "grad_norm": 1.6083205938339233, "learning_rate": 3.6681390096007315e-06, "loss": 2.441, "step": 630 }, { "epoch": 0.6284860557768924, "grad_norm": 1.1788543462753296, "learning_rate": 3.651380283788942e-06, "loss": 2.0867, "step": 631 }, { "epoch": 0.6294820717131474, "grad_norm": 1.6041985750198364, "learning_rate": 3.634637881478196e-06, "loss": 2.7786, "step": 632 }, { "epoch": 0.6304780876494024, "grad_norm": 0.7498704195022583, "learning_rate": 3.617912005316142e-06, "loss": 2.5885, "step": 633 }, { "epoch": 0.6314741035856574, "grad_norm": 1.2260042428970337, "learning_rate": 3.6012028577504106e-06, "loss": 2.5491, "step": 634 }, { "epoch": 0.6324701195219123, "grad_norm": 0.766639232635498, "learning_rate": 3.5845106410261417e-06, "loss": 2.6436, "step": 635 }, { "epoch": 0.6334661354581673, "grad_norm": 0.8522284626960754, "learning_rate": 3.56783555718355e-06, "loss": 2.361, "step": 636 }, { "epoch": 0.6344621513944223, "grad_norm": 1.09912109375, "learning_rate": 3.551177808055476e-06, "loss": 2.5303, "step": 637 }, { "epoch": 0.6354581673306773, "grad_norm": 1.4560422897338867, "learning_rate": 3.534537595264944e-06, "loss": 2.6122, "step": 638 }, { "epoch": 0.6364541832669323, "grad_norm": 0.858035147190094, "learning_rate": 3.5179151202227214e-06, "loss": 2.3591, "step": 639 }, { "epoch": 0.6374501992031872, "grad_norm": 4.184999942779541, "learning_rate": 3.5013105841248794e-06, "loss": 2.3339, "step": 640 }, { "epoch": 0.6384462151394422, "grad_norm": 1.2636277675628662, "learning_rate": 3.4847241879503574e-06, "loss": 2.6084, "step": 641 }, { "epoch": 0.6394422310756972, "grad_norm": 1.3735069036483765, "learning_rate": 3.4681561324585356e-06, "loss": 2.4582, "step": 642 }, { "epoch": 0.6404382470119522, "grad_norm": 1.3198506832122803, "learning_rate": 3.451606618186796e-06, "loss": 2.2207, "step": 643 }, { "epoch": 0.6414342629482072, "grad_norm": 0.895077109336853, "learning_rate": 3.435075845448105e-06, "loss": 2.141, "step": 644 }, { "epoch": 0.6424302788844621, "grad_norm": 1.5022435188293457, "learning_rate": 3.418564014328583e-06, "loss": 2.5608, "step": 645 }, { "epoch": 0.6434262948207171, "grad_norm": 0.9838452935218811, "learning_rate": 3.402071324685082e-06, "loss": 2.372, "step": 646 }, { "epoch": 0.6444223107569721, "grad_norm": 0.6322600841522217, "learning_rate": 3.3855979761427705e-06, "loss": 2.5499, "step": 647 }, { "epoch": 0.6454183266932271, "grad_norm": 1.3608890771865845, "learning_rate": 3.3691441680927105e-06, "loss": 2.4689, "step": 648 }, { "epoch": 0.646414342629482, "grad_norm": 0.9520907998085022, "learning_rate": 3.352710099689457e-06, "loss": 2.5012, "step": 649 }, { "epoch": 0.647410358565737, "grad_norm": 3.0419979095458984, "learning_rate": 3.3362959698486307e-06, "loss": 2.2773, "step": 650 }, { "epoch": 0.648406374501992, "grad_norm": 1.0915313959121704, "learning_rate": 3.3199019772445253e-06, "loss": 2.3744, "step": 651 }, { "epoch": 0.649402390438247, "grad_norm": 1.543050765991211, "learning_rate": 3.3035283203076877e-06, "loss": 2.3499, "step": 652 }, { "epoch": 0.650398406374502, "grad_norm": 1.0574357509613037, "learning_rate": 3.287175197222537e-06, "loss": 2.516, "step": 653 }, { "epoch": 0.651394422310757, "grad_norm": 1.3127410411834717, "learning_rate": 3.2708428059249437e-06, "loss": 2.4012, "step": 654 }, { "epoch": 0.6523904382470119, "grad_norm": 0.9456487894058228, "learning_rate": 3.254531344099847e-06, "loss": 2.773, "step": 655 }, { "epoch": 0.6533864541832669, "grad_norm": 1.1509116888046265, "learning_rate": 3.2382410091788567e-06, "loss": 2.7622, "step": 656 }, { "epoch": 0.6543824701195219, "grad_norm": 1.0328110456466675, "learning_rate": 3.221971998337872e-06, "loss": 2.5343, "step": 657 }, { "epoch": 0.6553784860557769, "grad_norm": 1.723029613494873, "learning_rate": 3.2057245084946796e-06, "loss": 2.8968, "step": 658 }, { "epoch": 0.6563745019920318, "grad_norm": 1.13263738155365, "learning_rate": 3.189498736306584e-06, "loss": 2.4689, "step": 659 }, { "epoch": 0.6573705179282868, "grad_norm": 1.4411126375198364, "learning_rate": 3.173294878168025e-06, "loss": 2.0715, "step": 660 }, { "epoch": 0.6583665338645418, "grad_norm": 2.4463119506835938, "learning_rate": 3.1571131302081916e-06, "loss": 2.5004, "step": 661 }, { "epoch": 0.6593625498007968, "grad_norm": 1.063270926475525, "learning_rate": 3.140953688288658e-06, "loss": 2.4079, "step": 662 }, { "epoch": 0.6603585657370518, "grad_norm": 1.860757827758789, "learning_rate": 3.1248167480010083e-06, "loss": 2.6755, "step": 663 }, { "epoch": 0.6613545816733067, "grad_norm": 1.5435043573379517, "learning_rate": 3.1087025046644704e-06, "loss": 2.7499, "step": 664 }, { "epoch": 0.6623505976095617, "grad_norm": 0.7970728278160095, "learning_rate": 3.0926111533235526e-06, "loss": 2.1911, "step": 665 }, { "epoch": 0.6633466135458167, "grad_norm": 1.1135482788085938, "learning_rate": 3.0765428887456794e-06, "loss": 2.6387, "step": 666 }, { "epoch": 0.6643426294820717, "grad_norm": 1.2876728773117065, "learning_rate": 3.0604979054188367e-06, "loss": 2.6715, "step": 667 }, { "epoch": 0.6653386454183267, "grad_norm": 0.6579734683036804, "learning_rate": 3.044476397549221e-06, "loss": 2.1833, "step": 668 }, { "epoch": 0.6663346613545816, "grad_norm": 1.7546638250350952, "learning_rate": 3.0284785590588804e-06, "loss": 2.5761, "step": 669 }, { "epoch": 0.6673306772908366, "grad_norm": 1.1617887020111084, "learning_rate": 3.012504583583374e-06, "loss": 2.4205, "step": 670 }, { "epoch": 0.6683266932270916, "grad_norm": 1.4457294940948486, "learning_rate": 2.9965546644694287e-06, "loss": 2.178, "step": 671 }, { "epoch": 0.6693227091633466, "grad_norm": 0.9334515333175659, "learning_rate": 2.9806289947725947e-06, "loss": 2.5343, "step": 672 }, { "epoch": 0.6703187250996016, "grad_norm": 1.115212082862854, "learning_rate": 2.9647277672549093e-06, "loss": 2.1731, "step": 673 }, { "epoch": 0.6713147410358565, "grad_norm": 1.1038217544555664, "learning_rate": 2.948851174382565e-06, "loss": 2.3589, "step": 674 }, { "epoch": 0.6723107569721115, "grad_norm": 1.4897500276565552, "learning_rate": 2.9329994083235857e-06, "loss": 2.4302, "step": 675 }, { "epoch": 0.6733067729083665, "grad_norm": 1.7196754217147827, "learning_rate": 2.9171726609454875e-06, "loss": 2.5387, "step": 676 }, { "epoch": 0.6743027888446215, "grad_norm": 1.271872878074646, "learning_rate": 2.9013711238129693e-06, "loss": 2.1938, "step": 677 }, { "epoch": 0.6752988047808764, "grad_norm": 1.0383085012435913, "learning_rate": 2.885594988185587e-06, "loss": 2.5842, "step": 678 }, { "epoch": 0.6762948207171314, "grad_norm": 1.9233471155166626, "learning_rate": 2.8698444450154395e-06, "loss": 2.4421, "step": 679 }, { "epoch": 0.6772908366533864, "grad_norm": 0.9812890291213989, "learning_rate": 2.8541196849448582e-06, "loss": 2.1876, "step": 680 }, { "epoch": 0.6782868525896414, "grad_norm": 1.217011570930481, "learning_rate": 2.8384208983040997e-06, "loss": 2.5115, "step": 681 }, { "epoch": 0.6792828685258964, "grad_norm": 1.4093648195266724, "learning_rate": 2.8227482751090445e-06, "loss": 2.5296, "step": 682 }, { "epoch": 0.6802788844621513, "grad_norm": 1.0479772090911865, "learning_rate": 2.8071020050588927e-06, "loss": 2.3801, "step": 683 }, { "epoch": 0.6812749003984063, "grad_norm": 0.761779248714447, "learning_rate": 2.7914822775338678e-06, "loss": 2.397, "step": 684 }, { "epoch": 0.6822709163346613, "grad_norm": 0.7536188364028931, "learning_rate": 2.775889281592927e-06, "loss": 2.2802, "step": 685 }, { "epoch": 0.6832669322709163, "grad_norm": 1.1621276140213013, "learning_rate": 2.760323205971476e-06, "loss": 2.3802, "step": 686 }, { "epoch": 0.6842629482071713, "grad_norm": 1.2401965856552124, "learning_rate": 2.744784239079077e-06, "loss": 1.9567, "step": 687 }, { "epoch": 0.6852589641434262, "grad_norm": 0.9456545114517212, "learning_rate": 2.7292725689971732e-06, "loss": 2.547, "step": 688 }, { "epoch": 0.6862549800796812, "grad_norm": 1.3253943920135498, "learning_rate": 2.7137883834768076e-06, "loss": 2.2105, "step": 689 }, { "epoch": 0.6872509960159362, "grad_norm": 1.525397777557373, "learning_rate": 2.6983318699363627e-06, "loss": 2.3682, "step": 690 }, { "epoch": 0.6882470119521913, "grad_norm": 0.9517590403556824, "learning_rate": 2.6829032154592745e-06, "loss": 2.2159, "step": 691 }, { "epoch": 0.6892430278884463, "grad_norm": 0.8040021061897278, "learning_rate": 2.6675026067917808e-06, "loss": 2.3967, "step": 692 }, { "epoch": 0.6902390438247012, "grad_norm": 1.6833242177963257, "learning_rate": 2.652130230340655e-06, "loss": 2.9864, "step": 693 }, { "epoch": 0.6912350597609562, "grad_norm": 1.104771614074707, "learning_rate": 2.636786272170956e-06, "loss": 2.5124, "step": 694 }, { "epoch": 0.6922310756972112, "grad_norm": 1.1763907670974731, "learning_rate": 2.621470918003768e-06, "loss": 2.6426, "step": 695 }, { "epoch": 0.6932270916334662, "grad_norm": 2.304222345352173, "learning_rate": 2.6061843532139563e-06, "loss": 2.4522, "step": 696 }, { "epoch": 0.6942231075697212, "grad_norm": 0.8599796891212463, "learning_rate": 2.5909267628279234e-06, "loss": 2.6796, "step": 697 }, { "epoch": 0.6952191235059761, "grad_norm": 1.0061733722686768, "learning_rate": 2.5756983315213748e-06, "loss": 2.5076, "step": 698 }, { "epoch": 0.6962151394422311, "grad_norm": 1.392606258392334, "learning_rate": 2.560499243617074e-06, "loss": 2.5134, "step": 699 }, { "epoch": 0.6972111553784861, "grad_norm": 1.2116351127624512, "learning_rate": 2.5453296830826135e-06, "loss": 2.0634, "step": 700 }, { "epoch": 0.6982071713147411, "grad_norm": 0.7071558237075806, "learning_rate": 2.5301898335281994e-06, "loss": 2.1104, "step": 701 }, { "epoch": 0.6992031872509961, "grad_norm": 1.8307946920394897, "learning_rate": 2.5150798782044123e-06, "loss": 2.8147, "step": 702 }, { "epoch": 0.700199203187251, "grad_norm": 0.9716182351112366, "learning_rate": 2.5000000000000015e-06, "loss": 2.4836, "step": 703 }, { "epoch": 0.701195219123506, "grad_norm": 0.7655389308929443, "learning_rate": 2.4849503814396624e-06, "loss": 2.2803, "step": 704 }, { "epoch": 0.702191235059761, "grad_norm": 1.1354485750198364, "learning_rate": 2.469931204681841e-06, "loss": 2.5936, "step": 705 }, { "epoch": 0.703187250996016, "grad_norm": 0.9272159337997437, "learning_rate": 2.4549426515165116e-06, "loss": 2.6629, "step": 706 }, { "epoch": 0.704183266932271, "grad_norm": 1.20318603515625, "learning_rate": 2.439984903362988e-06, "loss": 2.341, "step": 707 }, { "epoch": 0.7051792828685259, "grad_norm": 1.0813405513763428, "learning_rate": 2.425058141267722e-06, "loss": 2.5484, "step": 708 }, { "epoch": 0.7061752988047809, "grad_norm": 0.6365978121757507, "learning_rate": 2.4101625459021212e-06, "loss": 2.2276, "step": 709 }, { "epoch": 0.7071713147410359, "grad_norm": 1.4600951671600342, "learning_rate": 2.3952982975603494e-06, "loss": 2.7489, "step": 710 }, { "epoch": 0.7081673306772909, "grad_norm": 1.0905722379684448, "learning_rate": 2.3804655761571517e-06, "loss": 2.7045, "step": 711 }, { "epoch": 0.7091633466135459, "grad_norm": 1.2118492126464844, "learning_rate": 2.3656645612256747e-06, "loss": 2.2625, "step": 712 }, { "epoch": 0.7101593625498008, "grad_norm": 2.2730562686920166, "learning_rate": 2.3508954319153e-06, "loss": 2.9233, "step": 713 }, { "epoch": 0.7111553784860558, "grad_norm": 1.1532260179519653, "learning_rate": 2.3361583669894634e-06, "loss": 2.6882, "step": 714 }, { "epoch": 0.7121513944223108, "grad_norm": 1.74001944065094, "learning_rate": 2.321453544823499e-06, "loss": 2.296, "step": 715 }, { "epoch": 0.7131474103585658, "grad_norm": 1.6281747817993164, "learning_rate": 2.306781143402485e-06, "loss": 2.5453, "step": 716 }, { "epoch": 0.7141434262948207, "grad_norm": 1.2188794612884521, "learning_rate": 2.2921413403190774e-06, "loss": 2.3351, "step": 717 }, { "epoch": 0.7151394422310757, "grad_norm": 1.180245280265808, "learning_rate": 2.2775343127713685e-06, "loss": 2.8909, "step": 718 }, { "epoch": 0.7161354581673307, "grad_norm": 1.207853078842163, "learning_rate": 2.2629602375607373e-06, "loss": 2.2249, "step": 719 }, { "epoch": 0.7171314741035857, "grad_norm": 1.1911535263061523, "learning_rate": 2.24841929108972e-06, "loss": 1.9728, "step": 720 }, { "epoch": 0.7181274900398407, "grad_norm": 2.913970470428467, "learning_rate": 2.23391164935986e-06, "loss": 3.3, "step": 721 }, { "epoch": 0.7191235059760956, "grad_norm": 1.0136604309082031, "learning_rate": 2.219437487969588e-06, "loss": 2.3078, "step": 722 }, { "epoch": 0.7201195219123506, "grad_norm": 1.5370888710021973, "learning_rate": 2.20499698211209e-06, "loss": 2.4226, "step": 723 }, { "epoch": 0.7211155378486056, "grad_norm": 0.8609825968742371, "learning_rate": 2.190590306573198e-06, "loss": 2.5837, "step": 724 }, { "epoch": 0.7221115537848606, "grad_norm": 0.7969903945922852, "learning_rate": 2.1762176357292582e-06, "loss": 2.5065, "step": 725 }, { "epoch": 0.7231075697211156, "grad_norm": 0.6895061135292053, "learning_rate": 2.1618791435450334e-06, "loss": 2.4443, "step": 726 }, { "epoch": 0.7241035856573705, "grad_norm": 1.005803108215332, "learning_rate": 2.1475750035715914e-06, "loss": 2.8449, "step": 727 }, { "epoch": 0.7250996015936255, "grad_norm": 1.464055061340332, "learning_rate": 2.1333053889442033e-06, "loss": 2.436, "step": 728 }, { "epoch": 0.7260956175298805, "grad_norm": 0.7166134715080261, "learning_rate": 2.1190704723802587e-06, "loss": 2.6141, "step": 729 }, { "epoch": 0.7270916334661355, "grad_norm": 1.2269198894500732, "learning_rate": 2.104870426177157e-06, "loss": 2.0039, "step": 730 }, { "epoch": 0.7280876494023905, "grad_norm": 1.233473539352417, "learning_rate": 2.0907054222102367e-06, "loss": 2.4503, "step": 731 }, { "epoch": 0.7290836653386454, "grad_norm": 0.8751947283744812, "learning_rate": 2.0765756319306897e-06, "loss": 2.482, "step": 732 }, { "epoch": 0.7300796812749004, "grad_norm": 1.954285979270935, "learning_rate": 2.0624812263634847e-06, "loss": 2.6237, "step": 733 }, { "epoch": 0.7310756972111554, "grad_norm": 0.7244362235069275, "learning_rate": 2.048422376105299e-06, "loss": 2.3371, "step": 734 }, { "epoch": 0.7320717131474104, "grad_norm": 0.7712534666061401, "learning_rate": 2.034399251322458e-06, "loss": 2.6775, "step": 735 }, { "epoch": 0.7330677290836654, "grad_norm": 1.0466793775558472, "learning_rate": 2.020412021748866e-06, "loss": 2.2773, "step": 736 }, { "epoch": 0.7340637450199203, "grad_norm": 1.417794942855835, "learning_rate": 2.0064608566839584e-06, "loss": 2.7359, "step": 737 }, { "epoch": 0.7350597609561753, "grad_norm": 0.5706871747970581, "learning_rate": 1.9925459249906488e-06, "loss": 2.246, "step": 738 }, { "epoch": 0.7360557768924303, "grad_norm": 1.112219214439392, "learning_rate": 1.978667395093293e-06, "loss": 2.5444, "step": 739 }, { "epoch": 0.7370517928286853, "grad_norm": 1.5537924766540527, "learning_rate": 1.964825434975639e-06, "loss": 2.2497, "step": 740 }, { "epoch": 0.7380478087649402, "grad_norm": 0.7418034672737122, "learning_rate": 1.9510202121788003e-06, "loss": 2.4711, "step": 741 }, { "epoch": 0.7390438247011952, "grad_norm": 1.0376439094543457, "learning_rate": 1.9372518937992306e-06, "loss": 2.1369, "step": 742 }, { "epoch": 0.7400398406374502, "grad_norm": 1.6938295364379883, "learning_rate": 1.923520646486695e-06, "loss": 2.7013, "step": 743 }, { "epoch": 0.7410358565737052, "grad_norm": 1.1227657794952393, "learning_rate": 1.9098266364422554e-06, "loss": 2.1956, "step": 744 }, { "epoch": 0.7420318725099602, "grad_norm": 0.8521560430526733, "learning_rate": 1.8961700294162578e-06, "loss": 2.7621, "step": 745 }, { "epoch": 0.7430278884462151, "grad_norm": 1.3367222547531128, "learning_rate": 1.8825509907063328e-06, "loss": 2.3669, "step": 746 }, { "epoch": 0.7440239043824701, "grad_norm": 1.0971968173980713, "learning_rate": 1.8689696851553847e-06, "loss": 2.2727, "step": 747 }, { "epoch": 0.7450199203187251, "grad_norm": 0.7232230305671692, "learning_rate": 1.8554262771496017e-06, "loss": 2.4247, "step": 748 }, { "epoch": 0.7460159362549801, "grad_norm": 0.779901921749115, "learning_rate": 1.8419209306164653e-06, "loss": 2.4956, "step": 749 }, { "epoch": 0.7470119521912351, "grad_norm": 0.9150820970535278, "learning_rate": 1.82845380902277e-06, "loss": 2.6319, "step": 750 }, { "epoch": 0.74800796812749, "grad_norm": 1.0264116525650024, "learning_rate": 1.8150250753726363e-06, "loss": 2.537, "step": 751 }, { "epoch": 0.749003984063745, "grad_norm": 1.3325294256210327, "learning_rate": 1.8016348922055448e-06, "loss": 2.5891, "step": 752 }, { "epoch": 0.75, "grad_norm": 0.9217858910560608, "learning_rate": 1.7882834215943645e-06, "loss": 2.3572, "step": 753 }, { "epoch": 0.750996015936255, "grad_norm": 1.020738959312439, "learning_rate": 1.7749708251433983e-06, "loss": 2.4734, "step": 754 }, { "epoch": 0.75199203187251, "grad_norm": 0.9455721378326416, "learning_rate": 1.7616972639864166e-06, "loss": 2.4533, "step": 755 }, { "epoch": 0.7529880478087649, "grad_norm": 1.7625263929367065, "learning_rate": 1.7484628987847125e-06, "loss": 2.5292, "step": 756 }, { "epoch": 0.7539840637450199, "grad_norm": 1.2456424236297607, "learning_rate": 1.7352678897251606e-06, "loss": 2.5379, "step": 757 }, { "epoch": 0.7549800796812749, "grad_norm": 1.9081121683120728, "learning_rate": 1.7221123965182712e-06, "loss": 2.413, "step": 758 }, { "epoch": 0.7559760956175299, "grad_norm": 1.062225341796875, "learning_rate": 1.7089965783962608e-06, "loss": 2.458, "step": 759 }, { "epoch": 0.7569721115537849, "grad_norm": 1.1116987466812134, "learning_rate": 1.6959205941111228e-06, "loss": 2.4556, "step": 760 }, { "epoch": 0.7579681274900398, "grad_norm": 1.6234967708587646, "learning_rate": 1.6828846019327128e-06, "loss": 2.5499, "step": 761 }, { "epoch": 0.7589641434262948, "grad_norm": 0.741877555847168, "learning_rate": 1.6698887596468232e-06, "loss": 2.1629, "step": 762 }, { "epoch": 0.7599601593625498, "grad_norm": 2.112726926803589, "learning_rate": 1.6569332245532777e-06, "loss": 2.85, "step": 763 }, { "epoch": 0.7609561752988048, "grad_norm": 0.9073076844215393, "learning_rate": 1.6440181534640277e-06, "loss": 2.4195, "step": 764 }, { "epoch": 0.7619521912350598, "grad_norm": 1.2635924816131592, "learning_rate": 1.6311437027012582e-06, "loss": 2.3853, "step": 765 }, { "epoch": 0.7629482071713147, "grad_norm": 1.6688510179519653, "learning_rate": 1.618310028095486e-06, "loss": 2.186, "step": 766 }, { "epoch": 0.7639442231075697, "grad_norm": 1.076957106590271, "learning_rate": 1.6055172849836826e-06, "loss": 2.3887, "step": 767 }, { "epoch": 0.7649402390438247, "grad_norm": 1.0081124305725098, "learning_rate": 1.5927656282073861e-06, "loss": 2.4315, "step": 768 }, { "epoch": 0.7659362549800797, "grad_norm": 1.2986465692520142, "learning_rate": 1.5800552121108392e-06, "loss": 2.6633, "step": 769 }, { "epoch": 0.7669322709163346, "grad_norm": 0.7207338213920593, "learning_rate": 1.567386190539107e-06, "loss": 2.6924, "step": 770 }, { "epoch": 0.7679282868525896, "grad_norm": 0.6458574533462524, "learning_rate": 1.5547587168362204e-06, "loss": 2.7688, "step": 771 }, { "epoch": 0.7689243027888446, "grad_norm": 1.0633124113082886, "learning_rate": 1.5421729438433274e-06, "loss": 2.1328, "step": 772 }, { "epoch": 0.7699203187250996, "grad_norm": 2.143666982650757, "learning_rate": 1.5296290238968303e-06, "loss": 2.29, "step": 773 }, { "epoch": 0.7709163346613546, "grad_norm": 0.5651401281356812, "learning_rate": 1.517127108826551e-06, "loss": 2.4732, "step": 774 }, { "epoch": 0.7719123505976095, "grad_norm": 0.8489325642585754, "learning_rate": 1.5046673499538893e-06, "loss": 2.3174, "step": 775 }, { "epoch": 0.7729083665338645, "grad_norm": 1.1251336336135864, "learning_rate": 1.4922498980899907e-06, "loss": 2.2915, "step": 776 }, { "epoch": 0.7739043824701195, "grad_norm": 0.7484387755393982, "learning_rate": 1.4798749035339278e-06, "loss": 2.3685, "step": 777 }, { "epoch": 0.7749003984063745, "grad_norm": 1.1463130712509155, "learning_rate": 1.4675425160708723e-06, "loss": 2.468, "step": 778 }, { "epoch": 0.7758964143426295, "grad_norm": 1.5645790100097656, "learning_rate": 1.4552528849702852e-06, "loss": 2.6442, "step": 779 }, { "epoch": 0.7768924302788844, "grad_norm": 1.8811829090118408, "learning_rate": 1.4430061589841122e-06, "loss": 2.5609, "step": 780 }, { "epoch": 0.7778884462151394, "grad_norm": 0.8737534284591675, "learning_rate": 1.4308024863449805e-06, "loss": 2.6824, "step": 781 }, { "epoch": 0.7788844621513944, "grad_norm": 1.1957892179489136, "learning_rate": 1.4186420147644053e-06, "loss": 2.3529, "step": 782 }, { "epoch": 0.7798804780876494, "grad_norm": 1.2302711009979248, "learning_rate": 1.4065248914310066e-06, "loss": 2.513, "step": 783 }, { "epoch": 0.7808764940239044, "grad_norm": 0.5240752100944519, "learning_rate": 1.3944512630087182e-06, "loss": 2.4043, "step": 784 }, { "epoch": 0.7818725099601593, "grad_norm": 1.9195410013198853, "learning_rate": 1.3824212756350196e-06, "loss": 2.8095, "step": 785 }, { "epoch": 0.7828685258964143, "grad_norm": 0.9604887962341309, "learning_rate": 1.3704350749191642e-06, "loss": 2.3252, "step": 786 }, { "epoch": 0.7838645418326693, "grad_norm": 0.9721193313598633, "learning_rate": 1.3584928059404207e-06, "loss": 2.4578, "step": 787 }, { "epoch": 0.7848605577689243, "grad_norm": 1.9194726943969727, "learning_rate": 1.3465946132463125e-06, "loss": 2.623, "step": 788 }, { "epoch": 0.7858565737051793, "grad_norm": 1.482784390449524, "learning_rate": 1.3347406408508695e-06, "loss": 2.7708, "step": 789 }, { "epoch": 0.7868525896414342, "grad_norm": 0.7451381683349609, "learning_rate": 1.3229310322328847e-06, "loss": 2.4386, "step": 790 }, { "epoch": 0.7878486055776892, "grad_norm": 0.6679832339286804, "learning_rate": 1.3111659303341824e-06, "loss": 2.37, "step": 791 }, { "epoch": 0.7888446215139442, "grad_norm": 0.8974138498306274, "learning_rate": 1.2994454775578785e-06, "loss": 2.2855, "step": 792 }, { "epoch": 0.7898406374501992, "grad_norm": 1.3459084033966064, "learning_rate": 1.2877698157666663e-06, "loss": 2.7191, "step": 793 }, { "epoch": 0.7908366533864541, "grad_norm": 1.0975403785705566, "learning_rate": 1.2761390862810907e-06, "loss": 2.2521, "step": 794 }, { "epoch": 0.7918326693227091, "grad_norm": 0.9908530712127686, "learning_rate": 1.2645534298778506e-06, "loss": 2.603, "step": 795 }, { "epoch": 0.7928286852589641, "grad_norm": 0.7879658937454224, "learning_rate": 1.253012986788078e-06, "loss": 2.6744, "step": 796 }, { "epoch": 0.7938247011952191, "grad_norm": 0.9611647129058838, "learning_rate": 1.2415178966956531e-06, "loss": 2.3191, "step": 797 }, { "epoch": 0.7948207171314741, "grad_norm": 2.035386085510254, "learning_rate": 1.2300682987355122e-06, "loss": 2.46, "step": 798 }, { "epoch": 0.795816733067729, "grad_norm": 1.4089851379394531, "learning_rate": 1.2186643314919571e-06, "loss": 2.164, "step": 799 }, { "epoch": 0.796812749003984, "grad_norm": 0.944324254989624, "learning_rate": 1.2073061329969843e-06, "loss": 2.4692, "step": 800 }, { "epoch": 0.797808764940239, "grad_norm": 1.0982811450958252, "learning_rate": 1.1959938407286099e-06, "loss": 2.2721, "step": 801 }, { "epoch": 0.798804780876494, "grad_norm": 0.8596687316894531, "learning_rate": 1.1847275916092116e-06, "loss": 2.3065, "step": 802 }, { "epoch": 0.799800796812749, "grad_norm": 0.8045834898948669, "learning_rate": 1.1735075220038634e-06, "loss": 2.1781, "step": 803 }, { "epoch": 0.8007968127490039, "grad_norm": 1.5618336200714111, "learning_rate": 1.1623337677186902e-06, "loss": 2.4166, "step": 804 }, { "epoch": 0.8017928286852589, "grad_norm": 1.3380889892578125, "learning_rate": 1.151206463999222e-06, "loss": 2.5112, "step": 805 }, { "epoch": 0.8027888446215139, "grad_norm": 1.37197744846344, "learning_rate": 1.1401257455287612e-06, "loss": 2.4657, "step": 806 }, { "epoch": 0.8037848605577689, "grad_norm": 1.3427671194076538, "learning_rate": 1.1290917464267458e-06, "loss": 2.5771, "step": 807 }, { "epoch": 0.8047808764940239, "grad_norm": 0.5480353832244873, "learning_rate": 1.1181046002471292e-06, "loss": 2.2615, "step": 808 }, { "epoch": 0.8057768924302788, "grad_norm": 1.03799569606781, "learning_rate": 1.107164439976764e-06, "loss": 2.1258, "step": 809 }, { "epoch": 0.8067729083665338, "grad_norm": 0.874638020992279, "learning_rate": 1.0962713980337947e-06, "loss": 2.3982, "step": 810 }, { "epoch": 0.8077689243027888, "grad_norm": 0.7668205499649048, "learning_rate": 1.085425606266049e-06, "loss": 2.5981, "step": 811 }, { "epoch": 0.8087649402390438, "grad_norm": 1.033339262008667, "learning_rate": 1.0746271959494453e-06, "loss": 2.425, "step": 812 }, { "epoch": 0.8097609561752988, "grad_norm": 0.7686687707901001, "learning_rate": 1.063876297786407e-06, "loss": 2.287, "step": 813 }, { "epoch": 0.8107569721115537, "grad_norm": 0.8868098855018616, "learning_rate": 1.0531730419042736e-06, "loss": 2.4047, "step": 814 }, { "epoch": 0.8117529880478087, "grad_norm": 0.9515554308891296, "learning_rate": 1.04251755785373e-06, "loss": 2.2864, "step": 815 }, { "epoch": 0.8127490039840638, "grad_norm": 0.96048903465271, "learning_rate": 1.0319099746072375e-06, "loss": 2.305, "step": 816 }, { "epoch": 0.8137450199203188, "grad_norm": 1.2892875671386719, "learning_rate": 1.0213504205574758e-06, "loss": 2.8046, "step": 817 }, { "epoch": 0.8147410358565738, "grad_norm": 1.2894792556762695, "learning_rate": 1.0108390235157828e-06, "loss": 2.3662, "step": 818 }, { "epoch": 0.8157370517928287, "grad_norm": 0.894437849521637, "learning_rate": 1.0003759107106116e-06, "loss": 2.3213, "step": 819 }, { "epoch": 0.8167330677290837, "grad_norm": 0.8480390906333923, "learning_rate": 9.899612087859883e-07, "loss": 2.2743, "step": 820 }, { "epoch": 0.8177290836653387, "grad_norm": 0.6957425475120544, "learning_rate": 9.795950437999852e-07, "loss": 2.6014, "step": 821 }, { "epoch": 0.8187250996015937, "grad_norm": 1.1951571702957153, "learning_rate": 9.692775412231863e-07, "loss": 2.5359, "step": 822 }, { "epoch": 0.8197211155378487, "grad_norm": 1.5518149137496948, "learning_rate": 9.590088259371738e-07, "loss": 2.5717, "step": 823 }, { "epoch": 0.8207171314741036, "grad_norm": 0.9850301146507263, "learning_rate": 9.487890222330137e-07, "loss": 2.3225, "step": 824 }, { "epoch": 0.8217131474103586, "grad_norm": 1.0830625295639038, "learning_rate": 9.386182538097582e-07, "loss": 2.49, "step": 825 }, { "epoch": 0.8227091633466136, "grad_norm": 1.1903777122497559, "learning_rate": 9.284966437729387e-07, "loss": 2.3532, "step": 826 }, { "epoch": 0.8237051792828686, "grad_norm": 1.2485320568084717, "learning_rate": 9.184243146330829e-07, "loss": 2.286, "step": 827 }, { "epoch": 0.8247011952191236, "grad_norm": 1.5328834056854248, "learning_rate": 9.084013883042276e-07, "loss": 2.5148, "step": 828 }, { "epoch": 0.8256972111553785, "grad_norm": 1.7866473197937012, "learning_rate": 8.984279861024453e-07, "loss": 2.7636, "step": 829 }, { "epoch": 0.8266932270916335, "grad_norm": 0.8635814785957336, "learning_rate": 8.885042287443785e-07, "loss": 2.579, "step": 830 }, { "epoch": 0.8276892430278885, "grad_norm": 1.480765461921692, "learning_rate": 8.786302363457733e-07, "loss": 2.6228, "step": 831 }, { "epoch": 0.8286852589641435, "grad_norm": 0.6803283095359802, "learning_rate": 8.688061284200266e-07, "loss": 2.4377, "step": 832 }, { "epoch": 0.8296812749003984, "grad_norm": 1.2872114181518555, "learning_rate": 8.590320238767425e-07, "loss": 2.4269, "step": 833 }, { "epoch": 0.8306772908366534, "grad_norm": 1.054971694946289, "learning_rate": 8.493080410202914e-07, "loss": 2.4407, "step": 834 }, { "epoch": 0.8316733067729084, "grad_norm": 0.8980826735496521, "learning_rate": 8.396342975483751e-07, "loss": 2.352, "step": 835 }, { "epoch": 0.8326693227091634, "grad_norm": 0.7561918497085571, "learning_rate": 8.30010910550611e-07, "loss": 2.3562, "step": 836 }, { "epoch": 0.8336653386454184, "grad_norm": 0.6130694150924683, "learning_rate": 8.204379965071036e-07, "loss": 2.3059, "step": 837 }, { "epoch": 0.8346613545816733, "grad_norm": 1.1262505054473877, "learning_rate": 8.109156712870397e-07, "loss": 2.7065, "step": 838 }, { "epoch": 0.8356573705179283, "grad_norm": 2.2383527755737305, "learning_rate": 8.014440501472909e-07, "loss": 2.6112, "step": 839 }, { "epoch": 0.8366533864541833, "grad_norm": 1.0032474994659424, "learning_rate": 7.920232477310102e-07, "loss": 2.4155, "step": 840 }, { "epoch": 0.8376494023904383, "grad_norm": 1.5110204219818115, "learning_rate": 7.826533780662481e-07, "loss": 2.2312, "step": 841 }, { "epoch": 0.8386454183266933, "grad_norm": 1.206811785697937, "learning_rate": 7.733345545645726e-07, "loss": 2.7882, "step": 842 }, { "epoch": 0.8396414342629482, "grad_norm": 1.1954952478408813, "learning_rate": 7.640668900196985e-07, "loss": 2.5765, "step": 843 }, { "epoch": 0.8406374501992032, "grad_norm": 0.8079789280891418, "learning_rate": 7.54850496606117e-07, "loss": 2.0988, "step": 844 }, { "epoch": 0.8416334661354582, "grad_norm": 1.204300880432129, "learning_rate": 7.456854858777418e-07, "loss": 2.6726, "step": 845 }, { "epoch": 0.8426294820717132, "grad_norm": 2.1050732135772705, "learning_rate": 7.365719687665568e-07, "loss": 2.4657, "step": 846 }, { "epoch": 0.8436254980079682, "grad_norm": 1.1028344631195068, "learning_rate": 7.27510055581278e-07, "loss": 2.7165, "step": 847 }, { "epoch": 0.8446215139442231, "grad_norm": 0.9454997777938843, "learning_rate": 7.184998560060114e-07, "loss": 2.4538, "step": 848 }, { "epoch": 0.8456175298804781, "grad_norm": 1.000157117843628, "learning_rate": 7.095414790989292e-07, "loss": 2.2186, "step": 849 }, { "epoch": 0.8466135458167331, "grad_norm": 1.7517778873443604, "learning_rate": 7.006350332909495e-07, "loss": 2.4932, "step": 850 }, { "epoch": 0.8476095617529881, "grad_norm": 1.2720731496810913, "learning_rate": 6.917806263844268e-07, "loss": 2.287, "step": 851 }, { "epoch": 0.848605577689243, "grad_norm": 1.1185457706451416, "learning_rate": 6.829783655518402e-07, "loss": 2.4009, "step": 852 }, { "epoch": 0.849601593625498, "grad_norm": 1.228535771369934, "learning_rate": 6.742283573345004e-07, "loss": 2.4729, "step": 853 }, { "epoch": 0.850597609561753, "grad_norm": 1.3845924139022827, "learning_rate": 6.655307076412637e-07, "loss": 2.5723, "step": 854 }, { "epoch": 0.851593625498008, "grad_norm": 1.2278180122375488, "learning_rate": 6.568855217472425e-07, "loss": 2.4958, "step": 855 }, { "epoch": 0.852589641434263, "grad_norm": 0.7195264101028442, "learning_rate": 6.482929042925363e-07, "loss": 2.5879, "step": 856 }, { "epoch": 0.853585657370518, "grad_norm": 1.1685850620269775, "learning_rate": 6.397529592809615e-07, "loss": 2.3411, "step": 857 }, { "epoch": 0.8545816733067729, "grad_norm": 0.935212254524231, "learning_rate": 6.312657900788e-07, "loss": 2.3266, "step": 858 }, { "epoch": 0.8555776892430279, "grad_norm": 1.0523936748504639, "learning_rate": 6.228314994135376e-07, "loss": 2.5915, "step": 859 }, { "epoch": 0.8565737051792829, "grad_norm": 0.7740164399147034, "learning_rate": 6.14450189372628e-07, "loss": 2.7521, "step": 860 }, { "epoch": 0.8575697211155379, "grad_norm": 2.0991780757904053, "learning_rate": 6.061219614022535e-07, "loss": 2.2285, "step": 861 }, { "epoch": 0.8585657370517928, "grad_norm": 0.990088164806366, "learning_rate": 5.978469163061018e-07, "loss": 2.3091, "step": 862 }, { "epoch": 0.8595617529880478, "grad_norm": 1.3237099647521973, "learning_rate": 5.896251542441395e-07, "loss": 2.4856, "step": 863 }, { "epoch": 0.8605577689243028, "grad_norm": 0.9690184593200684, "learning_rate": 5.814567747314049e-07, "loss": 2.3608, "step": 864 }, { "epoch": 0.8615537848605578, "grad_norm": 2.3030495643615723, "learning_rate": 5.733418766367988e-07, "loss": 2.4188, "step": 865 }, { "epoch": 0.8625498007968128, "grad_norm": 1.1400495767593384, "learning_rate": 5.652805581818943e-07, "loss": 2.2339, "step": 866 }, { "epoch": 0.8635458167330677, "grad_norm": 1.5023630857467651, "learning_rate": 5.572729169397422e-07, "loss": 2.1393, "step": 867 }, { "epoch": 0.8645418326693227, "grad_norm": 0.9161491394042969, "learning_rate": 5.493190498336903e-07, "loss": 2.3602, "step": 868 }, { "epoch": 0.8655378486055777, "grad_norm": 0.9129965901374817, "learning_rate": 5.414190531362162e-07, "loss": 2.3639, "step": 869 }, { "epoch": 0.8665338645418327, "grad_norm": 1.9090954065322876, "learning_rate": 5.335730224677538e-07, "loss": 2.2505, "step": 870 }, { "epoch": 0.8675298804780877, "grad_norm": 0.8173048496246338, "learning_rate": 5.25781052795541e-07, "loss": 2.2072, "step": 871 }, { "epoch": 0.8685258964143426, "grad_norm": 2.2648603916168213, "learning_rate": 5.180432384324691e-07, "loss": 2.458, "step": 872 }, { "epoch": 0.8695219123505976, "grad_norm": 0.906757652759552, "learning_rate": 5.103596730359428e-07, "loss": 2.5185, "step": 873 }, { "epoch": 0.8705179282868526, "grad_norm": 0.9674282073974609, "learning_rate": 5.027304496067431e-07, "loss": 2.3208, "step": 874 }, { "epoch": 0.8715139442231076, "grad_norm": 1.4019418954849243, "learning_rate": 4.951556604879049e-07, "loss": 2.531, "step": 875 }, { "epoch": 0.8725099601593626, "grad_norm": 1.1170932054519653, "learning_rate": 4.876353973635955e-07, "loss": 2.2724, "step": 876 }, { "epoch": 0.8735059760956175, "grad_norm": 0.8795150518417358, "learning_rate": 4.8016975125801e-07, "loss": 2.4447, "step": 877 }, { "epoch": 0.8745019920318725, "grad_norm": 1.4134328365325928, "learning_rate": 4.727588125342669e-07, "loss": 2.4728, "step": 878 }, { "epoch": 0.8754980079681275, "grad_norm": 0.8537651300430298, "learning_rate": 4.6540267089331294e-07, "loss": 1.8693, "step": 879 }, { "epoch": 0.8764940239043825, "grad_norm": 0.973147988319397, "learning_rate": 4.581014153728386e-07, "loss": 2.6805, "step": 880 }, { "epoch": 0.8774900398406374, "grad_norm": 2.0054306983947754, "learning_rate": 4.508551343462014e-07, "loss": 2.8345, "step": 881 }, { "epoch": 0.8784860557768924, "grad_norm": 1.3600786924362183, "learning_rate": 4.4366391552135567e-07, "loss": 2.4685, "step": 882 }, { "epoch": 0.8794820717131474, "grad_norm": 1.042197823524475, "learning_rate": 4.3652784593978927e-07, "loss": 2.2609, "step": 883 }, { "epoch": 0.8804780876494024, "grad_norm": 1.342214822769165, "learning_rate": 4.29447011975474e-07, "loss": 2.3443, "step": 884 }, { "epoch": 0.8814741035856574, "grad_norm": 1.0559214353561401, "learning_rate": 4.224214993338149e-07, "loss": 2.5197, "step": 885 }, { "epoch": 0.8824701195219123, "grad_norm": 0.9082587361335754, "learning_rate": 4.154513930506171e-07, "loss": 2.4609, "step": 886 }, { "epoch": 0.8834661354581673, "grad_norm": 1.555578589439392, "learning_rate": 4.0853677749105426e-07, "loss": 2.5883, "step": 887 }, { "epoch": 0.8844621513944223, "grad_norm": 0.6700481176376343, "learning_rate": 4.0167773634865017e-07, "loss": 2.4754, "step": 888 }, { "epoch": 0.8854581673306773, "grad_norm": 1.7728710174560547, "learning_rate": 3.9487435264426056e-07, "loss": 2.5669, "step": 889 }, { "epoch": 0.8864541832669323, "grad_norm": 0.8938197493553162, "learning_rate": 3.8812670872507454e-07, "loss": 2.7976, "step": 890 }, { "epoch": 0.8874501992031872, "grad_norm": 1.6636402606964111, "learning_rate": 3.8143488626361135e-07, "loss": 2.4172, "step": 891 }, { "epoch": 0.8884462151394422, "grad_norm": 1.0676062107086182, "learning_rate": 3.747989662567403e-07, "loss": 2.582, "step": 892 }, { "epoch": 0.8894422310756972, "grad_norm": 1.280415654182434, "learning_rate": 3.6821902902469066e-07, "loss": 2.2295, "step": 893 }, { "epoch": 0.8904382470119522, "grad_norm": 1.0814982652664185, "learning_rate": 3.6169515421008494e-07, "loss": 2.7001, "step": 894 }, { "epoch": 0.8914342629482072, "grad_norm": 1.077160120010376, "learning_rate": 3.5522742077697734e-07, "loss": 2.49, "step": 895 }, { "epoch": 0.8924302788844621, "grad_norm": 1.2483303546905518, "learning_rate": 3.4881590700989175e-07, "loss": 2.3397, "step": 896 }, { "epoch": 0.8934262948207171, "grad_norm": 0.6463543772697449, "learning_rate": 3.4246069051287747e-07, "loss": 2.3176, "step": 897 }, { "epoch": 0.8944223107569721, "grad_norm": 0.8918944001197815, "learning_rate": 3.3616184820856936e-07, "loss": 2.5445, "step": 898 }, { "epoch": 0.8954183266932271, "grad_norm": 0.963518500328064, "learning_rate": 3.299194563372604e-07, "loss": 2.3521, "step": 899 }, { "epoch": 0.896414342629482, "grad_norm": 1.8490091562271118, "learning_rate": 3.237335904559713e-07, "loss": 1.8583, "step": 900 }, { "epoch": 0.897410358565737, "grad_norm": 1.2349917888641357, "learning_rate": 3.176043254375422e-07, "loss": 2.5093, "step": 901 }, { "epoch": 0.898406374501992, "grad_norm": 1.3500816822052002, "learning_rate": 3.1153173546972395e-07, "loss": 2.3507, "step": 902 }, { "epoch": 0.899402390438247, "grad_norm": 1.2740628719329834, "learning_rate": 3.055158940542818e-07, "loss": 2.4008, "step": 903 }, { "epoch": 0.900398406374502, "grad_norm": 33.82315444946289, "learning_rate": 2.9955687400610336e-07, "loss": 2.3794, "step": 904 }, { "epoch": 0.901394422310757, "grad_norm": 0.8968676328659058, "learning_rate": 2.9365474745231935e-07, "loss": 2.3304, "step": 905 }, { "epoch": 0.9023904382470119, "grad_norm": 1.474859356880188, "learning_rate": 2.878095858314278e-07, "loss": 2.2761, "step": 906 }, { "epoch": 0.9033864541832669, "grad_norm": 1.0338733196258545, "learning_rate": 2.820214598924348e-07, "loss": 2.222, "step": 907 }, { "epoch": 0.9043824701195219, "grad_norm": 1.2461026906967163, "learning_rate": 2.7629043969399193e-07, "loss": 2.129, "step": 908 }, { "epoch": 0.9053784860557769, "grad_norm": 1.0711873769760132, "learning_rate": 2.7061659460355047e-07, "loss": 2.5671, "step": 909 }, { "epoch": 0.9063745019920318, "grad_norm": 2.0772130489349365, "learning_rate": 2.6499999329652525e-07, "loss": 2.4514, "step": 910 }, { "epoch": 0.9073705179282868, "grad_norm": 0.8682186603546143, "learning_rate": 2.594407037554586e-07, "loss": 2.4655, "step": 911 }, { "epoch": 0.9083665338645418, "grad_norm": 2.363909959793091, "learning_rate": 2.539387932691995e-07, "loss": 3.0087, "step": 912 }, { "epoch": 0.9093625498007968, "grad_norm": 1.1324294805526733, "learning_rate": 2.4849432843208786e-07, "loss": 2.6718, "step": 913 }, { "epoch": 0.9103585657370518, "grad_norm": 1.2291409969329834, "learning_rate": 2.431073751431529e-07, "loss": 2.4459, "step": 914 }, { "epoch": 0.9113545816733067, "grad_norm": 1.1384942531585693, "learning_rate": 2.377779986053097e-07, "loss": 2.3891, "step": 915 }, { "epoch": 0.9123505976095617, "grad_norm": 0.9154942631721497, "learning_rate": 2.3250626332457226e-07, "loss": 2.3965, "step": 916 }, { "epoch": 0.9133466135458167, "grad_norm": 3.3389575481414795, "learning_rate": 2.2729223310927473e-07, "loss": 2.4395, "step": 917 }, { "epoch": 0.9143426294820717, "grad_norm": 1.6811953783035278, "learning_rate": 2.2213597106929608e-07, "loss": 2.6017, "step": 918 }, { "epoch": 0.9153386454183267, "grad_norm": 0.8396251201629639, "learning_rate": 2.1703753961529906e-07, "loss": 2.7736, "step": 919 }, { "epoch": 0.9163346613545816, "grad_norm": 1.4626351594924927, "learning_rate": 2.1199700045797077e-07, "loss": 2.2861, "step": 920 }, { "epoch": 0.9173306772908366, "grad_norm": 1.5617847442626953, "learning_rate": 2.070144146072789e-07, "loss": 2.6273, "step": 921 }, { "epoch": 0.9183266932270916, "grad_norm": 1.4151337146759033, "learning_rate": 2.0208984237173546e-07, "loss": 2.1617, "step": 922 }, { "epoch": 0.9193227091633466, "grad_norm": 0.9167352318763733, "learning_rate": 1.9722334335766092e-07, "loss": 2.3805, "step": 923 }, { "epoch": 0.9203187250996016, "grad_norm": 0.840559720993042, "learning_rate": 1.9241497646846463e-07, "loss": 2.3247, "step": 924 }, { "epoch": 0.9213147410358565, "grad_norm": 0.9646689891815186, "learning_rate": 1.876647999039377e-07, "loss": 2.5301, "step": 925 }, { "epoch": 0.9223107569721115, "grad_norm": 0.7615554928779602, "learning_rate": 1.829728711595391e-07, "loss": 2.3153, "step": 926 }, { "epoch": 0.9233067729083665, "grad_norm": 1.138389229774475, "learning_rate": 1.7833924702570725e-07, "loss": 2.573, "step": 927 }, { "epoch": 0.9243027888446215, "grad_norm": 1.0490334033966064, "learning_rate": 1.7376398358716852e-07, "loss": 2.6447, "step": 928 }, { "epoch": 0.9252988047808764, "grad_norm": 0.9057884812355042, "learning_rate": 1.6924713622225975e-07, "loss": 2.6325, "step": 929 }, { "epoch": 0.9262948207171314, "grad_norm": 2.5391931533813477, "learning_rate": 1.6478875960225904e-07, "loss": 2.7326, "step": 930 }, { "epoch": 0.9272908366533864, "grad_norm": 1.9867279529571533, "learning_rate": 1.6038890769072223e-07, "loss": 2.4295, "step": 931 }, { "epoch": 0.9282868525896414, "grad_norm": 0.7426679730415344, "learning_rate": 1.5604763374283073e-07, "loss": 2.3705, "step": 932 }, { "epoch": 0.9292828685258964, "grad_norm": 1.033311367034912, "learning_rate": 1.5176499030474578e-07, "loss": 2.2442, "step": 933 }, { "epoch": 0.9302788844621513, "grad_norm": 1.394702672958374, "learning_rate": 1.4754102921297363e-07, "loss": 2.2417, "step": 934 }, { "epoch": 0.9312749003984063, "grad_norm": 1.109711766242981, "learning_rate": 1.4337580159373864e-07, "loss": 2.1194, "step": 935 }, { "epoch": 0.9322709163346613, "grad_norm": 0.810978353023529, "learning_rate": 1.3926935786236218e-07, "loss": 2.251, "step": 936 }, { "epoch": 0.9332669322709163, "grad_norm": 0.9075368642807007, "learning_rate": 1.3522174772265585e-07, "loss": 2.3295, "step": 937 }, { "epoch": 0.9342629482071713, "grad_norm": 1.0174400806427002, "learning_rate": 1.3123302016631477e-07, "loss": 2.4231, "step": 938 }, { "epoch": 0.9352589641434262, "grad_norm": 1.4848259687423706, "learning_rate": 1.2730322347233037e-07, "loss": 2.327, "step": 939 }, { "epoch": 0.9362549800796812, "grad_norm": 0.7531813979148865, "learning_rate": 1.2343240520640287e-07, "loss": 2.5221, "step": 940 }, { "epoch": 0.9372509960159362, "grad_norm": 2.318554162979126, "learning_rate": 1.196206122203647e-07, "loss": 2.3946, "step": 941 }, { "epoch": 0.9382470119521913, "grad_norm": 0.8973721861839294, "learning_rate": 1.158678906516153e-07, "loss": 2.1919, "step": 942 }, { "epoch": 0.9392430278884463, "grad_norm": 1.4439376592636108, "learning_rate": 1.1217428592256218e-07, "loss": 2.3653, "step": 943 }, { "epoch": 0.9402390438247012, "grad_norm": 1.6569935083389282, "learning_rate": 1.0853984274007246e-07, "loss": 2.6982, "step": 944 }, { "epoch": 0.9412350597609562, "grad_norm": 1.5078299045562744, "learning_rate": 1.0496460509492767e-07, "loss": 2.4708, "step": 945 }, { "epoch": 0.9422310756972112, "grad_norm": 2.2158310413360596, "learning_rate": 1.0144861626129599e-07, "loss": 2.5999, "step": 946 }, { "epoch": 0.9432270916334662, "grad_norm": 1.4059021472930908, "learning_rate": 9.799191879620474e-08, "loss": 2.4373, "step": 947 }, { "epoch": 0.9442231075697212, "grad_norm": 2.9164271354675293, "learning_rate": 9.459455453902866e-08, "loss": 2.6697, "step": 948 }, { "epoch": 0.9452191235059761, "grad_norm": 1.275817632675171, "learning_rate": 9.125656461098142e-08, "loss": 2.48, "step": 949 }, { "epoch": 0.9462151394422311, "grad_norm": 0.821499764919281, "learning_rate": 8.797798941461655e-08, "loss": 2.4301, "step": 950 }, { "epoch": 0.9472111553784861, "grad_norm": 1.4214954376220703, "learning_rate": 8.475886863334282e-08, "loss": 2.2847, "step": 951 }, { "epoch": 0.9482071713147411, "grad_norm": 0.9953071475028992, "learning_rate": 8.15992412309391e-08, "loss": 2.464, "step": 952 }, { "epoch": 0.9492031872509961, "grad_norm": 1.4973928928375244, "learning_rate": 7.84991454510864e-08, "loss": 2.0293, "step": 953 }, { "epoch": 0.950199203187251, "grad_norm": 1.326232671737671, "learning_rate": 7.545861881690097e-08, "loss": 2.4445, "step": 954 }, { "epoch": 0.951195219123506, "grad_norm": 1.0779331922531128, "learning_rate": 7.247769813048644e-08, "loss": 2.7232, "step": 955 }, { "epoch": 0.952191235059761, "grad_norm": 1.0593082904815674, "learning_rate": 6.955641947248127e-08, "loss": 2.8634, "step": 956 }, { "epoch": 0.953187250996016, "grad_norm": 0.8761929869651794, "learning_rate": 6.669481820162638e-08, "loss": 2.3328, "step": 957 }, { "epoch": 0.954183266932271, "grad_norm": 0.9143054485321045, "learning_rate": 6.389292895433608e-08, "loss": 2.2261, "step": 958 }, { "epoch": 0.9551792828685259, "grad_norm": 0.8204777240753174, "learning_rate": 6.115078564427946e-08, "loss": 2.4155, "step": 959 }, { "epoch": 0.9561752988047809, "grad_norm": 0.7546234726905823, "learning_rate": 5.8468421461968517e-08, "loss": 2.267, "step": 960 }, { "epoch": 0.9571713147410359, "grad_norm": 1.0376356840133667, "learning_rate": 5.584586887435739e-08, "loss": 2.4496, "step": 961 }, { "epoch": 0.9581673306772909, "grad_norm": 0.6573870182037354, "learning_rate": 5.3283159624448745e-08, "loss": 2.4904, "step": 962 }, { "epoch": 0.9591633466135459, "grad_norm": 1.3613762855529785, "learning_rate": 5.0780324730911877e-08, "loss": 2.4824, "step": 963 }, { "epoch": 0.9601593625498008, "grad_norm": 1.4304169416427612, "learning_rate": 4.833739448770247e-08, "loss": 2.9062, "step": 964 }, { "epoch": 0.9611553784860558, "grad_norm": 0.9547715783119202, "learning_rate": 4.5954398463700647e-08, "loss": 2.4977, "step": 965 }, { "epoch": 0.9621513944223108, "grad_norm": 1.3909553289413452, "learning_rate": 4.3631365502351805e-08, "loss": 2.2116, "step": 966 }, { "epoch": 0.9631474103585658, "grad_norm": 0.7387050986289978, "learning_rate": 4.136832372131583e-08, "loss": 2.5225, "step": 967 }, { "epoch": 0.9641434262948207, "grad_norm": 1.2469770908355713, "learning_rate": 3.916530051212841e-08, "loss": 2.4759, "step": 968 }, { "epoch": 0.9651394422310757, "grad_norm": 1.3780826330184937, "learning_rate": 3.702232253986804e-08, "loss": 2.0538, "step": 969 }, { "epoch": 0.9661354581673307, "grad_norm": 0.9699292778968811, "learning_rate": 3.4939415742835655e-08, "loss": 2.3441, "step": 970 }, { "epoch": 0.9671314741035857, "grad_norm": 1.147615671157837, "learning_rate": 3.2916605332238284e-08, "loss": 2.4042, "step": 971 }, { "epoch": 0.9681274900398407, "grad_norm": 1.1738359928131104, "learning_rate": 3.095391579188589e-08, "loss": 2.539, "step": 972 }, { "epoch": 0.9691235059760956, "grad_norm": 0.640042781829834, "learning_rate": 2.9051370877892226e-08, "loss": 2.3044, "step": 973 }, { "epoch": 0.9701195219123506, "grad_norm": 0.7774790525436401, "learning_rate": 2.7208993618390578e-08, "loss": 2.2616, "step": 974 }, { "epoch": 0.9711155378486056, "grad_norm": 1.0389803647994995, "learning_rate": 2.5426806313252895e-08, "loss": 2.4425, "step": 975 }, { "epoch": 0.9721115537848606, "grad_norm": 1.3041914701461792, "learning_rate": 2.370483053382111e-08, "loss": 2.39, "step": 976 }, { "epoch": 0.9731075697211156, "grad_norm": 0.6934490203857422, "learning_rate": 2.2043087122644023e-08, "loss": 2.2232, "step": 977 }, { "epoch": 0.9741035856573705, "grad_norm": 0.6868986487388611, "learning_rate": 2.0441596193227497e-08, "loss": 2.4806, "step": 978 }, { "epoch": 0.9750996015936255, "grad_norm": 1.1924256086349487, "learning_rate": 1.8900377129790205e-08, "loss": 2.4314, "step": 979 }, { "epoch": 0.9760956175298805, "grad_norm": 0.7938891053199768, "learning_rate": 1.741944858702771e-08, "loss": 2.4715, "step": 980 }, { "epoch": 0.9770916334661355, "grad_norm": 0.9900745749473572, "learning_rate": 1.5998828489888762e-08, "loss": 2.0915, "step": 981 }, { "epoch": 0.9780876494023905, "grad_norm": 3.0510518550872803, "learning_rate": 1.4638534033356578e-08, "loss": 3.3239, "step": 982 }, { "epoch": 0.9790836653386454, "grad_norm": 1.1174182891845703, "learning_rate": 1.333858168224178e-08, "loss": 2.1631, "step": 983 }, { "epoch": 0.9800796812749004, "grad_norm": 0.7828091979026794, "learning_rate": 1.2098987170982013e-08, "loss": 2.2998, "step": 984 }, { "epoch": 0.9810756972111554, "grad_norm": 0.8114204406738281, "learning_rate": 1.0919765503453195e-08, "loss": 2.3996, "step": 985 }, { "epoch": 0.9820717131474104, "grad_norm": 0.625230073928833, "learning_rate": 9.800930952786336e-09, "loss": 2.6785, "step": 986 }, { "epoch": 0.9830677290836654, "grad_norm": 1.6107351779937744, "learning_rate": 8.742497061195455e-09, "loss": 2.6999, "step": 987 }, { "epoch": 0.9840637450199203, "grad_norm": 1.5219416618347168, "learning_rate": 7.744476639813814e-09, "loss": 2.3396, "step": 988 }, { "epoch": 0.9850597609561753, "grad_norm": 1.0381386280059814, "learning_rate": 6.806881768539053e-09, "loss": 2.2097, "step": 989 }, { "epoch": 0.9860557768924303, "grad_norm": 1.1309791803359985, "learning_rate": 5.929723795884967e-09, "loss": 2.4901, "step": 990 }, { "epoch": 0.9870517928286853, "grad_norm": 0.979224443435669, "learning_rate": 5.113013338847173e-09, "loss": 2.355, "step": 991 }, { "epoch": 0.9880478087649402, "grad_norm": 0.9343250393867493, "learning_rate": 4.356760282773209e-09, "loss": 2.682, "step": 992 }, { "epoch": 0.9890438247011952, "grad_norm": 4.211667060852051, "learning_rate": 3.660973781242083e-09, "loss": 3.9138, "step": 993 }, { "epoch": 0.9900398406374502, "grad_norm": 0.9402066469192505, "learning_rate": 3.0256622559543537e-09, "loss": 2.1841, "step": 994 }, { "epoch": 0.9910358565737052, "grad_norm": 1.136916995048523, "learning_rate": 2.4508333966305473e-09, "loss": 2.2469, "step": 995 }, { "epoch": 0.9920318725099602, "grad_norm": 1.080809473991394, "learning_rate": 1.936494160916791e-09, "loss": 2.4922, "step": 996 }, { "epoch": 0.9930278884462151, "grad_norm": 0.9956486821174622, "learning_rate": 1.4826507743032071e-09, "loss": 2.5901, "step": 997 }, { "epoch": 0.9940239043824701, "grad_norm": 2.1008529663085938, "learning_rate": 1.089308730043981e-09, "loss": 2.8828, "step": 998 }, { "epoch": 0.9950199203187251, "grad_norm": 1.3167147636413574, "learning_rate": 7.564727890968515e-10, "loss": 2.3331, "step": 999 }, { "epoch": 0.9960159362549801, "grad_norm": 1.1669059991836548, "learning_rate": 4.841469800592746e-10, "loss": 1.9942, "step": 1000 }, { "epoch": 0.9970119521912351, "grad_norm": 0.9161001443862915, "learning_rate": 2.723345991245685e-10, "loss": 2.2932, "step": 1001 }, { "epoch": 0.99800796812749, "grad_norm": 0.8516436815261841, "learning_rate": 1.210382100397256e-10, "loss": 2.2065, "step": 1002 }, { "epoch": 0.999003984063745, "grad_norm": 0.7772925496101379, "learning_rate": 3.0259644074326355e-11, "loss": 2.4638, "step": 1003 }, { "epoch": 1.0, "grad_norm": 1.244535207748413, "learning_rate": 0.0, "loss": 2.3463, "step": 1004 } ], "logging_steps": 1, "max_steps": 1004, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.5282748355775386e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }