| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.994671403197158, |
| "global_step": 843, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.01, |
| "learning_rate": 1.4117647058823532e-05, |
| "loss": 5.6404, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 2.8235294117647063e-05, |
| "loss": 4.5746, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 4.235294117647059e-05, |
| "loss": 3.3627, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 5.6470588235294126e-05, |
| "loss": 2.7997, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 7.058823529411765e-05, |
| "loss": 2.4203, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 7.999971068621779e-05, |
| "loss": 2.217, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 7.999537106318664e-05, |
| "loss": 2.2324, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 7.998582444493811e-05, |
| "loss": 2.089, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 7.997107207434592e-05, |
| "loss": 2.0586, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 7.995111587202043e-05, |
| "loss": 1.9011, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 7.992595843605852e-05, |
| "loss": 1.8797, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 7.989560304170546e-05, |
| "loss": 1.8855, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 7.98600536409284e-05, |
| "loss": 1.7122, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 7.981931486190192e-05, |
| "loss": 1.7285, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 7.97733920084055e-05, |
| "loss": 1.6935, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 7.972229105913295e-05, |
| "loss": 1.5963, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 7.966601866691413e-05, |
| "loss": 1.6079, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 7.96045821578488e-05, |
| "loss": 1.6835, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 7.953798953035274e-05, |
| "loss": 1.5928, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 7.946624945411658e-05, |
| "loss": 1.6176, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 7.938937126897702e-05, |
| "loss": 1.5979, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 7.930736498370085e-05, |
| "loss": 1.5759, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 7.922024127468196e-05, |
| "loss": 1.633, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 7.912801148455134e-05, |
| "loss": 1.5562, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 7.903068762070047e-05, |
| "loss": 1.5349, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 7.892828235371795e-05, |
| "loss": 1.5454, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 7.882080901574001e-05, |
| "loss": 1.5811, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 7.870828159871478e-05, |
| "loss": 1.5442, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 7.859071475258065e-05, |
| "loss": 1.5513, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 7.8468123783359e-05, |
| "loss": 1.5349, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 7.83405246511616e-05, |
| "loss": 1.5035, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 7.820793396811258e-05, |
| "loss": 1.5159, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 7.80703689961859e-05, |
| "loss": 1.4197, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 7.792784764495791e-05, |
| "loss": 1.4168, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 7.778038846927566e-05, |
| "loss": 1.4997, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 7.762801066684133e-05, |
| "loss": 1.4985, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 7.747073407571291e-05, |
| "loss": 1.2951, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 7.730857917172132e-05, |
| "loss": 1.3738, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 7.714156706580487e-05, |
| "loss": 1.385, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 7.696971950126068e-05, |
| "loss": 1.4614, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 7.679305885091398e-05, |
| "loss": 1.3669, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 7.661160811420536e-05, |
| "loss": 1.3745, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 7.642539091419654e-05, |
| "loss": 1.3872, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 7.623443149449479e-05, |
| "loss": 1.3118, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 7.603875471609677e-05, |
| "loss": 1.3973, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 7.583838605415177e-05, |
| "loss": 1.4051, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 7.563335159464515e-05, |
| "loss": 1.4145, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 7.542367803100223e-05, |
| "loss": 1.342, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 7.520939266061304e-05, |
| "loss": 1.3072, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 7.499052338127845e-05, |
| "loss": 1.3066, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 7.476709868757825e-05, |
| "loss": 1.3681, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 7.453914766716133e-05, |
| "loss": 1.3416, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 7.430669999695885e-05, |
| "loss": 1.4098, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 7.406978593932055e-05, |
| "loss": 1.4484, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 7.382843633807488e-05, |
| "loss": 1.2463, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 7.358268261451353e-05, |
| "loss": 1.3308, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 7.33325567633006e-05, |
| "loss": 1.2935, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 7.307809134830729e-05, |
| "loss": 1.2542, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 7.281931949837233e-05, |
| "loss": 1.2856, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 7.255627490298902e-05, |
| "loss": 1.3299, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 7.228899180791914e-05, |
| "loss": 1.2866, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 7.201750501073452e-05, |
| "loss": 1.2212, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 7.174184985628679e-05, |
| "loss": 1.2903, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 7.146206223210571e-05, |
| "loss": 1.2567, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 7.117817856372709e-05, |
| "loss": 1.2662, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 7.089023580995057e-05, |
| "loss": 1.2442, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 7.059827145802777e-05, |
| "loss": 1.2454, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 7.03023235187821e-05, |
| "loss": 1.3152, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 7.000243052165989e-05, |
| "loss": 1.2658, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 6.969863150971439e-05, |
| "loss": 1.1871, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 6.939096603452269e-05, |
| "loss": 1.2346, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 6.907947415103656e-05, |
| "loss": 1.1971, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 6.87641964123676e-05, |
| "loss": 1.2636, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 6.844517386450775e-05, |
| "loss": 1.2181, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 6.812244804098544e-05, |
| "loss": 1.2436, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 6.779606095745832e-05, |
| "loss": 1.249, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 6.746605510624337e-05, |
| "loss": 1.2698, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 6.713247345078465e-05, |
| "loss": 1.1873, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 6.679535942006004e-05, |
| "loss": 1.1497, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 6.645475690292716e-05, |
| "loss": 1.2872, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 6.611071024240946e-05, |
| "loss": 1.2413, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 6.576326422992326e-05, |
| "loss": 1.213, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 6.541246409944634e-05, |
| "loss": 1.2931, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 6.505835552162887e-05, |
| "loss": 1.2878, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 6.470098459784768e-05, |
| "loss": 1.2502, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 6.434039785420417e-05, |
| "loss": 1.2416, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 6.397664223546721e-05, |
| "loss": 1.2468, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 6.360976509896133e-05, |
| "loss": 1.2207, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 6.323981420840124e-05, |
| "loss": 1.2038, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 6.286683772767365e-05, |
| "loss": 1.3596, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 6.249088421456656e-05, |
| "loss": 1.1464, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 6.211200261444774e-05, |
| "loss": 1.2218, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 6.173024225389242e-05, |
| "loss": 1.1165, |
| "step": 279 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_gen_len": 28.29241516966068, |
| "eval_loss": 1.1089943647384644, |
| "eval_rouge1": 57.9239, |
| "eval_rouge2": 31.9259, |
| "eval_rougeL": 53.8737, |
| "eval_rougeLsum": 54.9811, |
| "eval_runtime": 90.9427, |
| "eval_samples_per_second": 11.018, |
| "eval_steps_per_second": 2.76, |
| "step": 281 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 6.134565283426149e-05, |
| "loss": 1.1982, |
| "step": 282 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 6.0958284425230914e-05, |
| "loss": 1.117, |
| "step": 285 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 6.0568187458273134e-05, |
| "loss": 1.1124, |
| "step": 288 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 6.017541272009135e-05, |
| "loss": 1.0754, |
| "step": 291 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 5.9780011346007745e-05, |
| "loss": 1.046, |
| "step": 294 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 5.9382034813306014e-05, |
| "loss": 1.1768, |
| "step": 297 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 5.898153493452965e-05, |
| "loss": 1.1407, |
| "step": 300 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 5.857856385073643e-05, |
| "loss": 1.1402, |
| "step": 303 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 5.817317402471016e-05, |
| "loss": 1.0946, |
| "step": 306 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 5.776541823413062e-05, |
| "loss": 1.0745, |
| "step": 309 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 5.735534956470233e-05, |
| "loss": 1.1032, |
| "step": 312 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 5.6943021403243415e-05, |
| "loss": 1.0543, |
| "step": 315 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 5.652848743073513e-05, |
| "loss": 1.0755, |
| "step": 318 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 5.611180161533318e-05, |
| "loss": 1.0945, |
| "step": 321 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 5.569301820534154e-05, |
| "loss": 1.158, |
| "step": 324 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 5.5272191722149965e-05, |
| "loss": 1.1357, |
| "step": 327 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 5.484937695313573e-05, |
| "loss": 1.1591, |
| "step": 330 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 5.4424628944531024e-05, |
| "loss": 1.0965, |
| "step": 333 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 5.399800299425631e-05, |
| "loss": 1.0746, |
| "step": 336 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 5.356955464472121e-05, |
| "loss": 1.1068, |
| "step": 339 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 5.3139339675593446e-05, |
| "loss": 1.0997, |
| "step": 342 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 5.2707414096536793e-05, |
| "loss": 1.1239, |
| "step": 345 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 5.227383413991925e-05, |
| "loss": 1.1499, |
| "step": 348 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 5.1838656253492175e-05, |
| "loss": 1.0583, |
| "step": 351 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 5.1401937093041305e-05, |
| "loss": 0.9659, |
| "step": 354 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 5.096373351501074e-05, |
| "loss": 1.0781, |
| "step": 357 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 5.052410256910085e-05, |
| "loss": 1.0969, |
| "step": 360 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 5.008310149084097e-05, |
| "loss": 1.1756, |
| "step": 363 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 4.964078769413782e-05, |
| "loss": 1.0718, |
| "step": 366 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 4.919721876380094e-05, |
| "loss": 1.1295, |
| "step": 369 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 4.8752452448045626e-05, |
| "loss": 1.1278, |
| "step": 372 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 4.830654665097469e-05, |
| "loss": 1.1014, |
| "step": 375 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 4.785955942504001e-05, |
| "loss": 1.0543, |
| "step": 378 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 4.741154896348458e-05, |
| "loss": 1.126, |
| "step": 381 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 4.696257359276641e-05, |
| "loss": 1.0462, |
| "step": 384 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 4.651269176496496e-05, |
| "loss": 1.1578, |
| "step": 387 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 4.606196205017133e-05, |
| "loss": 1.1495, |
| "step": 390 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 4.561044312886294e-05, |
| "loss": 1.0439, |
| "step": 393 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 4.5158193784263975e-05, |
| "loss": 1.0773, |
| "step": 396 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 4.4705272894692425e-05, |
| "loss": 1.1168, |
| "step": 399 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 4.425173942589462e-05, |
| "loss": 1.0929, |
| "step": 402 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 4.379765242336859e-05, |
| "loss": 1.1132, |
| "step": 405 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 4.3343071004676905e-05, |
| "loss": 0.9778, |
| "step": 408 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 4.288805435175011e-05, |
| "loss": 1.0954, |
| "step": 411 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 4.2432661703181875e-05, |
| "loss": 1.0817, |
| "step": 414 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 4.197695234651674e-05, |
| "loss": 1.1092, |
| "step": 417 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 4.152098561053148e-05, |
| "loss": 1.1176, |
| "step": 420 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 4.1064820857511e-05, |
| "loss": 1.0565, |
| "step": 423 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 4.0608517475520084e-05, |
| "loss": 1.1288, |
| "step": 426 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 4.015213487067163e-05, |
| "loss": 1.1744, |
| "step": 429 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 3.96957324593925e-05, |
| "loss": 1.1086, |
| "step": 432 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 3.92393696606882e-05, |
| "loss": 1.0737, |
| "step": 435 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 3.878310588840703e-05, |
| "loss": 1.1223, |
| "step": 438 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 3.832700054350511e-05, |
| "loss": 1.0578, |
| "step": 441 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 3.787111300631287e-05, |
| "loss": 1.0141, |
| "step": 444 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 3.7415502628804334e-05, |
| "loss": 1.0913, |
| "step": 447 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 3.6960228726870186e-05, |
| "loss": 1.1208, |
| "step": 450 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 3.650535057259536e-05, |
| "loss": 1.073, |
| "step": 453 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 3.6050927386542365e-05, |
| "loss": 1.1406, |
| "step": 456 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 3.5597018330041476e-05, |
| "loss": 1.0705, |
| "step": 459 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 3.514368249748849e-05, |
| "loss": 1.0571, |
| "step": 462 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 3.469097890865113e-05, |
| "loss": 1.0793, |
| "step": 465 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 3.4238966500985405e-05, |
| "loss": 1.1336, |
| "step": 468 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 3.378770412196249e-05, |
| "loss": 1.0964, |
| "step": 471 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 3.333725052140732e-05, |
| "loss": 1.0749, |
| "step": 474 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 3.2887664343850045e-05, |
| "loss": 1.0811, |
| "step": 477 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 3.243900412089106e-05, |
| "loss": 1.082, |
| "step": 480 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 3.199132826358075e-05, |
| "loss": 1.0476, |
| "step": 483 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 3.154469505481503e-05, |
| "loss": 1.1107, |
| "step": 486 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 3.109916264174743e-05, |
| "loss": 1.1545, |
| "step": 489 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 3.065478902821894e-05, |
| "loss": 1.0738, |
| "step": 492 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 3.0211632067206533e-05, |
| "loss": 1.0883, |
| "step": 495 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 2.976974945329125e-05, |
| "loss": 1.0172, |
| "step": 498 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 2.932919871514689e-05, |
| "loss": 1.076, |
| "step": 501 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 2.8890037208050486e-05, |
| "loss": 1.1079, |
| "step": 504 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 2.845232210641517e-05, |
| "loss": 1.0263, |
| "step": 507 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 2.801611039634654e-05, |
| "loss": 1.0274, |
| "step": 510 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 2.7581458868223796e-05, |
| "loss": 1.0736, |
| "step": 513 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 2.714842410930612e-05, |
| "loss": 1.0472, |
| "step": 516 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 2.6717062496365574e-05, |
| "loss": 1.0982, |
| "step": 519 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 2.6287430188347456e-05, |
| "loss": 1.0779, |
| "step": 522 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 2.5859583119058967e-05, |
| "loss": 1.0563, |
| "step": 525 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 2.5433576989887115e-05, |
| "loss": 1.0575, |
| "step": 528 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 2.5009467262547072e-05, |
| "loss": 1.0763, |
| "step": 531 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 2.458730915186153e-05, |
| "loss": 1.0557, |
| "step": 534 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 2.4167157618572312e-05, |
| "loss": 0.9993, |
| "step": 537 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 2.374906736218506e-05, |
| "loss": 1.0631, |
| "step": 540 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 2.333309281384784e-05, |
| "loss": 0.9966, |
| "step": 543 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 2.2919288129264837e-05, |
| "loss": 1.0742, |
| "step": 546 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 2.250770718164579e-05, |
| "loss": 1.0572, |
| "step": 549 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 2.209840355469223e-05, |
| "loss": 1.0628, |
| "step": 552 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 2.1691430535621432e-05, |
| "loss": 1.0401, |
| "step": 555 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 2.1286841108228987e-05, |
| "loss": 0.9666, |
| "step": 558 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 2.0884687945990744e-05, |
| "loss": 1.0763, |
| "step": 561 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_gen_len": 29.696606786427147, |
| "eval_loss": 1.0267277956008911, |
| "eval_rouge1": 59.9605, |
| "eval_rouge2": 34.0298, |
| "eval_rougeL": 55.7523, |
| "eval_rougeLsum": 56.8021, |
| "eval_runtime": 96.5427, |
| "eval_samples_per_second": 10.379, |
| "eval_steps_per_second": 2.6, |
| "step": 563 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 2.048502340520538e-05, |
| "loss": 1.0732, |
| "step": 564 |
| }, |
| { |
| "epoch": 2.01, |
| "learning_rate": 2.008789951817809e-05, |
| "loss": 0.9671, |
| "step": 567 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 1.9693367986446415e-05, |
| "loss": 1.0325, |
| "step": 570 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 1.930148017404936e-05, |
| "loss": 0.9656, |
| "step": 573 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 1.8912287100840275e-05, |
| "loss": 0.9455, |
| "step": 576 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 1.852583943584449e-05, |
| "loss": 0.993, |
| "step": 579 |
| }, |
| { |
| "epoch": 2.07, |
| "learning_rate": 1.8142187490662827e-05, |
| "loss": 0.9598, |
| "step": 582 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 1.7761381212921508e-05, |
| "loss": 0.978, |
| "step": 585 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 1.7383470179769416e-05, |
| "loss": 1.0325, |
| "step": 588 |
| }, |
| { |
| "epoch": 2.1, |
| "learning_rate": 1.700850359142373e-05, |
| "loss": 0.9173, |
| "step": 591 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 1.663653026476452e-05, |
| "loss": 1.0184, |
| "step": 594 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 1.6267598626979264e-05, |
| "loss": 0.9831, |
| "step": 597 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 1.5901756709258133e-05, |
| "loss": 0.9981, |
| "step": 600 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 1.553905214054082e-05, |
| "loss": 0.9645, |
| "step": 603 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 1.5179532141315685e-05, |
| "loss": 0.9968, |
| "step": 606 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 1.4823243517472187e-05, |
| "loss": 0.9537, |
| "step": 609 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 1.4470232654207208e-05, |
| "loss": 0.9923, |
| "step": 612 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 1.4120545509986103e-05, |
| "loss": 0.9497, |
| "step": 615 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 1.3774227610559461e-05, |
| "loss": 0.9502, |
| "step": 618 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 1.3431324043036104e-05, |
| "loss": 0.9703, |
| "step": 621 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 1.3091879450013086e-05, |
| "loss": 0.9592, |
| "step": 624 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 1.2755938023763821e-05, |
| "loss": 0.9748, |
| "step": 627 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 1.2423543500484607e-05, |
| "loss": 0.9978, |
| "step": 630 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 1.2094739154600616e-05, |
| "loss": 1.0247, |
| "step": 633 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 1.1769567793132048e-05, |
| "loss": 0.9923, |
| "step": 636 |
| }, |
| { |
| "epoch": 2.27, |
| "learning_rate": 1.1448071750121042e-05, |
| "loss": 0.9638, |
| "step": 639 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 1.113029288112022e-05, |
| "loss": 0.9403, |
| "step": 642 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 1.0816272557743525e-05, |
| "loss": 1.0162, |
| "step": 645 |
| }, |
| { |
| "epoch": 2.3, |
| "learning_rate": 1.050605166228006e-05, |
| "loss": 0.9988, |
| "step": 648 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 1.0199670582371573e-05, |
| "loss": 1.004, |
| "step": 651 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 9.897169205754461e-06, |
| "loss": 0.9909, |
| "step": 654 |
| }, |
| { |
| "epoch": 2.33, |
| "learning_rate": 9.598586915066766e-06, |
| "loss": 0.977, |
| "step": 657 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 9.303962582720887e-06, |
| "loss": 1.001, |
| "step": 660 |
| }, |
| { |
| "epoch": 2.36, |
| "learning_rate": 9.013334565842929e-06, |
| "loss": 0.9925, |
| "step": 663 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 8.72674070127881e-06, |
| "loss": 0.9803, |
| "step": 666 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 8.44421830066843e-06, |
| "loss": 0.9129, |
| "step": 669 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 8.165804145588008e-06, |
| "loss": 1.0281, |
| "step": 672 |
| }, |
| { |
| "epoch": 2.4, |
| "learning_rate": 7.891534482761463e-06, |
| "loss": 0.9433, |
| "step": 675 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 7.621445019341522e-06, |
| "loss": 0.9536, |
| "step": 678 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 7.355570918260979e-06, |
| "loss": 1.0091, |
| "step": 681 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 7.0939467936548045e-06, |
| "loss": 1.0031, |
| "step": 684 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 6.836606706353786e-06, |
| "loss": 0.9501, |
| "step": 687 |
| }, |
| { |
| "epoch": 2.45, |
| "learning_rate": 6.583584159450116e-06, |
| "loss": 1.0059, |
| "step": 690 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 6.334912093935605e-06, |
| "loss": 1.0037, |
| "step": 693 |
| }, |
| { |
| "epoch": 2.47, |
| "learning_rate": 6.090622884413164e-06, |
| "loss": 0.9347, |
| "step": 696 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 5.85074833488191e-06, |
| "loss": 0.9206, |
| "step": 699 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 5.6153196745965774e-06, |
| "loss": 0.9739, |
| "step": 702 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 5.384367554001864e-06, |
| "loss": 0.9265, |
| "step": 705 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 5.157922040741996e-06, |
| "loss": 0.9867, |
| "step": 708 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 4.936012615746193e-06, |
| "loss": 0.9579, |
| "step": 711 |
| }, |
| { |
| "epoch": 2.54, |
| "learning_rate": 4.71866816939063e-06, |
| "loss": 0.9567, |
| "step": 714 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 4.505916997737143e-06, |
| "loss": 0.9833, |
| "step": 717 |
| }, |
| { |
| "epoch": 2.56, |
| "learning_rate": 4.29778679884937e-06, |
| "loss": 0.9967, |
| "step": 720 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 4.094304669186761e-06, |
| "loss": 0.9324, |
| "step": 723 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 3.895497100076892e-06, |
| "loss": 0.9994, |
| "step": 726 |
| }, |
| { |
| "epoch": 2.59, |
| "learning_rate": 3.701389974266536e-06, |
| "loss": 0.9551, |
| "step": 729 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 3.512008562552005e-06, |
| "loss": 1.001, |
| "step": 732 |
| }, |
| { |
| "epoch": 2.61, |
| "learning_rate": 3.3273775204891678e-06, |
| "loss": 1.0111, |
| "step": 735 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 3.1475208851834815e-06, |
| "loss": 1.0089, |
| "step": 738 |
| }, |
| { |
| "epoch": 2.63, |
| "learning_rate": 2.972462072160682e-06, |
| "loss": 0.9709, |
| "step": 741 |
| }, |
| { |
| "epoch": 2.64, |
| "learning_rate": 2.8022238723182635e-06, |
| "loss": 0.9676, |
| "step": 744 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 2.6368284489583396e-06, |
| "loss": 0.9643, |
| "step": 747 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 2.4762973349022355e-06, |
| "loss": 0.9753, |
| "step": 750 |
| }, |
| { |
| "epoch": 2.67, |
| "learning_rate": 2.320651429687093e-06, |
| "loss": 0.8655, |
| "step": 753 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 2.169910996844968e-06, |
| "loss": 1.0049, |
| "step": 756 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 2.0240956612647487e-06, |
| "loss": 1.0138, |
| "step": 759 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 1.8832244066371696e-06, |
| "loss": 1.0226, |
| "step": 762 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 1.7473155729833234e-06, |
| "loss": 0.9316, |
| "step": 765 |
| }, |
| { |
| "epoch": 2.73, |
| "learning_rate": 1.616386854266989e-06, |
| "loss": 0.9972, |
| "step": 768 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 1.4904552960910512e-06, |
| "loss": 0.9322, |
| "step": 771 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 1.3695372934783157e-06, |
| "loss": 0.9403, |
| "step": 774 |
| }, |
| { |
| "epoch": 2.76, |
| "learning_rate": 1.2536485887370754e-06, |
| "loss": 0.998, |
| "step": 777 |
| }, |
| { |
| "epoch": 2.77, |
| "learning_rate": 1.142804269411606e-06, |
| "loss": 1.0496, |
| "step": 780 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 1.0370187663178854e-06, |
| "loss": 0.964, |
| "step": 783 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 9.363058516649138e-07, |
| "loss": 0.9786, |
| "step": 786 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 8.406786372616405e-07, |
| "loss": 0.9675, |
| "step": 789 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 7.501495728100061e-07, |
| "loss": 0.9009, |
| "step": 792 |
| }, |
| { |
| "epoch": 2.82, |
| "learning_rate": 6.647304442840652e-07, |
| "loss": 0.9913, |
| "step": 795 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 5.844323723955736e-07, |
| "loss": 1.0268, |
| "step": 798 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 5.092658111462179e-07, |
| "loss": 0.9936, |
| "step": 801 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 4.392405464665883e-07, |
| "loss": 0.9537, |
| "step": 804 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 3.743656949421315e-07, |
| "loss": 1.0263, |
| "step": 807 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 3.14649702626304e-07, |
| "loss": 1.0091, |
| "step": 810 |
| }, |
| { |
| "epoch": 2.89, |
| "learning_rate": 2.6010034394095e-07, |
| "loss": 0.9997, |
| "step": 813 |
| }, |
| { |
| "epoch": 2.9, |
| "learning_rate": 2.1072472066414606e-07, |
| "loss": 1.0062, |
| "step": 816 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 1.6652926100565237e-07, |
| "loss": 0.9276, |
| "step": 819 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 1.2751971876999504e-07, |
| "loss": 0.9591, |
| "step": 822 |
| }, |
| { |
| "epoch": 2.93, |
| "learning_rate": 9.37011726073811e-08, |
| "loss": 0.949, |
| "step": 825 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 6.507802535253405e-08, |
| "loss": 1.0096, |
| "step": 828 |
| }, |
| { |
| "epoch": 2.95, |
| "learning_rate": 4.165400345145898e-08, |
| "loss": 0.9327, |
| "step": 831 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 2.3432156476310698e-08, |
| "loss": 1.0365, |
| "step": 834 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 1.0414856728373501e-08, |
| "loss": 1.0014, |
| "step": 837 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 2.603798929206036e-09, |
| "loss": 0.9712, |
| "step": 840 |
| }, |
| { |
| "epoch": 2.99, |
| "learning_rate": 0.0, |
| "loss": 0.9595, |
| "step": 843 |
| }, |
| { |
| "epoch": 2.99, |
| "eval_gen_len": 29.714570858283434, |
| "eval_loss": 1.013558268547058, |
| "eval_rouge1": 59.9513, |
| "eval_rouge2": 33.9118, |
| "eval_rougeL": 55.7815, |
| "eval_rougeLsum": 56.9064, |
| "eval_runtime": 94.9567, |
| "eval_samples_per_second": 10.552, |
| "eval_steps_per_second": 2.643, |
| "step": 843 |
| }, |
| { |
| "epoch": 2.99, |
| "step": 843, |
| "total_flos": 3.289406245752013e+16, |
| "train_loss": 1.2042206371664859, |
| "train_runtime": 2861.2173, |
| "train_samples_per_second": 18.888, |
| "train_steps_per_second": 0.295 |
| } |
| ], |
| "max_steps": 843, |
| "num_train_epochs": 3, |
| "total_flos": 3.289406245752013e+16, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|