| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 5.0, | |
| "eval_steps": 500, | |
| "global_step": 585, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0, | |
| "forget_Q_A_ROUGE": 0.3188276274014988, | |
| "step": 0 | |
| }, | |
| { | |
| "epoch": 0.042735042735042736, | |
| "grad_norm": 764.0, | |
| "learning_rate": 4.3103448275862073e-07, | |
| "loss": 14.2, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.08547008547008547, | |
| "grad_norm": 652.0, | |
| "learning_rate": 8.620689655172415e-07, | |
| "loss": 14.0957, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.1282051282051282, | |
| "grad_norm": 700.0, | |
| "learning_rate": 1.2931034482758623e-06, | |
| "loss": 13.9925, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.17094017094017094, | |
| "grad_norm": 740.0, | |
| "learning_rate": 1.724137931034483e-06, | |
| "loss": 13.8799, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.21367521367521367, | |
| "grad_norm": 692.0, | |
| "learning_rate": 2.1551724137931035e-06, | |
| "loss": 13.7433, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.2564102564102564, | |
| "grad_norm": 608.0, | |
| "learning_rate": 2.5862068965517246e-06, | |
| "loss": 13.5505, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.29914529914529914, | |
| "grad_norm": 676.0, | |
| "learning_rate": 3.017241379310345e-06, | |
| "loss": 13.3642, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.3418803418803419, | |
| "grad_norm": 760.0, | |
| "learning_rate": 3.448275862068966e-06, | |
| "loss": 12.9769, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.38461538461538464, | |
| "grad_norm": 728.0, | |
| "learning_rate": 3.8793103448275865e-06, | |
| "loss": 11.7681, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.42735042735042733, | |
| "grad_norm": 728.0, | |
| "learning_rate": 4.310344827586207e-06, | |
| "loss": 10.5377, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.4700854700854701, | |
| "grad_norm": 676.0, | |
| "learning_rate": 4.741379310344828e-06, | |
| "loss": 8.6128, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.5128205128205128, | |
| "grad_norm": 664.0, | |
| "learning_rate": 5.172413793103449e-06, | |
| "loss": 6.0024, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.5555555555555556, | |
| "grad_norm": 312.0, | |
| "learning_rate": 5.603448275862069e-06, | |
| "loss": 2.7774, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.5982905982905983, | |
| "grad_norm": 96.5, | |
| "learning_rate": 6.03448275862069e-06, | |
| "loss": 1.5545, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.6410256410256411, | |
| "grad_norm": 79.0, | |
| "learning_rate": 6.465517241379311e-06, | |
| "loss": 1.4243, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.6837606837606838, | |
| "grad_norm": 77.0, | |
| "learning_rate": 6.896551724137932e-06, | |
| "loss": 1.159, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.7264957264957265, | |
| "grad_norm": 81.0, | |
| "learning_rate": 7.327586206896552e-06, | |
| "loss": 1.0177, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.7692307692307693, | |
| "grad_norm": 46.25, | |
| "learning_rate": 7.758620689655173e-06, | |
| "loss": 0.849, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.811965811965812, | |
| "grad_norm": 31.375, | |
| "learning_rate": 8.189655172413794e-06, | |
| "loss": 0.816, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.8547008547008547, | |
| "grad_norm": 39.5, | |
| "learning_rate": 8.620689655172414e-06, | |
| "loss": 0.728, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.8974358974358975, | |
| "grad_norm": 99.0, | |
| "learning_rate": 9.051724137931036e-06, | |
| "loss": 0.7868, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.9401709401709402, | |
| "grad_norm": 26.875, | |
| "learning_rate": 9.482758620689655e-06, | |
| "loss": 0.5595, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.9829059829059829, | |
| "grad_norm": 22.625, | |
| "learning_rate": 9.913793103448277e-06, | |
| "loss": 0.5535, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "forget_Q_A_ROUGE": 0.18181675542335068, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 1.0256410256410255, | |
| "grad_norm": 17.75, | |
| "learning_rate": 9.914712153518125e-06, | |
| "loss": 0.5077, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.0683760683760684, | |
| "grad_norm": 12.125, | |
| "learning_rate": 9.80810234541578e-06, | |
| "loss": 0.4655, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 1.1111111111111112, | |
| "grad_norm": 31.25, | |
| "learning_rate": 9.701492537313434e-06, | |
| "loss": 0.4229, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.1538461538461537, | |
| "grad_norm": 24.125, | |
| "learning_rate": 9.594882729211089e-06, | |
| "loss": 0.4191, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 1.1965811965811965, | |
| "grad_norm": 21.25, | |
| "learning_rate": 9.488272921108744e-06, | |
| "loss": 0.3993, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.2393162393162394, | |
| "grad_norm": 10.75, | |
| "learning_rate": 9.381663113006397e-06, | |
| "loss": 0.3551, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 1.282051282051282, | |
| "grad_norm": 7.71875, | |
| "learning_rate": 9.275053304904051e-06, | |
| "loss": 0.3124, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.3247863247863247, | |
| "grad_norm": 8.0625, | |
| "learning_rate": 9.168443496801706e-06, | |
| "loss": 0.3136, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 1.3675213675213675, | |
| "grad_norm": 15.4375, | |
| "learning_rate": 9.06183368869936e-06, | |
| "loss": 0.3554, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 1.4102564102564101, | |
| "grad_norm": 7.46875, | |
| "learning_rate": 8.955223880597016e-06, | |
| "loss": 0.2964, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 1.452991452991453, | |
| "grad_norm": 62.0, | |
| "learning_rate": 8.84861407249467e-06, | |
| "loss": 0.2874, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 1.4957264957264957, | |
| "grad_norm": 13.9375, | |
| "learning_rate": 8.742004264392325e-06, | |
| "loss": 0.2957, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 1.5384615384615383, | |
| "grad_norm": 8.125, | |
| "learning_rate": 8.63539445628998e-06, | |
| "loss": 0.3001, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.5811965811965814, | |
| "grad_norm": 6.625, | |
| "learning_rate": 8.528784648187633e-06, | |
| "loss": 0.3171, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 1.623931623931624, | |
| "grad_norm": 6.0625, | |
| "learning_rate": 8.42217484008529e-06, | |
| "loss": 0.2736, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.6666666666666665, | |
| "grad_norm": 7.59375, | |
| "learning_rate": 8.315565031982942e-06, | |
| "loss": 0.3145, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 1.7094017094017095, | |
| "grad_norm": 13.4375, | |
| "learning_rate": 8.208955223880599e-06, | |
| "loss": 0.2538, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.7521367521367521, | |
| "grad_norm": 18.75, | |
| "learning_rate": 8.102345415778252e-06, | |
| "loss": 0.2419, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 1.7948717948717947, | |
| "grad_norm": 7.09375, | |
| "learning_rate": 7.995735607675907e-06, | |
| "loss": 0.2446, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.8376068376068377, | |
| "grad_norm": 5.5625, | |
| "learning_rate": 7.889125799573561e-06, | |
| "loss": 0.2172, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 1.8803418803418803, | |
| "grad_norm": 6.75, | |
| "learning_rate": 7.782515991471216e-06, | |
| "loss": 0.2354, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.9230769230769231, | |
| "grad_norm": 6.28125, | |
| "learning_rate": 7.67590618336887e-06, | |
| "loss": 0.2031, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 1.965811965811966, | |
| "grad_norm": 7.09375, | |
| "learning_rate": 7.569296375266525e-06, | |
| "loss": 0.2396, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "forget_Q_A_ROUGE": 0.16664429628381783, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 2.0085470085470085, | |
| "grad_norm": 7.03125, | |
| "learning_rate": 7.46268656716418e-06, | |
| "loss": 0.2064, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 2.051282051282051, | |
| "grad_norm": 5.5625, | |
| "learning_rate": 7.356076759061834e-06, | |
| "loss": 0.1929, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 2.094017094017094, | |
| "grad_norm": 6.4375, | |
| "learning_rate": 7.249466950959488e-06, | |
| "loss": 0.1718, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 2.1367521367521367, | |
| "grad_norm": 6.1875, | |
| "learning_rate": 7.1428571428571436e-06, | |
| "loss": 0.2016, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 2.1794871794871793, | |
| "grad_norm": 6.0, | |
| "learning_rate": 7.0362473347547975e-06, | |
| "loss": 0.1856, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 2.2222222222222223, | |
| "grad_norm": 5.0625, | |
| "learning_rate": 6.929637526652453e-06, | |
| "loss": 0.1694, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 2.264957264957265, | |
| "grad_norm": 5.40625, | |
| "learning_rate": 6.823027718550107e-06, | |
| "loss": 0.1571, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 2.3076923076923075, | |
| "grad_norm": 7.40625, | |
| "learning_rate": 6.7164179104477625e-06, | |
| "loss": 0.1763, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 2.3504273504273505, | |
| "grad_norm": 5.5, | |
| "learning_rate": 6.609808102345416e-06, | |
| "loss": 0.1777, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 2.393162393162393, | |
| "grad_norm": 6.0, | |
| "learning_rate": 6.50319829424307e-06, | |
| "loss": 0.1754, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 2.435897435897436, | |
| "grad_norm": 5.8125, | |
| "learning_rate": 6.396588486140726e-06, | |
| "loss": 0.158, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 2.4786324786324787, | |
| "grad_norm": 6.59375, | |
| "learning_rate": 6.28997867803838e-06, | |
| "loss": 0.1666, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 2.5213675213675213, | |
| "grad_norm": 5.28125, | |
| "learning_rate": 6.183368869936035e-06, | |
| "loss": 0.1561, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 2.564102564102564, | |
| "grad_norm": 5.25, | |
| "learning_rate": 6.076759061833689e-06, | |
| "loss": 0.1373, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 2.606837606837607, | |
| "grad_norm": 6.46875, | |
| "learning_rate": 5.970149253731343e-06, | |
| "loss": 0.1559, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 2.6495726495726495, | |
| "grad_norm": 5.90625, | |
| "learning_rate": 5.863539445628999e-06, | |
| "loss": 0.1453, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 2.6923076923076925, | |
| "grad_norm": 4.9375, | |
| "learning_rate": 5.756929637526653e-06, | |
| "loss": 0.153, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 2.735042735042735, | |
| "grad_norm": 5.5, | |
| "learning_rate": 5.650319829424308e-06, | |
| "loss": 0.1456, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 2.7777777777777777, | |
| "grad_norm": 6.6875, | |
| "learning_rate": 5.543710021321962e-06, | |
| "loss": 0.1375, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 2.8205128205128203, | |
| "grad_norm": 5.125, | |
| "learning_rate": 5.437100213219617e-06, | |
| "loss": 0.1433, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 2.8632478632478633, | |
| "grad_norm": 6.34375, | |
| "learning_rate": 5.3304904051172716e-06, | |
| "loss": 0.1354, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 2.905982905982906, | |
| "grad_norm": 7.5, | |
| "learning_rate": 5.2238805970149255e-06, | |
| "loss": 0.1252, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 2.948717948717949, | |
| "grad_norm": 5.15625, | |
| "learning_rate": 5.11727078891258e-06, | |
| "loss": 0.1273, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 2.9914529914529915, | |
| "grad_norm": 4.53125, | |
| "learning_rate": 5.010660980810235e-06, | |
| "loss": 0.1179, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "forget_Q_A_ROUGE": 0.16380368545856266, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 3.034188034188034, | |
| "grad_norm": 5.4375, | |
| "learning_rate": 4.90405117270789e-06, | |
| "loss": 0.1379, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 3.076923076923077, | |
| "grad_norm": 5.34375, | |
| "learning_rate": 4.797441364605544e-06, | |
| "loss": 0.1277, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 3.1196581196581197, | |
| "grad_norm": 5.0, | |
| "learning_rate": 4.690831556503198e-06, | |
| "loss": 0.1305, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 3.1623931623931623, | |
| "grad_norm": 4.71875, | |
| "learning_rate": 4.584221748400853e-06, | |
| "loss": 0.118, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 3.2051282051282053, | |
| "grad_norm": 4.84375, | |
| "learning_rate": 4.477611940298508e-06, | |
| "loss": 0.1176, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 3.247863247863248, | |
| "grad_norm": 4.75, | |
| "learning_rate": 4.3710021321961625e-06, | |
| "loss": 0.12, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 3.2905982905982905, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 4.264392324093816e-06, | |
| "loss": 0.1288, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 3.3333333333333335, | |
| "grad_norm": 5.34375, | |
| "learning_rate": 4.157782515991471e-06, | |
| "loss": 0.1244, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 3.376068376068376, | |
| "grad_norm": 4.4375, | |
| "learning_rate": 4.051172707889126e-06, | |
| "loss": 0.1189, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 3.4188034188034186, | |
| "grad_norm": 4.71875, | |
| "learning_rate": 3.944562899786781e-06, | |
| "loss": 0.1073, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 3.4615384615384617, | |
| "grad_norm": 4.6875, | |
| "learning_rate": 3.837953091684435e-06, | |
| "loss": 0.1056, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 3.5042735042735043, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 3.73134328358209e-06, | |
| "loss": 0.0968, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 3.547008547008547, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 3.624733475479744e-06, | |
| "loss": 0.114, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 3.58974358974359, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 3.5181236673773987e-06, | |
| "loss": 0.0984, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 3.6324786324786325, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 3.4115138592750535e-06, | |
| "loss": 0.1039, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 3.6752136752136755, | |
| "grad_norm": 5.15625, | |
| "learning_rate": 3.304904051172708e-06, | |
| "loss": 0.1201, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 3.717948717948718, | |
| "grad_norm": 4.78125, | |
| "learning_rate": 3.198294243070363e-06, | |
| "loss": 0.1138, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 3.7606837606837606, | |
| "grad_norm": 4.4375, | |
| "learning_rate": 3.0916844349680177e-06, | |
| "loss": 0.1009, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 3.8034188034188032, | |
| "grad_norm": 5.0, | |
| "learning_rate": 2.9850746268656716e-06, | |
| "loss": 0.1073, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 3.8461538461538463, | |
| "grad_norm": 3.890625, | |
| "learning_rate": 2.8784648187633263e-06, | |
| "loss": 0.1065, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 3.888888888888889, | |
| "grad_norm": 5.65625, | |
| "learning_rate": 2.771855010660981e-06, | |
| "loss": 0.1164, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 3.931623931623932, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 2.6652452025586358e-06, | |
| "loss": 0.0976, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 3.9743589743589745, | |
| "grad_norm": 4.375, | |
| "learning_rate": 2.55863539445629e-06, | |
| "loss": 0.1027, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "forget_Q_A_ROUGE": 0.1577463934994216, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 4.017094017094017, | |
| "grad_norm": 5.03125, | |
| "learning_rate": 2.452025586353945e-06, | |
| "loss": 0.108, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 4.05982905982906, | |
| "grad_norm": 3.53125, | |
| "learning_rate": 2.345415778251599e-06, | |
| "loss": 0.0954, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 4.102564102564102, | |
| "grad_norm": 4.53125, | |
| "learning_rate": 2.238805970149254e-06, | |
| "loss": 0.107, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 4.145299145299146, | |
| "grad_norm": 4.625, | |
| "learning_rate": 2.132196162046908e-06, | |
| "loss": 0.1001, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 4.188034188034188, | |
| "grad_norm": 5.28125, | |
| "learning_rate": 2.025586353944563e-06, | |
| "loss": 0.109, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 4.230769230769231, | |
| "grad_norm": 4.4375, | |
| "learning_rate": 1.9189765458422177e-06, | |
| "loss": 0.1133, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 4.273504273504273, | |
| "grad_norm": 4.0, | |
| "learning_rate": 1.812366737739872e-06, | |
| "loss": 0.092, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 4.316239316239316, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 1.7057569296375267e-06, | |
| "loss": 0.1034, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 4.358974358974359, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 1.5991471215351815e-06, | |
| "loss": 0.0986, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 4.401709401709402, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 1.4925373134328358e-06, | |
| "loss": 0.1168, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 4.444444444444445, | |
| "grad_norm": 4.90625, | |
| "learning_rate": 1.3859275053304905e-06, | |
| "loss": 0.1033, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 4.487179487179487, | |
| "grad_norm": 4.65625, | |
| "learning_rate": 1.279317697228145e-06, | |
| "loss": 0.1006, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 4.52991452991453, | |
| "grad_norm": 5.21875, | |
| "learning_rate": 1.1727078891257996e-06, | |
| "loss": 0.109, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 4.572649572649572, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 1.066098081023454e-06, | |
| "loss": 0.1129, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 4.615384615384615, | |
| "grad_norm": 5.34375, | |
| "learning_rate": 9.594882729211088e-07, | |
| "loss": 0.0991, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 4.6581196581196584, | |
| "grad_norm": 4.84375, | |
| "learning_rate": 8.528784648187634e-07, | |
| "loss": 0.1047, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 4.700854700854701, | |
| "grad_norm": 5.59375, | |
| "learning_rate": 7.462686567164179e-07, | |
| "loss": 0.1099, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 4.743589743589744, | |
| "grad_norm": 3.984375, | |
| "learning_rate": 6.396588486140725e-07, | |
| "loss": 0.0981, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 4.786324786324786, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 5.33049040511727e-07, | |
| "loss": 0.0975, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 4.829059829059829, | |
| "grad_norm": 5.40625, | |
| "learning_rate": 4.264392324093817e-07, | |
| "loss": 0.1066, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 4.871794871794872, | |
| "grad_norm": 4.53125, | |
| "learning_rate": 3.1982942430703626e-07, | |
| "loss": 0.1064, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 4.914529914529915, | |
| "grad_norm": 3.953125, | |
| "learning_rate": 2.1321961620469084e-07, | |
| "loss": 0.0975, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 4.957264957264957, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 1.0660980810234542e-07, | |
| "loss": 0.1096, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 0.0, | |
| "loss": 0.0957, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "forget_Q_A_ROUGE": 0.16323620599043673, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "step": 585, | |
| "total_flos": 0.0, | |
| "train_loss": 1.4970051565231421, | |
| "train_runtime": 6804.7106, | |
| "train_samples_per_second": 5.499, | |
| "train_steps_per_second": 0.086 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 585, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": false, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |