diff --git "a/checkpoint-14000/trainer_state.json" "b/checkpoint-14000/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoint-14000/trainer_state.json" @@ -0,0 +1,9834 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.0295748613678373, + "eval_steps": 500, + "global_step": 14000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0014497481062665362, + "grad_norm": 2.875, + "learning_rate": 1.9987142857142856e-06, + "loss": 1.1375, + "step": 10 + }, + { + "epoch": 0.0028994962125330724, + "grad_norm": 2.953125, + "learning_rate": 1.9972857142857144e-06, + "loss": 1.2129, + "step": 20 + }, + { + "epoch": 0.004349244318799609, + "grad_norm": 2.390625, + "learning_rate": 1.9958571428571427e-06, + "loss": 1.1235, + "step": 30 + }, + { + "epoch": 0.005798992425066145, + "grad_norm": 2.515625, + "learning_rate": 1.9944285714285715e-06, + "loss": 1.0452, + "step": 40 + }, + { + "epoch": 0.007248740531332681, + "grad_norm": 2.375, + "learning_rate": 1.993e-06, + "loss": 1.0621, + "step": 50 + }, + { + "epoch": 0.008698488637599217, + "grad_norm": 2.140625, + "learning_rate": 1.9915714285714286e-06, + "loss": 1.0783, + "step": 60 + }, + { + "epoch": 0.010148236743865753, + "grad_norm": 1.90625, + "learning_rate": 1.990142857142857e-06, + "loss": 1.1023, + "step": 70 + }, + { + "epoch": 0.01159798485013229, + "grad_norm": 2.265625, + "learning_rate": 1.9887142857142857e-06, + "loss": 1.0718, + "step": 80 + }, + { + "epoch": 0.013047732956398826, + "grad_norm": 1.9765625, + "learning_rate": 1.987285714285714e-06, + "loss": 1.0128, + "step": 90 + }, + { + "epoch": 0.014497481062665362, + "grad_norm": 1.8359375, + "learning_rate": 1.985857142857143e-06, + "loss": 1.0176, + "step": 100 + }, + { + "epoch": 0.015947229168931897, + "grad_norm": 2.0, + "learning_rate": 1.984428571428571e-06, + "loss": 1.014, + "step": 110 + }, + { + "epoch": 0.017396977275198434, + "grad_norm": 2.3125, + "learning_rate": 1.983e-06, + "loss": 1.0257, + "step": 120 + }, + { + "epoch": 0.01884672538146497, + "grad_norm": 3.015625, + "learning_rate": 1.9815714285714282e-06, + "loss": 1.0194, + "step": 130 + }, + { + "epoch": 0.020296473487731507, + "grad_norm": 1.6796875, + "learning_rate": 1.980142857142857e-06, + "loss": 1.0052, + "step": 140 + }, + { + "epoch": 0.02174622159399804, + "grad_norm": 2.09375, + "learning_rate": 1.9787142857142854e-06, + "loss": 1.0326, + "step": 150 + }, + { + "epoch": 0.02319596970026458, + "grad_norm": 2.15625, + "learning_rate": 1.977285714285714e-06, + "loss": 1.0111, + "step": 160 + }, + { + "epoch": 0.024645717806531114, + "grad_norm": 2.125, + "learning_rate": 1.9758571428571425e-06, + "loss": 0.979, + "step": 170 + }, + { + "epoch": 0.02609546591279765, + "grad_norm": 2.125, + "learning_rate": 1.9744285714285712e-06, + "loss": 0.9513, + "step": 180 + }, + { + "epoch": 0.027545214019064186, + "grad_norm": 2.421875, + "learning_rate": 1.973e-06, + "loss": 0.931, + "step": 190 + }, + { + "epoch": 0.028994962125330724, + "grad_norm": 2.421875, + "learning_rate": 1.9715714285714283e-06, + "loss": 0.9783, + "step": 200 + }, + { + "epoch": 0.03044471023159726, + "grad_norm": 2.4375, + "learning_rate": 1.970142857142857e-06, + "loss": 0.9473, + "step": 210 + }, + { + "epoch": 0.03189445833786379, + "grad_norm": 2.65625, + "learning_rate": 1.9687142857142855e-06, + "loss": 0.9434, + "step": 220 + }, + { + "epoch": 0.03334420644413033, + "grad_norm": 2.84375, + "learning_rate": 1.9672857142857142e-06, + "loss": 0.9568, + "step": 230 + }, + { + "epoch": 0.03479395455039687, + "grad_norm": 3.359375, + "learning_rate": 1.9658571428571426e-06, + "loss": 0.943, + "step": 240 + }, + { + "epoch": 0.03624370265666341, + "grad_norm": 2.375, + "learning_rate": 1.9644285714285713e-06, + "loss": 0.9249, + "step": 250 + }, + { + "epoch": 0.03769345076292994, + "grad_norm": 2.5625, + "learning_rate": 1.963e-06, + "loss": 0.9565, + "step": 260 + }, + { + "epoch": 0.039143198869196476, + "grad_norm": 2.390625, + "learning_rate": 1.9615714285714284e-06, + "loss": 0.9772, + "step": 270 + }, + { + "epoch": 0.040592946975463014, + "grad_norm": 2.484375, + "learning_rate": 1.960142857142857e-06, + "loss": 0.8959, + "step": 280 + }, + { + "epoch": 0.04204269508172955, + "grad_norm": 2.015625, + "learning_rate": 1.9587142857142855e-06, + "loss": 0.8747, + "step": 290 + }, + { + "epoch": 0.04349244318799608, + "grad_norm": 2.75, + "learning_rate": 1.9572857142857143e-06, + "loss": 0.8972, + "step": 300 + }, + { + "epoch": 0.04494219129426262, + "grad_norm": 2.53125, + "learning_rate": 1.955857142857143e-06, + "loss": 0.8711, + "step": 310 + }, + { + "epoch": 0.04639193940052916, + "grad_norm": 2.5625, + "learning_rate": 1.9544285714285714e-06, + "loss": 0.8684, + "step": 320 + }, + { + "epoch": 0.047841687506795696, + "grad_norm": 2.421875, + "learning_rate": 1.953e-06, + "loss": 0.9538, + "step": 330 + }, + { + "epoch": 0.04929143561306223, + "grad_norm": 2.5625, + "learning_rate": 1.9515714285714285e-06, + "loss": 0.917, + "step": 340 + }, + { + "epoch": 0.050741183719328765, + "grad_norm": 2.5, + "learning_rate": 1.9501428571428573e-06, + "loss": 0.9764, + "step": 350 + }, + { + "epoch": 0.0521909318255953, + "grad_norm": 2.5625, + "learning_rate": 1.9487142857142856e-06, + "loss": 0.9164, + "step": 360 + }, + { + "epoch": 0.05364067993186184, + "grad_norm": 2.5, + "learning_rate": 1.9472857142857144e-06, + "loss": 0.9172, + "step": 370 + }, + { + "epoch": 0.05509042803812837, + "grad_norm": 2.609375, + "learning_rate": 1.9458571428571428e-06, + "loss": 0.8655, + "step": 380 + }, + { + "epoch": 0.05654017614439491, + "grad_norm": 2.328125, + "learning_rate": 1.944428571428571e-06, + "loss": 0.9029, + "step": 390 + }, + { + "epoch": 0.05798992425066145, + "grad_norm": 2.5625, + "learning_rate": 1.943e-06, + "loss": 0.9203, + "step": 400 + }, + { + "epoch": 0.059439672356927986, + "grad_norm": 2.828125, + "learning_rate": 1.941571428571428e-06, + "loss": 0.9151, + "step": 410 + }, + { + "epoch": 0.06088942046319452, + "grad_norm": 3.03125, + "learning_rate": 1.940142857142857e-06, + "loss": 0.8941, + "step": 420 + }, + { + "epoch": 0.062339168569461055, + "grad_norm": 2.359375, + "learning_rate": 1.9387142857142853e-06, + "loss": 0.7872, + "step": 430 + }, + { + "epoch": 0.06378891667572759, + "grad_norm": 2.421875, + "learning_rate": 1.937285714285714e-06, + "loss": 0.8826, + "step": 440 + }, + { + "epoch": 0.06523866478199412, + "grad_norm": 2.6875, + "learning_rate": 1.935857142857143e-06, + "loss": 0.8736, + "step": 450 + }, + { + "epoch": 0.06668841288826066, + "grad_norm": 2.890625, + "learning_rate": 1.934428571428571e-06, + "loss": 0.8882, + "step": 460 + }, + { + "epoch": 0.0681381609945272, + "grad_norm": 2.546875, + "learning_rate": 1.933e-06, + "loss": 0.8923, + "step": 470 + }, + { + "epoch": 0.06958790910079374, + "grad_norm": 2.3125, + "learning_rate": 1.9315714285714283e-06, + "loss": 0.9457, + "step": 480 + }, + { + "epoch": 0.07103765720706028, + "grad_norm": 2.875, + "learning_rate": 1.930142857142857e-06, + "loss": 0.887, + "step": 490 + }, + { + "epoch": 0.07248740531332681, + "grad_norm": 2.359375, + "learning_rate": 1.9287142857142854e-06, + "loss": 0.9032, + "step": 500 + }, + { + "epoch": 0.07393715341959335, + "grad_norm": 3.5, + "learning_rate": 1.927285714285714e-06, + "loss": 0.9526, + "step": 510 + }, + { + "epoch": 0.07538690152585988, + "grad_norm": 2.53125, + "learning_rate": 1.925857142857143e-06, + "loss": 0.8792, + "step": 520 + }, + { + "epoch": 0.07683664963212641, + "grad_norm": 2.78125, + "learning_rate": 1.9244285714285713e-06, + "loss": 0.9215, + "step": 530 + }, + { + "epoch": 0.07828639773839295, + "grad_norm": 2.578125, + "learning_rate": 1.923e-06, + "loss": 0.8741, + "step": 540 + }, + { + "epoch": 0.07973614584465949, + "grad_norm": 2.390625, + "learning_rate": 1.9215714285714284e-06, + "loss": 0.8701, + "step": 550 + }, + { + "epoch": 0.08118589395092603, + "grad_norm": 2.40625, + "learning_rate": 1.920142857142857e-06, + "loss": 0.8203, + "step": 560 + }, + { + "epoch": 0.08263564205719257, + "grad_norm": 2.484375, + "learning_rate": 1.918714285714286e-06, + "loss": 0.8338, + "step": 570 + }, + { + "epoch": 0.0840853901634591, + "grad_norm": 2.265625, + "learning_rate": 1.9172857142857143e-06, + "loss": 0.8522, + "step": 580 + }, + { + "epoch": 0.08553513826972564, + "grad_norm": 2.796875, + "learning_rate": 1.915857142857143e-06, + "loss": 0.8359, + "step": 590 + }, + { + "epoch": 0.08698488637599217, + "grad_norm": 2.71875, + "learning_rate": 1.9144285714285714e-06, + "loss": 0.8508, + "step": 600 + }, + { + "epoch": 0.0884346344822587, + "grad_norm": 2.5625, + "learning_rate": 1.913e-06, + "loss": 0.8721, + "step": 610 + }, + { + "epoch": 0.08988438258852524, + "grad_norm": 2.671875, + "learning_rate": 1.9115714285714285e-06, + "loss": 0.8673, + "step": 620 + }, + { + "epoch": 0.09133413069479178, + "grad_norm": 2.40625, + "learning_rate": 1.9101428571428573e-06, + "loss": 0.925, + "step": 630 + }, + { + "epoch": 0.09278387880105832, + "grad_norm": 20.375, + "learning_rate": 1.9087142857142856e-06, + "loss": 0.782, + "step": 640 + }, + { + "epoch": 0.09423362690732486, + "grad_norm": 2.46875, + "learning_rate": 1.9072857142857142e-06, + "loss": 0.8539, + "step": 650 + }, + { + "epoch": 0.09568337501359139, + "grad_norm": 2.546875, + "learning_rate": 1.905857142857143e-06, + "loss": 0.854, + "step": 660 + }, + { + "epoch": 0.09713312311985793, + "grad_norm": 3.328125, + "learning_rate": 1.9044285714285713e-06, + "loss": 0.91, + "step": 670 + }, + { + "epoch": 0.09858287122612445, + "grad_norm": 3.109375, + "learning_rate": 1.903e-06, + "loss": 0.8562, + "step": 680 + }, + { + "epoch": 0.10003261933239099, + "grad_norm": 2.640625, + "learning_rate": 1.9015714285714284e-06, + "loss": 0.885, + "step": 690 + }, + { + "epoch": 0.10148236743865753, + "grad_norm": 2.5625, + "learning_rate": 1.9001428571428571e-06, + "loss": 0.8854, + "step": 700 + }, + { + "epoch": 0.10293211554492407, + "grad_norm": 2.5, + "learning_rate": 1.8987142857142857e-06, + "loss": 0.835, + "step": 710 + }, + { + "epoch": 0.1043818636511906, + "grad_norm": 2.28125, + "learning_rate": 1.8972857142857143e-06, + "loss": 0.8941, + "step": 720 + }, + { + "epoch": 0.10583161175745714, + "grad_norm": 2.171875, + "learning_rate": 1.8958571428571428e-06, + "loss": 0.8378, + "step": 730 + }, + { + "epoch": 0.10728135986372368, + "grad_norm": 2.734375, + "learning_rate": 1.8944285714285714e-06, + "loss": 0.831, + "step": 740 + }, + { + "epoch": 0.10873110796999022, + "grad_norm": 2.453125, + "learning_rate": 1.893e-06, + "loss": 0.847, + "step": 750 + }, + { + "epoch": 0.11018085607625674, + "grad_norm": 2.59375, + "learning_rate": 1.8915714285714287e-06, + "loss": 0.8401, + "step": 760 + }, + { + "epoch": 0.11163060418252328, + "grad_norm": 2.859375, + "learning_rate": 1.890142857142857e-06, + "loss": 0.8815, + "step": 770 + }, + { + "epoch": 0.11308035228878982, + "grad_norm": 3.4375, + "learning_rate": 1.8887142857142858e-06, + "loss": 0.8646, + "step": 780 + }, + { + "epoch": 0.11453010039505636, + "grad_norm": 2.1875, + "learning_rate": 1.8872857142857141e-06, + "loss": 0.852, + "step": 790 + }, + { + "epoch": 0.1159798485013229, + "grad_norm": 2.78125, + "learning_rate": 1.885857142857143e-06, + "loss": 0.8398, + "step": 800 + }, + { + "epoch": 0.11742959660758943, + "grad_norm": 2.21875, + "learning_rate": 1.8844285714285712e-06, + "loss": 0.8127, + "step": 810 + }, + { + "epoch": 0.11887934471385597, + "grad_norm": 3.5, + "learning_rate": 1.883e-06, + "loss": 0.8451, + "step": 820 + }, + { + "epoch": 0.12032909282012251, + "grad_norm": 2.4375, + "learning_rate": 1.8815714285714286e-06, + "loss": 0.8524, + "step": 830 + }, + { + "epoch": 0.12177884092638903, + "grad_norm": 2.5, + "learning_rate": 1.8801428571428571e-06, + "loss": 0.8272, + "step": 840 + }, + { + "epoch": 0.12322858903265557, + "grad_norm": 2.328125, + "learning_rate": 1.8787142857142857e-06, + "loss": 0.8707, + "step": 850 + }, + { + "epoch": 0.12467833713892211, + "grad_norm": 2.359375, + "learning_rate": 1.8772857142857142e-06, + "loss": 0.8004, + "step": 860 + }, + { + "epoch": 0.12612808524518865, + "grad_norm": 2.8125, + "learning_rate": 1.8758571428571428e-06, + "loss": 0.899, + "step": 870 + }, + { + "epoch": 0.12757783335145517, + "grad_norm": 2.546875, + "learning_rate": 1.8744285714285713e-06, + "loss": 0.8591, + "step": 880 + }, + { + "epoch": 0.12902758145772172, + "grad_norm": 2.296875, + "learning_rate": 1.873e-06, + "loss": 0.8044, + "step": 890 + }, + { + "epoch": 0.13047732956398825, + "grad_norm": 2.84375, + "learning_rate": 1.8715714285714287e-06, + "loss": 0.8738, + "step": 900 + }, + { + "epoch": 0.1319270776702548, + "grad_norm": 2.640625, + "learning_rate": 1.870142857142857e-06, + "loss": 0.8545, + "step": 910 + }, + { + "epoch": 0.13337682577652132, + "grad_norm": 2.375, + "learning_rate": 1.8687142857142858e-06, + "loss": 0.8652, + "step": 920 + }, + { + "epoch": 0.13482657388278788, + "grad_norm": 2.4375, + "learning_rate": 1.8672857142857141e-06, + "loss": 0.8423, + "step": 930 + }, + { + "epoch": 0.1362763219890544, + "grad_norm": 2.921875, + "learning_rate": 1.8658571428571429e-06, + "loss": 0.7856, + "step": 940 + }, + { + "epoch": 0.13772607009532095, + "grad_norm": 2.5625, + "learning_rate": 1.8644285714285712e-06, + "loss": 0.8558, + "step": 950 + }, + { + "epoch": 0.13917581820158748, + "grad_norm": 2.25, + "learning_rate": 1.863e-06, + "loss": 0.7797, + "step": 960 + }, + { + "epoch": 0.140625566307854, + "grad_norm": 2.609375, + "learning_rate": 1.8615714285714285e-06, + "loss": 0.8118, + "step": 970 + }, + { + "epoch": 0.14207531441412055, + "grad_norm": 2.625, + "learning_rate": 1.860142857142857e-06, + "loss": 0.8365, + "step": 980 + }, + { + "epoch": 0.14352506252038708, + "grad_norm": 2.515625, + "learning_rate": 1.8587142857142857e-06, + "loss": 0.8777, + "step": 990 + }, + { + "epoch": 0.14497481062665363, + "grad_norm": 2.5625, + "learning_rate": 1.8572857142857142e-06, + "loss": 0.8701, + "step": 1000 + }, + { + "epoch": 0.14642455873292015, + "grad_norm": 2.4375, + "learning_rate": 1.8558571428571428e-06, + "loss": 0.8523, + "step": 1010 + }, + { + "epoch": 0.1478743068391867, + "grad_norm": 3.03125, + "learning_rate": 1.8544285714285715e-06, + "loss": 0.8182, + "step": 1020 + }, + { + "epoch": 0.14932405494545323, + "grad_norm": 2.375, + "learning_rate": 1.8529999999999999e-06, + "loss": 0.8119, + "step": 1030 + }, + { + "epoch": 0.15077380305171975, + "grad_norm": 2.578125, + "learning_rate": 1.8515714285714286e-06, + "loss": 0.8361, + "step": 1040 + }, + { + "epoch": 0.1522235511579863, + "grad_norm": 2.40625, + "learning_rate": 1.850142857142857e-06, + "loss": 0.8576, + "step": 1050 + }, + { + "epoch": 0.15367329926425283, + "grad_norm": 2.109375, + "learning_rate": 1.8487142857142858e-06, + "loss": 0.8071, + "step": 1060 + }, + { + "epoch": 0.15512304737051938, + "grad_norm": 3.09375, + "learning_rate": 1.847285714285714e-06, + "loss": 0.8253, + "step": 1070 + }, + { + "epoch": 0.1565727954767859, + "grad_norm": 2.328125, + "learning_rate": 1.8458571428571429e-06, + "loss": 0.8162, + "step": 1080 + }, + { + "epoch": 0.15802254358305245, + "grad_norm": 2.703125, + "learning_rate": 1.8444285714285714e-06, + "loss": 0.872, + "step": 1090 + }, + { + "epoch": 0.15947229168931898, + "grad_norm": 2.609375, + "learning_rate": 1.843e-06, + "loss": 0.8784, + "step": 1100 + }, + { + "epoch": 0.16092203979558553, + "grad_norm": 2.453125, + "learning_rate": 1.8415714285714285e-06, + "loss": 0.8684, + "step": 1110 + }, + { + "epoch": 0.16237178790185205, + "grad_norm": 3.28125, + "learning_rate": 1.840142857142857e-06, + "loss": 0.8648, + "step": 1120 + }, + { + "epoch": 0.16382153600811858, + "grad_norm": 2.265625, + "learning_rate": 1.8387142857142856e-06, + "loss": 0.7867, + "step": 1130 + }, + { + "epoch": 0.16527128411438513, + "grad_norm": 2.328125, + "learning_rate": 1.8372857142857142e-06, + "loss": 0.8465, + "step": 1140 + }, + { + "epoch": 0.16672103222065165, + "grad_norm": 2.1875, + "learning_rate": 1.8358571428571427e-06, + "loss": 0.8598, + "step": 1150 + }, + { + "epoch": 0.1681707803269182, + "grad_norm": 2.5, + "learning_rate": 1.8344285714285715e-06, + "loss": 0.7857, + "step": 1160 + }, + { + "epoch": 0.16962052843318473, + "grad_norm": 2.421875, + "learning_rate": 1.8329999999999999e-06, + "loss": 0.8571, + "step": 1170 + }, + { + "epoch": 0.17107027653945128, + "grad_norm": 2.875, + "learning_rate": 1.8315714285714286e-06, + "loss": 0.8597, + "step": 1180 + }, + { + "epoch": 0.1725200246457178, + "grad_norm": 2.421875, + "learning_rate": 1.830142857142857e-06, + "loss": 0.8475, + "step": 1190 + }, + { + "epoch": 0.17396977275198433, + "grad_norm": 2.359375, + "learning_rate": 1.8287142857142857e-06, + "loss": 0.8534, + "step": 1200 + }, + { + "epoch": 0.17541952085825088, + "grad_norm": 2.484375, + "learning_rate": 1.827285714285714e-06, + "loss": 0.8545, + "step": 1210 + }, + { + "epoch": 0.1768692689645174, + "grad_norm": 2.203125, + "learning_rate": 1.8258571428571428e-06, + "loss": 0.7938, + "step": 1220 + }, + { + "epoch": 0.17831901707078396, + "grad_norm": 2.703125, + "learning_rate": 1.8244285714285714e-06, + "loss": 0.7759, + "step": 1230 + }, + { + "epoch": 0.17976876517705048, + "grad_norm": 2.546875, + "learning_rate": 1.823e-06, + "loss": 0.8706, + "step": 1240 + }, + { + "epoch": 0.18121851328331703, + "grad_norm": 2.46875, + "learning_rate": 1.8215714285714285e-06, + "loss": 0.8412, + "step": 1250 + }, + { + "epoch": 0.18266826138958356, + "grad_norm": 2.609375, + "learning_rate": 1.820142857142857e-06, + "loss": 0.8472, + "step": 1260 + }, + { + "epoch": 0.18411800949585008, + "grad_norm": 2.234375, + "learning_rate": 1.8187142857142856e-06, + "loss": 0.8629, + "step": 1270 + }, + { + "epoch": 0.18556775760211663, + "grad_norm": 2.828125, + "learning_rate": 1.8172857142857144e-06, + "loss": 0.8426, + "step": 1280 + }, + { + "epoch": 0.18701750570838316, + "grad_norm": 2.078125, + "learning_rate": 1.8158571428571427e-06, + "loss": 0.8221, + "step": 1290 + }, + { + "epoch": 0.1884672538146497, + "grad_norm": 2.234375, + "learning_rate": 1.8144285714285715e-06, + "loss": 0.8036, + "step": 1300 + }, + { + "epoch": 0.18991700192091623, + "grad_norm": 2.5625, + "learning_rate": 1.8129999999999998e-06, + "loss": 0.8334, + "step": 1310 + }, + { + "epoch": 0.19136675002718279, + "grad_norm": 2.84375, + "learning_rate": 1.8115714285714286e-06, + "loss": 0.8216, + "step": 1320 + }, + { + "epoch": 0.1928164981334493, + "grad_norm": 2.125, + "learning_rate": 1.810142857142857e-06, + "loss": 0.8597, + "step": 1330 + }, + { + "epoch": 0.19426624623971586, + "grad_norm": 2.5, + "learning_rate": 1.8087142857142857e-06, + "loss": 0.8585, + "step": 1340 + }, + { + "epoch": 0.19571599434598239, + "grad_norm": 2.453125, + "learning_rate": 1.8072857142857143e-06, + "loss": 0.7812, + "step": 1350 + }, + { + "epoch": 0.1971657424522489, + "grad_norm": 2.953125, + "learning_rate": 1.8058571428571428e-06, + "loss": 0.8329, + "step": 1360 + }, + { + "epoch": 0.19861549055851546, + "grad_norm": 3.03125, + "learning_rate": 1.8044285714285714e-06, + "loss": 0.8393, + "step": 1370 + }, + { + "epoch": 0.20006523866478199, + "grad_norm": 2.546875, + "learning_rate": 1.803e-06, + "loss": 0.8863, + "step": 1380 + }, + { + "epoch": 0.20151498677104854, + "grad_norm": 2.453125, + "learning_rate": 1.8015714285714285e-06, + "loss": 0.8211, + "step": 1390 + }, + { + "epoch": 0.20296473487731506, + "grad_norm": 2.546875, + "learning_rate": 1.800142857142857e-06, + "loss": 0.8667, + "step": 1400 + }, + { + "epoch": 0.2044144829835816, + "grad_norm": 2.671875, + "learning_rate": 1.7987142857142856e-06, + "loss": 0.8191, + "step": 1410 + }, + { + "epoch": 0.20586423108984814, + "grad_norm": 2.953125, + "learning_rate": 1.7972857142857144e-06, + "loss": 0.8361, + "step": 1420 + }, + { + "epoch": 0.20731397919611466, + "grad_norm": 2.484375, + "learning_rate": 1.7958571428571427e-06, + "loss": 0.7969, + "step": 1430 + }, + { + "epoch": 0.2087637273023812, + "grad_norm": 2.75, + "learning_rate": 1.7944285714285715e-06, + "loss": 0.874, + "step": 1440 + }, + { + "epoch": 0.21021347540864774, + "grad_norm": 2.421875, + "learning_rate": 1.7929999999999998e-06, + "loss": 0.8136, + "step": 1450 + }, + { + "epoch": 0.2116632235149143, + "grad_norm": 2.09375, + "learning_rate": 1.7915714285714286e-06, + "loss": 0.8695, + "step": 1460 + }, + { + "epoch": 0.2131129716211808, + "grad_norm": 2.296875, + "learning_rate": 1.790142857142857e-06, + "loss": 0.7892, + "step": 1470 + }, + { + "epoch": 0.21456271972744737, + "grad_norm": 2.109375, + "learning_rate": 1.7887142857142857e-06, + "loss": 0.8258, + "step": 1480 + }, + { + "epoch": 0.2160124678337139, + "grad_norm": 2.390625, + "learning_rate": 1.7872857142857142e-06, + "loss": 0.8091, + "step": 1490 + }, + { + "epoch": 0.21746221593998044, + "grad_norm": 3.0, + "learning_rate": 1.7858571428571428e-06, + "loss": 0.7505, + "step": 1500 + }, + { + "epoch": 0.21891196404624697, + "grad_norm": 2.65625, + "learning_rate": 1.7844285714285714e-06, + "loss": 0.8311, + "step": 1510 + }, + { + "epoch": 0.2203617121525135, + "grad_norm": 2.296875, + "learning_rate": 1.783e-06, + "loss": 0.7984, + "step": 1520 + }, + { + "epoch": 0.22181146025878004, + "grad_norm": 2.984375, + "learning_rate": 1.7815714285714285e-06, + "loss": 0.8338, + "step": 1530 + }, + { + "epoch": 0.22326120836504657, + "grad_norm": 3.0, + "learning_rate": 1.7801428571428572e-06, + "loss": 0.8754, + "step": 1540 + }, + { + "epoch": 0.22471095647131312, + "grad_norm": 2.359375, + "learning_rate": 1.7787142857142856e-06, + "loss": 0.8289, + "step": 1550 + }, + { + "epoch": 0.22616070457757964, + "grad_norm": 2.65625, + "learning_rate": 1.7772857142857143e-06, + "loss": 0.8351, + "step": 1560 + }, + { + "epoch": 0.2276104526838462, + "grad_norm": 2.59375, + "learning_rate": 1.7758571428571427e-06, + "loss": 0.8523, + "step": 1570 + }, + { + "epoch": 0.22906020079011272, + "grad_norm": 2.625, + "learning_rate": 1.7744285714285715e-06, + "loss": 0.8484, + "step": 1580 + }, + { + "epoch": 0.23050994889637924, + "grad_norm": 2.90625, + "learning_rate": 1.7729999999999998e-06, + "loss": 0.787, + "step": 1590 + }, + { + "epoch": 0.2319596970026458, + "grad_norm": 2.21875, + "learning_rate": 1.7715714285714286e-06, + "loss": 0.8085, + "step": 1600 + }, + { + "epoch": 0.23340944510891232, + "grad_norm": 2.5, + "learning_rate": 1.7701428571428571e-06, + "loss": 0.7822, + "step": 1610 + }, + { + "epoch": 0.23485919321517887, + "grad_norm": 2.484375, + "learning_rate": 1.7687142857142857e-06, + "loss": 0.8533, + "step": 1620 + }, + { + "epoch": 0.2363089413214454, + "grad_norm": 2.6875, + "learning_rate": 1.7672857142857142e-06, + "loss": 0.8305, + "step": 1630 + }, + { + "epoch": 0.23775868942771194, + "grad_norm": 2.796875, + "learning_rate": 1.7658571428571428e-06, + "loss": 0.8332, + "step": 1640 + }, + { + "epoch": 0.23920843753397847, + "grad_norm": 2.421875, + "learning_rate": 1.7644285714285713e-06, + "loss": 0.8734, + "step": 1650 + }, + { + "epoch": 0.24065818564024502, + "grad_norm": 3.171875, + "learning_rate": 1.7629999999999999e-06, + "loss": 0.8853, + "step": 1660 + }, + { + "epoch": 0.24210793374651154, + "grad_norm": 2.84375, + "learning_rate": 1.7615714285714284e-06, + "loss": 0.7921, + "step": 1670 + }, + { + "epoch": 0.24355768185277807, + "grad_norm": 2.609375, + "learning_rate": 1.7601428571428572e-06, + "loss": 0.8146, + "step": 1680 + }, + { + "epoch": 0.24500742995904462, + "grad_norm": 3.15625, + "learning_rate": 1.7587142857142856e-06, + "loss": 0.8655, + "step": 1690 + }, + { + "epoch": 0.24645717806531114, + "grad_norm": 2.65625, + "learning_rate": 1.7572857142857143e-06, + "loss": 0.761, + "step": 1700 + }, + { + "epoch": 0.2479069261715777, + "grad_norm": 2.65625, + "learning_rate": 1.7558571428571427e-06, + "loss": 0.798, + "step": 1710 + }, + { + "epoch": 0.24935667427784422, + "grad_norm": 3.03125, + "learning_rate": 1.7544285714285714e-06, + "loss": 0.836, + "step": 1720 + }, + { + "epoch": 0.25080642238411077, + "grad_norm": 2.5625, + "learning_rate": 1.7529999999999998e-06, + "loss": 0.8426, + "step": 1730 + }, + { + "epoch": 0.2522561704903773, + "grad_norm": 2.484375, + "learning_rate": 1.7515714285714285e-06, + "loss": 0.8221, + "step": 1740 + }, + { + "epoch": 0.2537059185966438, + "grad_norm": 2.234375, + "learning_rate": 1.750142857142857e-06, + "loss": 0.8206, + "step": 1750 + }, + { + "epoch": 0.25515566670291034, + "grad_norm": 1.984375, + "learning_rate": 1.7487142857142857e-06, + "loss": 0.8332, + "step": 1760 + }, + { + "epoch": 0.2566054148091769, + "grad_norm": 2.890625, + "learning_rate": 1.7472857142857142e-06, + "loss": 0.8594, + "step": 1770 + }, + { + "epoch": 0.25805516291544345, + "grad_norm": 2.671875, + "learning_rate": 1.7458571428571428e-06, + "loss": 0.8025, + "step": 1780 + }, + { + "epoch": 0.25950491102170997, + "grad_norm": 2.75, + "learning_rate": 1.7444285714285713e-06, + "loss": 0.8803, + "step": 1790 + }, + { + "epoch": 0.2609546591279765, + "grad_norm": 2.65625, + "learning_rate": 1.743e-06, + "loss": 0.8044, + "step": 1800 + }, + { + "epoch": 0.2624044072342431, + "grad_norm": 2.984375, + "learning_rate": 1.7415714285714284e-06, + "loss": 0.8391, + "step": 1810 + }, + { + "epoch": 0.2638541553405096, + "grad_norm": 2.3125, + "learning_rate": 1.7401428571428572e-06, + "loss": 0.794, + "step": 1820 + }, + { + "epoch": 0.2653039034467761, + "grad_norm": 2.84375, + "learning_rate": 1.7387142857142855e-06, + "loss": 0.8827, + "step": 1830 + }, + { + "epoch": 0.26675365155304265, + "grad_norm": 2.375, + "learning_rate": 1.7372857142857143e-06, + "loss": 0.8006, + "step": 1840 + }, + { + "epoch": 0.26820339965930917, + "grad_norm": 2.8125, + "learning_rate": 1.7358571428571426e-06, + "loss": 0.8394, + "step": 1850 + }, + { + "epoch": 0.26965314776557575, + "grad_norm": 2.375, + "learning_rate": 1.7344285714285714e-06, + "loss": 0.7879, + "step": 1860 + }, + { + "epoch": 0.2711028958718423, + "grad_norm": 2.625, + "learning_rate": 1.733e-06, + "loss": 0.8682, + "step": 1870 + }, + { + "epoch": 0.2725526439781088, + "grad_norm": 2.59375, + "learning_rate": 1.7315714285714285e-06, + "loss": 0.8546, + "step": 1880 + }, + { + "epoch": 0.2740023920843753, + "grad_norm": 2.5625, + "learning_rate": 1.730142857142857e-06, + "loss": 0.8424, + "step": 1890 + }, + { + "epoch": 0.2754521401906419, + "grad_norm": 2.453125, + "learning_rate": 1.7287142857142856e-06, + "loss": 0.8424, + "step": 1900 + }, + { + "epoch": 0.2769018882969084, + "grad_norm": 2.5, + "learning_rate": 1.7272857142857142e-06, + "loss": 0.8408, + "step": 1910 + }, + { + "epoch": 0.27835163640317495, + "grad_norm": 2.265625, + "learning_rate": 1.7258571428571427e-06, + "loss": 0.7772, + "step": 1920 + }, + { + "epoch": 0.2798013845094415, + "grad_norm": 2.609375, + "learning_rate": 1.7244285714285713e-06, + "loss": 0.7999, + "step": 1930 + }, + { + "epoch": 0.281251132615708, + "grad_norm": 2.578125, + "learning_rate": 1.723e-06, + "loss": 0.7943, + "step": 1940 + }, + { + "epoch": 0.2827008807219746, + "grad_norm": 2.5, + "learning_rate": 1.7215714285714284e-06, + "loss": 0.8283, + "step": 1950 + }, + { + "epoch": 0.2841506288282411, + "grad_norm": 2.6875, + "learning_rate": 1.7201428571428572e-06, + "loss": 0.8017, + "step": 1960 + }, + { + "epoch": 0.2856003769345076, + "grad_norm": 2.671875, + "learning_rate": 1.7187142857142855e-06, + "loss": 0.8182, + "step": 1970 + }, + { + "epoch": 0.28705012504077415, + "grad_norm": 2.203125, + "learning_rate": 1.7172857142857143e-06, + "loss": 0.802, + "step": 1980 + }, + { + "epoch": 0.2884998731470407, + "grad_norm": 2.421875, + "learning_rate": 1.7158571428571426e-06, + "loss": 0.7991, + "step": 1990 + }, + { + "epoch": 0.28994962125330725, + "grad_norm": 2.84375, + "learning_rate": 1.7144285714285714e-06, + "loss": 0.812, + "step": 2000 + }, + { + "epoch": 0.2913993693595738, + "grad_norm": 2.5, + "learning_rate": 1.713e-06, + "loss": 0.8085, + "step": 2010 + }, + { + "epoch": 0.2928491174658403, + "grad_norm": 2.65625, + "learning_rate": 1.7115714285714285e-06, + "loss": 0.9225, + "step": 2020 + }, + { + "epoch": 0.2942988655721068, + "grad_norm": 2.546875, + "learning_rate": 1.710142857142857e-06, + "loss": 0.8821, + "step": 2030 + }, + { + "epoch": 0.2957486136783734, + "grad_norm": 2.203125, + "learning_rate": 1.7087142857142856e-06, + "loss": 0.8367, + "step": 2040 + }, + { + "epoch": 0.29719836178463993, + "grad_norm": 2.671875, + "learning_rate": 1.7072857142857142e-06, + "loss": 0.8263, + "step": 2050 + }, + { + "epoch": 0.29864810989090645, + "grad_norm": 2.28125, + "learning_rate": 1.705857142857143e-06, + "loss": 0.8453, + "step": 2060 + }, + { + "epoch": 0.300097857997173, + "grad_norm": 2.375, + "learning_rate": 1.7044285714285713e-06, + "loss": 0.7457, + "step": 2070 + }, + { + "epoch": 0.3015476061034395, + "grad_norm": 3.09375, + "learning_rate": 1.703e-06, + "loss": 0.8192, + "step": 2080 + }, + { + "epoch": 0.3029973542097061, + "grad_norm": 2.625, + "learning_rate": 1.7015714285714284e-06, + "loss": 0.808, + "step": 2090 + }, + { + "epoch": 0.3044471023159726, + "grad_norm": 2.21875, + "learning_rate": 1.7001428571428572e-06, + "loss": 0.8019, + "step": 2100 + }, + { + "epoch": 0.30589685042223913, + "grad_norm": 2.703125, + "learning_rate": 1.6987142857142855e-06, + "loss": 0.8216, + "step": 2110 + }, + { + "epoch": 0.30734659852850565, + "grad_norm": 2.515625, + "learning_rate": 1.6972857142857143e-06, + "loss": 0.8336, + "step": 2120 + }, + { + "epoch": 0.30879634663477223, + "grad_norm": 2.5, + "learning_rate": 1.6958571428571428e-06, + "loss": 0.8398, + "step": 2130 + }, + { + "epoch": 0.31024609474103876, + "grad_norm": 2.328125, + "learning_rate": 1.6944285714285714e-06, + "loss": 0.7997, + "step": 2140 + }, + { + "epoch": 0.3116958428473053, + "grad_norm": 2.703125, + "learning_rate": 1.693e-06, + "loss": 0.7877, + "step": 2150 + }, + { + "epoch": 0.3131455909535718, + "grad_norm": 2.625, + "learning_rate": 1.6915714285714285e-06, + "loss": 0.8751, + "step": 2160 + }, + { + "epoch": 0.31459533905983833, + "grad_norm": 2.859375, + "learning_rate": 1.690142857142857e-06, + "loss": 0.8713, + "step": 2170 + }, + { + "epoch": 0.3160450871661049, + "grad_norm": 3.203125, + "learning_rate": 1.6887142857142856e-06, + "loss": 0.8164, + "step": 2180 + }, + { + "epoch": 0.31749483527237143, + "grad_norm": 2.796875, + "learning_rate": 1.6872857142857141e-06, + "loss": 0.7583, + "step": 2190 + }, + { + "epoch": 0.31894458337863796, + "grad_norm": 2.421875, + "learning_rate": 1.685857142857143e-06, + "loss": 0.7973, + "step": 2200 + }, + { + "epoch": 0.3203943314849045, + "grad_norm": 2.8125, + "learning_rate": 1.6844285714285713e-06, + "loss": 0.8176, + "step": 2210 + }, + { + "epoch": 0.32184407959117106, + "grad_norm": 2.5625, + "learning_rate": 1.683e-06, + "loss": 0.8629, + "step": 2220 + }, + { + "epoch": 0.3232938276974376, + "grad_norm": 2.875, + "learning_rate": 1.6815714285714284e-06, + "loss": 0.7541, + "step": 2230 + }, + { + "epoch": 0.3247435758037041, + "grad_norm": 2.53125, + "learning_rate": 1.6801428571428571e-06, + "loss": 0.7695, + "step": 2240 + }, + { + "epoch": 0.32619332390997063, + "grad_norm": 2.890625, + "learning_rate": 1.6787142857142855e-06, + "loss": 0.8176, + "step": 2250 + }, + { + "epoch": 0.32764307201623716, + "grad_norm": 2.453125, + "learning_rate": 1.6772857142857142e-06, + "loss": 0.7396, + "step": 2260 + }, + { + "epoch": 0.32909282012250374, + "grad_norm": 2.625, + "learning_rate": 1.6758571428571428e-06, + "loss": 0.8014, + "step": 2270 + }, + { + "epoch": 0.33054256822877026, + "grad_norm": 2.90625, + "learning_rate": 1.6744285714285714e-06, + "loss": 0.7933, + "step": 2280 + }, + { + "epoch": 0.3319923163350368, + "grad_norm": 2.859375, + "learning_rate": 1.673e-06, + "loss": 0.8283, + "step": 2290 + }, + { + "epoch": 0.3334420644413033, + "grad_norm": 2.625, + "learning_rate": 1.6715714285714285e-06, + "loss": 0.8429, + "step": 2300 + }, + { + "epoch": 0.33489181254756983, + "grad_norm": 2.390625, + "learning_rate": 1.670142857142857e-06, + "loss": 0.8863, + "step": 2310 + }, + { + "epoch": 0.3363415606538364, + "grad_norm": 2.46875, + "learning_rate": 1.6687142857142858e-06, + "loss": 0.8611, + "step": 2320 + }, + { + "epoch": 0.33779130876010294, + "grad_norm": 2.21875, + "learning_rate": 1.6672857142857141e-06, + "loss": 0.8482, + "step": 2330 + }, + { + "epoch": 0.33924105686636946, + "grad_norm": 2.546875, + "learning_rate": 1.6658571428571429e-06, + "loss": 0.7916, + "step": 2340 + }, + { + "epoch": 0.340690804972636, + "grad_norm": 2.265625, + "learning_rate": 1.6644285714285712e-06, + "loss": 0.8432, + "step": 2350 + }, + { + "epoch": 0.34214055307890257, + "grad_norm": 3.1875, + "learning_rate": 1.663e-06, + "loss": 0.8319, + "step": 2360 + }, + { + "epoch": 0.3435903011851691, + "grad_norm": 2.65625, + "learning_rate": 1.6615714285714283e-06, + "loss": 0.7659, + "step": 2370 + }, + { + "epoch": 0.3450400492914356, + "grad_norm": 2.828125, + "learning_rate": 1.6601428571428571e-06, + "loss": 0.809, + "step": 2380 + }, + { + "epoch": 0.34648979739770214, + "grad_norm": 2.59375, + "learning_rate": 1.6587142857142857e-06, + "loss": 0.8719, + "step": 2390 + }, + { + "epoch": 0.34793954550396866, + "grad_norm": 2.5625, + "learning_rate": 1.6572857142857142e-06, + "loss": 0.8141, + "step": 2400 + }, + { + "epoch": 0.34938929361023524, + "grad_norm": 2.40625, + "learning_rate": 1.6558571428571428e-06, + "loss": 0.8101, + "step": 2410 + }, + { + "epoch": 0.35083904171650176, + "grad_norm": 2.921875, + "learning_rate": 1.6544285714285713e-06, + "loss": 0.7923, + "step": 2420 + }, + { + "epoch": 0.3522887898227683, + "grad_norm": 2.890625, + "learning_rate": 1.6529999999999999e-06, + "loss": 0.8276, + "step": 2430 + }, + { + "epoch": 0.3537385379290348, + "grad_norm": 2.40625, + "learning_rate": 1.6515714285714284e-06, + "loss": 0.7846, + "step": 2440 + }, + { + "epoch": 0.3551882860353014, + "grad_norm": 2.546875, + "learning_rate": 1.650142857142857e-06, + "loss": 0.8342, + "step": 2450 + }, + { + "epoch": 0.3566380341415679, + "grad_norm": 2.625, + "learning_rate": 1.6487142857142858e-06, + "loss": 0.8012, + "step": 2460 + }, + { + "epoch": 0.35808778224783444, + "grad_norm": 2.4375, + "learning_rate": 1.647285714285714e-06, + "loss": 0.8666, + "step": 2470 + }, + { + "epoch": 0.35953753035410096, + "grad_norm": 2.875, + "learning_rate": 1.6458571428571429e-06, + "loss": 0.8359, + "step": 2480 + }, + { + "epoch": 0.3609872784603675, + "grad_norm": 2.71875, + "learning_rate": 1.6444285714285712e-06, + "loss": 0.7718, + "step": 2490 + }, + { + "epoch": 0.36243702656663407, + "grad_norm": 2.375, + "learning_rate": 1.643e-06, + "loss": 0.8129, + "step": 2500 + }, + { + "epoch": 0.3638867746729006, + "grad_norm": 2.1875, + "learning_rate": 1.6415714285714285e-06, + "loss": 0.796, + "step": 2510 + }, + { + "epoch": 0.3653365227791671, + "grad_norm": 3.46875, + "learning_rate": 1.640142857142857e-06, + "loss": 0.8485, + "step": 2520 + }, + { + "epoch": 0.36678627088543364, + "grad_norm": 2.46875, + "learning_rate": 1.6387142857142856e-06, + "loss": 0.8605, + "step": 2530 + }, + { + "epoch": 0.36823601899170016, + "grad_norm": 2.953125, + "learning_rate": 1.6372857142857142e-06, + "loss": 0.7999, + "step": 2540 + }, + { + "epoch": 0.36968576709796674, + "grad_norm": 3.4375, + "learning_rate": 1.6358571428571428e-06, + "loss": 0.8774, + "step": 2550 + }, + { + "epoch": 0.37113551520423327, + "grad_norm": 2.3125, + "learning_rate": 1.6344285714285713e-06, + "loss": 0.7905, + "step": 2560 + }, + { + "epoch": 0.3725852633104998, + "grad_norm": 2.484375, + "learning_rate": 1.6329999999999999e-06, + "loss": 0.8083, + "step": 2570 + }, + { + "epoch": 0.3740350114167663, + "grad_norm": 2.859375, + "learning_rate": 1.6315714285714286e-06, + "loss": 0.7904, + "step": 2580 + }, + { + "epoch": 0.3754847595230329, + "grad_norm": 2.640625, + "learning_rate": 1.630142857142857e-06, + "loss": 0.8099, + "step": 2590 + }, + { + "epoch": 0.3769345076292994, + "grad_norm": 2.5625, + "learning_rate": 1.6287142857142857e-06, + "loss": 0.8028, + "step": 2600 + }, + { + "epoch": 0.37838425573556594, + "grad_norm": 2.796875, + "learning_rate": 1.627285714285714e-06, + "loss": 0.817, + "step": 2610 + }, + { + "epoch": 0.37983400384183247, + "grad_norm": 2.171875, + "learning_rate": 1.6258571428571429e-06, + "loss": 0.8369, + "step": 2620 + }, + { + "epoch": 0.381283751948099, + "grad_norm": 2.421875, + "learning_rate": 1.6244285714285712e-06, + "loss": 0.7882, + "step": 2630 + }, + { + "epoch": 0.38273350005436557, + "grad_norm": 2.78125, + "learning_rate": 1.623e-06, + "loss": 0.7568, + "step": 2640 + }, + { + "epoch": 0.3841832481606321, + "grad_norm": 2.640625, + "learning_rate": 1.6215714285714285e-06, + "loss": 0.8551, + "step": 2650 + }, + { + "epoch": 0.3856329962668986, + "grad_norm": 2.453125, + "learning_rate": 1.620142857142857e-06, + "loss": 0.7759, + "step": 2660 + }, + { + "epoch": 0.38708274437316514, + "grad_norm": 2.9375, + "learning_rate": 1.6187142857142856e-06, + "loss": 0.8493, + "step": 2670 + }, + { + "epoch": 0.3885324924794317, + "grad_norm": 3.203125, + "learning_rate": 1.6172857142857142e-06, + "loss": 0.8246, + "step": 2680 + }, + { + "epoch": 0.38998224058569825, + "grad_norm": 2.59375, + "learning_rate": 1.6158571428571427e-06, + "loss": 0.7912, + "step": 2690 + }, + { + "epoch": 0.39143198869196477, + "grad_norm": 2.671875, + "learning_rate": 1.6144285714285713e-06, + "loss": 0.7963, + "step": 2700 + }, + { + "epoch": 0.3928817367982313, + "grad_norm": 2.796875, + "learning_rate": 1.6129999999999998e-06, + "loss": 0.8349, + "step": 2710 + }, + { + "epoch": 0.3943314849044978, + "grad_norm": 2.484375, + "learning_rate": 1.6115714285714286e-06, + "loss": 0.8791, + "step": 2720 + }, + { + "epoch": 0.3957812330107644, + "grad_norm": 2.53125, + "learning_rate": 1.610142857142857e-06, + "loss": 0.8467, + "step": 2730 + }, + { + "epoch": 0.3972309811170309, + "grad_norm": 2.625, + "learning_rate": 1.6087142857142857e-06, + "loss": 0.8632, + "step": 2740 + }, + { + "epoch": 0.39868072922329745, + "grad_norm": 2.578125, + "learning_rate": 1.607285714285714e-06, + "loss": 0.8356, + "step": 2750 + }, + { + "epoch": 0.40013047732956397, + "grad_norm": 2.078125, + "learning_rate": 1.6058571428571428e-06, + "loss": 0.7998, + "step": 2760 + }, + { + "epoch": 0.40158022543583055, + "grad_norm": 2.140625, + "learning_rate": 1.6044285714285714e-06, + "loss": 0.823, + "step": 2770 + }, + { + "epoch": 0.4030299735420971, + "grad_norm": 3.234375, + "learning_rate": 1.603e-06, + "loss": 0.8842, + "step": 2780 + }, + { + "epoch": 0.4044797216483636, + "grad_norm": 2.421875, + "learning_rate": 1.6015714285714285e-06, + "loss": 0.8275, + "step": 2790 + }, + { + "epoch": 0.4059294697546301, + "grad_norm": 2.703125, + "learning_rate": 1.600142857142857e-06, + "loss": 0.8138, + "step": 2800 + }, + { + "epoch": 0.40737921786089665, + "grad_norm": 2.421875, + "learning_rate": 1.5987142857142856e-06, + "loss": 0.7782, + "step": 2810 + }, + { + "epoch": 0.4088289659671632, + "grad_norm": 2.296875, + "learning_rate": 1.5972857142857142e-06, + "loss": 0.8185, + "step": 2820 + }, + { + "epoch": 0.41027871407342975, + "grad_norm": 2.75, + "learning_rate": 1.5958571428571427e-06, + "loss": 0.8008, + "step": 2830 + }, + { + "epoch": 0.4117284621796963, + "grad_norm": 2.21875, + "learning_rate": 1.5944285714285715e-06, + "loss": 0.9156, + "step": 2840 + }, + { + "epoch": 0.4131782102859628, + "grad_norm": 2.25, + "learning_rate": 1.5929999999999998e-06, + "loss": 0.8302, + "step": 2850 + }, + { + "epoch": 0.4146279583922293, + "grad_norm": 2.71875, + "learning_rate": 1.5915714285714286e-06, + "loss": 0.7983, + "step": 2860 + }, + { + "epoch": 0.4160777064984959, + "grad_norm": 2.578125, + "learning_rate": 1.590142857142857e-06, + "loss": 0.842, + "step": 2870 + }, + { + "epoch": 0.4175274546047624, + "grad_norm": 2.828125, + "learning_rate": 1.5887142857142857e-06, + "loss": 0.8377, + "step": 2880 + }, + { + "epoch": 0.41897720271102895, + "grad_norm": 2.421875, + "learning_rate": 1.587285714285714e-06, + "loss": 0.8462, + "step": 2890 + }, + { + "epoch": 0.4204269508172955, + "grad_norm": 2.65625, + "learning_rate": 1.5858571428571428e-06, + "loss": 0.8026, + "step": 2900 + }, + { + "epoch": 0.42187669892356205, + "grad_norm": 2.765625, + "learning_rate": 1.5844285714285714e-06, + "loss": 0.8224, + "step": 2910 + }, + { + "epoch": 0.4233264470298286, + "grad_norm": 2.515625, + "learning_rate": 1.583e-06, + "loss": 0.8332, + "step": 2920 + }, + { + "epoch": 0.4247761951360951, + "grad_norm": 2.84375, + "learning_rate": 1.5815714285714285e-06, + "loss": 0.8036, + "step": 2930 + }, + { + "epoch": 0.4262259432423616, + "grad_norm": 2.171875, + "learning_rate": 1.580142857142857e-06, + "loss": 0.9069, + "step": 2940 + }, + { + "epoch": 0.42767569134862815, + "grad_norm": 2.671875, + "learning_rate": 1.5787142857142856e-06, + "loss": 0.8566, + "step": 2950 + }, + { + "epoch": 0.42912543945489473, + "grad_norm": 2.1875, + "learning_rate": 1.5772857142857141e-06, + "loss": 0.8218, + "step": 2960 + }, + { + "epoch": 0.43057518756116125, + "grad_norm": 2.28125, + "learning_rate": 1.5758571428571427e-06, + "loss": 0.838, + "step": 2970 + }, + { + "epoch": 0.4320249356674278, + "grad_norm": 2.078125, + "learning_rate": 1.5744285714285715e-06, + "loss": 0.8333, + "step": 2980 + }, + { + "epoch": 0.4334746837736943, + "grad_norm": 2.40625, + "learning_rate": 1.5729999999999998e-06, + "loss": 0.743, + "step": 2990 + }, + { + "epoch": 0.4349244318799609, + "grad_norm": 2.578125, + "learning_rate": 1.5715714285714286e-06, + "loss": 0.8014, + "step": 3000 + }, + { + "epoch": 0.4363741799862274, + "grad_norm": 2.6875, + "learning_rate": 1.570142857142857e-06, + "loss": 0.8201, + "step": 3010 + }, + { + "epoch": 0.43782392809249393, + "grad_norm": 2.6875, + "learning_rate": 1.5687142857142857e-06, + "loss": 0.7979, + "step": 3020 + }, + { + "epoch": 0.43927367619876045, + "grad_norm": 3.015625, + "learning_rate": 1.5672857142857142e-06, + "loss": 0.7914, + "step": 3030 + }, + { + "epoch": 0.440723424305027, + "grad_norm": 2.4375, + "learning_rate": 1.5658571428571428e-06, + "loss": 0.8406, + "step": 3040 + }, + { + "epoch": 0.44217317241129356, + "grad_norm": 2.671875, + "learning_rate": 1.5644285714285713e-06, + "loss": 0.7772, + "step": 3050 + }, + { + "epoch": 0.4436229205175601, + "grad_norm": 2.8125, + "learning_rate": 1.563e-06, + "loss": 0.8082, + "step": 3060 + }, + { + "epoch": 0.4450726686238266, + "grad_norm": 2.859375, + "learning_rate": 1.5615714285714285e-06, + "loss": 0.8437, + "step": 3070 + }, + { + "epoch": 0.44652241673009313, + "grad_norm": 2.8125, + "learning_rate": 1.560142857142857e-06, + "loss": 0.8012, + "step": 3080 + }, + { + "epoch": 0.4479721648363597, + "grad_norm": 3.109375, + "learning_rate": 1.5587142857142856e-06, + "loss": 0.8487, + "step": 3090 + }, + { + "epoch": 0.44942191294262623, + "grad_norm": 2.625, + "learning_rate": 1.5572857142857143e-06, + "loss": 0.8683, + "step": 3100 + }, + { + "epoch": 0.45087166104889276, + "grad_norm": 3.09375, + "learning_rate": 1.5558571428571427e-06, + "loss": 0.7946, + "step": 3110 + }, + { + "epoch": 0.4523214091551593, + "grad_norm": 8.0625, + "learning_rate": 1.5544285714285714e-06, + "loss": 0.8102, + "step": 3120 + }, + { + "epoch": 0.4537711572614258, + "grad_norm": 2.609375, + "learning_rate": 1.5529999999999998e-06, + "loss": 0.823, + "step": 3130 + }, + { + "epoch": 0.4552209053676924, + "grad_norm": 2.453125, + "learning_rate": 1.5515714285714286e-06, + "loss": 0.838, + "step": 3140 + }, + { + "epoch": 0.4566706534739589, + "grad_norm": 3.109375, + "learning_rate": 1.550142857142857e-06, + "loss": 0.8373, + "step": 3150 + }, + { + "epoch": 0.45812040158022543, + "grad_norm": 2.515625, + "learning_rate": 1.5487142857142857e-06, + "loss": 0.7699, + "step": 3160 + }, + { + "epoch": 0.45957014968649196, + "grad_norm": 2.578125, + "learning_rate": 1.5472857142857142e-06, + "loss": 0.8584, + "step": 3170 + }, + { + "epoch": 0.4610198977927585, + "grad_norm": 2.625, + "learning_rate": 1.5458571428571428e-06, + "loss": 0.8133, + "step": 3180 + }, + { + "epoch": 0.46246964589902506, + "grad_norm": 2.609375, + "learning_rate": 1.5444285714285713e-06, + "loss": 0.8767, + "step": 3190 + }, + { + "epoch": 0.4639193940052916, + "grad_norm": 2.53125, + "learning_rate": 1.5429999999999999e-06, + "loss": 0.8222, + "step": 3200 + }, + { + "epoch": 0.4653691421115581, + "grad_norm": 2.5, + "learning_rate": 1.5415714285714284e-06, + "loss": 0.897, + "step": 3210 + }, + { + "epoch": 0.46681889021782463, + "grad_norm": 2.71875, + "learning_rate": 1.540142857142857e-06, + "loss": 0.8424, + "step": 3220 + }, + { + "epoch": 0.4682686383240912, + "grad_norm": 2.53125, + "learning_rate": 1.5387142857142855e-06, + "loss": 0.812, + "step": 3230 + }, + { + "epoch": 0.46971838643035774, + "grad_norm": 2.828125, + "learning_rate": 1.5372857142857143e-06, + "loss": 0.7614, + "step": 3240 + }, + { + "epoch": 0.47116813453662426, + "grad_norm": 3.171875, + "learning_rate": 1.5358571428571427e-06, + "loss": 0.7944, + "step": 3250 + }, + { + "epoch": 0.4726178826428908, + "grad_norm": 2.4375, + "learning_rate": 1.5344285714285714e-06, + "loss": 0.8452, + "step": 3260 + }, + { + "epoch": 0.4740676307491573, + "grad_norm": 2.78125, + "learning_rate": 1.5329999999999998e-06, + "loss": 0.8174, + "step": 3270 + }, + { + "epoch": 0.4755173788554239, + "grad_norm": 2.953125, + "learning_rate": 1.5315714285714285e-06, + "loss": 0.8349, + "step": 3280 + }, + { + "epoch": 0.4769671269616904, + "grad_norm": 2.765625, + "learning_rate": 1.530142857142857e-06, + "loss": 0.8118, + "step": 3290 + }, + { + "epoch": 0.47841687506795694, + "grad_norm": 2.875, + "learning_rate": 1.5287142857142856e-06, + "loss": 0.8324, + "step": 3300 + }, + { + "epoch": 0.47986662317422346, + "grad_norm": 2.78125, + "learning_rate": 1.5272857142857142e-06, + "loss": 0.8308, + "step": 3310 + }, + { + "epoch": 0.48131637128049004, + "grad_norm": 3.0625, + "learning_rate": 1.5258571428571428e-06, + "loss": 0.875, + "step": 3320 + }, + { + "epoch": 0.48276611938675656, + "grad_norm": 2.84375, + "learning_rate": 1.5244285714285713e-06, + "loss": 0.829, + "step": 3330 + }, + { + "epoch": 0.4842158674930231, + "grad_norm": 2.375, + "learning_rate": 1.5229999999999999e-06, + "loss": 0.7742, + "step": 3340 + }, + { + "epoch": 0.4856656155992896, + "grad_norm": 2.890625, + "learning_rate": 1.5215714285714284e-06, + "loss": 0.8065, + "step": 3350 + }, + { + "epoch": 0.48711536370555614, + "grad_norm": 2.34375, + "learning_rate": 1.5201428571428572e-06, + "loss": 0.829, + "step": 3360 + }, + { + "epoch": 0.4885651118118227, + "grad_norm": 2.578125, + "learning_rate": 1.5187142857142855e-06, + "loss": 0.837, + "step": 3370 + }, + { + "epoch": 0.49001485991808924, + "grad_norm": 2.28125, + "learning_rate": 1.5172857142857143e-06, + "loss": 0.8143, + "step": 3380 + }, + { + "epoch": 0.49146460802435576, + "grad_norm": 2.8125, + "learning_rate": 1.5158571428571426e-06, + "loss": 0.8379, + "step": 3390 + }, + { + "epoch": 0.4929143561306223, + "grad_norm": 2.3125, + "learning_rate": 1.5144285714285714e-06, + "loss": 0.8351, + "step": 3400 + }, + { + "epoch": 0.49436410423688887, + "grad_norm": 2.59375, + "learning_rate": 1.5129999999999997e-06, + "loss": 0.8256, + "step": 3410 + }, + { + "epoch": 0.4958138523431554, + "grad_norm": 2.296875, + "learning_rate": 1.5115714285714285e-06, + "loss": 0.7167, + "step": 3420 + }, + { + "epoch": 0.4972636004494219, + "grad_norm": 2.390625, + "learning_rate": 1.510142857142857e-06, + "loss": 0.8, + "step": 3430 + }, + { + "epoch": 0.49871334855568844, + "grad_norm": 2.65625, + "learning_rate": 1.5087142857142856e-06, + "loss": 0.8157, + "step": 3440 + }, + { + "epoch": 0.500163096661955, + "grad_norm": 3.265625, + "learning_rate": 1.5072857142857142e-06, + "loss": 0.8145, + "step": 3450 + }, + { + "epoch": 0.5016128447682215, + "grad_norm": 2.453125, + "learning_rate": 1.5058571428571427e-06, + "loss": 0.7918, + "step": 3460 + }, + { + "epoch": 0.5030625928744881, + "grad_norm": 2.703125, + "learning_rate": 1.5044285714285713e-06, + "loss": 0.8213, + "step": 3470 + }, + { + "epoch": 0.5045123409807546, + "grad_norm": 2.5625, + "learning_rate": 1.5029999999999998e-06, + "loss": 0.8377, + "step": 3480 + }, + { + "epoch": 0.5059620890870211, + "grad_norm": 2.78125, + "learning_rate": 1.5015714285714284e-06, + "loss": 0.7633, + "step": 3490 + }, + { + "epoch": 0.5074118371932876, + "grad_norm": 2.296875, + "learning_rate": 1.5001428571428572e-06, + "loss": 0.8008, + "step": 3500 + }, + { + "epoch": 0.5088615852995542, + "grad_norm": 3.15625, + "learning_rate": 1.4987142857142855e-06, + "loss": 0.8114, + "step": 3510 + }, + { + "epoch": 0.5103113334058207, + "grad_norm": 2.359375, + "learning_rate": 1.4972857142857143e-06, + "loss": 0.8536, + "step": 3520 + }, + { + "epoch": 0.5117610815120873, + "grad_norm": 3.140625, + "learning_rate": 1.4958571428571426e-06, + "loss": 0.8447, + "step": 3530 + }, + { + "epoch": 0.5132108296183538, + "grad_norm": 2.421875, + "learning_rate": 1.4944285714285714e-06, + "loss": 0.8397, + "step": 3540 + }, + { + "epoch": 0.5146605777246204, + "grad_norm": 3.0, + "learning_rate": 1.493e-06, + "loss": 0.8676, + "step": 3550 + }, + { + "epoch": 0.5161103258308869, + "grad_norm": 2.5, + "learning_rate": 1.4915714285714285e-06, + "loss": 0.8115, + "step": 3560 + }, + { + "epoch": 0.5175600739371534, + "grad_norm": 2.25, + "learning_rate": 1.490142857142857e-06, + "loss": 0.7876, + "step": 3570 + }, + { + "epoch": 0.5190098220434199, + "grad_norm": 2.703125, + "learning_rate": 1.4887142857142856e-06, + "loss": 0.8338, + "step": 3580 + }, + { + "epoch": 0.5204595701496865, + "grad_norm": 2.265625, + "learning_rate": 1.4872857142857142e-06, + "loss": 0.7992, + "step": 3590 + }, + { + "epoch": 0.521909318255953, + "grad_norm": 2.34375, + "learning_rate": 1.4858571428571427e-06, + "loss": 0.8447, + "step": 3600 + }, + { + "epoch": 0.5233590663622195, + "grad_norm": 2.515625, + "learning_rate": 1.4844285714285713e-06, + "loss": 0.8187, + "step": 3610 + }, + { + "epoch": 0.5248088144684862, + "grad_norm": 2.828125, + "learning_rate": 1.483e-06, + "loss": 0.8401, + "step": 3620 + }, + { + "epoch": 0.5262585625747527, + "grad_norm": 2.90625, + "learning_rate": 1.4815714285714284e-06, + "loss": 0.7889, + "step": 3630 + }, + { + "epoch": 0.5277083106810192, + "grad_norm": 2.15625, + "learning_rate": 1.4801428571428571e-06, + "loss": 0.8028, + "step": 3640 + }, + { + "epoch": 0.5291580587872857, + "grad_norm": 2.734375, + "learning_rate": 1.4787142857142855e-06, + "loss": 0.8515, + "step": 3650 + }, + { + "epoch": 0.5306078068935522, + "grad_norm": 2.328125, + "learning_rate": 1.4772857142857143e-06, + "loss": 0.8201, + "step": 3660 + }, + { + "epoch": 0.5320575549998188, + "grad_norm": 2.640625, + "learning_rate": 1.4758571428571426e-06, + "loss": 0.8526, + "step": 3670 + }, + { + "epoch": 0.5335073031060853, + "grad_norm": 2.453125, + "learning_rate": 1.4744285714285714e-06, + "loss": 0.8708, + "step": 3680 + }, + { + "epoch": 0.5349570512123518, + "grad_norm": 2.578125, + "learning_rate": 1.473e-06, + "loss": 0.8377, + "step": 3690 + }, + { + "epoch": 0.5364067993186183, + "grad_norm": 2.5, + "learning_rate": 1.4715714285714285e-06, + "loss": 0.8366, + "step": 3700 + }, + { + "epoch": 0.537856547424885, + "grad_norm": 2.671875, + "learning_rate": 1.470142857142857e-06, + "loss": 0.7561, + "step": 3710 + }, + { + "epoch": 0.5393062955311515, + "grad_norm": 2.921875, + "learning_rate": 1.4687142857142856e-06, + "loss": 0.8734, + "step": 3720 + }, + { + "epoch": 0.540756043637418, + "grad_norm": 3.125, + "learning_rate": 1.4672857142857141e-06, + "loss": 0.8377, + "step": 3730 + }, + { + "epoch": 0.5422057917436846, + "grad_norm": 3.125, + "learning_rate": 1.4658571428571427e-06, + "loss": 0.8081, + "step": 3740 + }, + { + "epoch": 0.5436555398499511, + "grad_norm": 2.4375, + "learning_rate": 1.4644285714285712e-06, + "loss": 0.7883, + "step": 3750 + }, + { + "epoch": 0.5451052879562176, + "grad_norm": 12.4375, + "learning_rate": 1.463e-06, + "loss": 0.8154, + "step": 3760 + }, + { + "epoch": 0.5465550360624841, + "grad_norm": 3.390625, + "learning_rate": 1.4615714285714284e-06, + "loss": 0.7484, + "step": 3770 + }, + { + "epoch": 0.5480047841687506, + "grad_norm": 2.1875, + "learning_rate": 1.4601428571428571e-06, + "loss": 0.8581, + "step": 3780 + }, + { + "epoch": 0.5494545322750172, + "grad_norm": 2.5, + "learning_rate": 1.4587142857142855e-06, + "loss": 0.8005, + "step": 3790 + }, + { + "epoch": 0.5509042803812838, + "grad_norm": 2.6875, + "learning_rate": 1.4572857142857142e-06, + "loss": 0.7916, + "step": 3800 + }, + { + "epoch": 0.5523540284875503, + "grad_norm": 2.765625, + "learning_rate": 1.455857142857143e-06, + "loss": 0.8062, + "step": 3810 + }, + { + "epoch": 0.5538037765938169, + "grad_norm": 2.296875, + "learning_rate": 1.4544285714285713e-06, + "loss": 0.77, + "step": 3820 + }, + { + "epoch": 0.5552535247000834, + "grad_norm": 2.96875, + "learning_rate": 1.4530000000000001e-06, + "loss": 0.7917, + "step": 3830 + }, + { + "epoch": 0.5567032728063499, + "grad_norm": 2.78125, + "learning_rate": 1.4515714285714284e-06, + "loss": 0.8262, + "step": 3840 + }, + { + "epoch": 0.5581530209126164, + "grad_norm": 2.5625, + "learning_rate": 1.4501428571428572e-06, + "loss": 0.7932, + "step": 3850 + }, + { + "epoch": 0.559602769018883, + "grad_norm": 2.5625, + "learning_rate": 1.4487142857142856e-06, + "loss": 0.7886, + "step": 3860 + }, + { + "epoch": 0.5610525171251495, + "grad_norm": 2.3125, + "learning_rate": 1.4472857142857143e-06, + "loss": 0.8574, + "step": 3870 + }, + { + "epoch": 0.562502265231416, + "grad_norm": 3.015625, + "learning_rate": 1.4458571428571429e-06, + "loss": 0.8001, + "step": 3880 + }, + { + "epoch": 0.5639520133376825, + "grad_norm": 2.546875, + "learning_rate": 1.4444285714285712e-06, + "loss": 0.8008, + "step": 3890 + }, + { + "epoch": 0.5654017614439492, + "grad_norm": 2.59375, + "learning_rate": 1.443e-06, + "loss": 0.8233, + "step": 3900 + }, + { + "epoch": 0.5668515095502157, + "grad_norm": 2.65625, + "learning_rate": 1.4415714285714283e-06, + "loss": 0.7831, + "step": 3910 + }, + { + "epoch": 0.5683012576564822, + "grad_norm": 2.53125, + "learning_rate": 1.440142857142857e-06, + "loss": 0.8156, + "step": 3920 + }, + { + "epoch": 0.5697510057627487, + "grad_norm": 2.59375, + "learning_rate": 1.4387142857142854e-06, + "loss": 0.8035, + "step": 3930 + }, + { + "epoch": 0.5712007538690153, + "grad_norm": 2.671875, + "learning_rate": 1.4372857142857142e-06, + "loss": 0.8253, + "step": 3940 + }, + { + "epoch": 0.5726505019752818, + "grad_norm": 2.90625, + "learning_rate": 1.435857142857143e-06, + "loss": 0.8293, + "step": 3950 + }, + { + "epoch": 0.5741002500815483, + "grad_norm": 2.609375, + "learning_rate": 1.4344285714285713e-06, + "loss": 0.8229, + "step": 3960 + }, + { + "epoch": 0.5755499981878148, + "grad_norm": 2.453125, + "learning_rate": 1.433e-06, + "loss": 0.842, + "step": 3970 + }, + { + "epoch": 0.5769997462940814, + "grad_norm": 2.5, + "learning_rate": 1.4315714285714284e-06, + "loss": 0.815, + "step": 3980 + }, + { + "epoch": 0.578449494400348, + "grad_norm": 2.71875, + "learning_rate": 1.4301428571428572e-06, + "loss": 0.7907, + "step": 3990 + }, + { + "epoch": 0.5798992425066145, + "grad_norm": 3.15625, + "learning_rate": 1.4287142857142855e-06, + "loss": 0.8314, + "step": 4000 + }, + { + "epoch": 0.581348990612881, + "grad_norm": 2.546875, + "learning_rate": 1.4272857142857143e-06, + "loss": 0.8048, + "step": 4010 + }, + { + "epoch": 0.5827987387191476, + "grad_norm": 2.5625, + "learning_rate": 1.4258571428571429e-06, + "loss": 0.7836, + "step": 4020 + }, + { + "epoch": 0.5842484868254141, + "grad_norm": 2.4375, + "learning_rate": 1.4244285714285714e-06, + "loss": 0.8082, + "step": 4030 + }, + { + "epoch": 0.5856982349316806, + "grad_norm": 2.5625, + "learning_rate": 1.423e-06, + "loss": 0.806, + "step": 4040 + }, + { + "epoch": 0.5871479830379471, + "grad_norm": 2.8125, + "learning_rate": 1.4215714285714285e-06, + "loss": 0.883, + "step": 4050 + }, + { + "epoch": 0.5885977311442137, + "grad_norm": 2.546875, + "learning_rate": 1.420142857142857e-06, + "loss": 0.8406, + "step": 4060 + }, + { + "epoch": 0.5900474792504802, + "grad_norm": 2.484375, + "learning_rate": 1.4187142857142858e-06, + "loss": 0.7933, + "step": 4070 + }, + { + "epoch": 0.5914972273567468, + "grad_norm": 2.59375, + "learning_rate": 1.4172857142857142e-06, + "loss": 0.8305, + "step": 4080 + }, + { + "epoch": 0.5929469754630133, + "grad_norm": 2.828125, + "learning_rate": 1.415857142857143e-06, + "loss": 0.8244, + "step": 4090 + }, + { + "epoch": 0.5943967235692799, + "grad_norm": 2.8125, + "learning_rate": 1.4144285714285713e-06, + "loss": 0.8106, + "step": 4100 + }, + { + "epoch": 0.5958464716755464, + "grad_norm": 2.546875, + "learning_rate": 1.413e-06, + "loss": 0.79, + "step": 4110 + }, + { + "epoch": 0.5972962197818129, + "grad_norm": 2.734375, + "learning_rate": 1.4115714285714284e-06, + "loss": 0.8551, + "step": 4120 + }, + { + "epoch": 0.5987459678880794, + "grad_norm": 2.421875, + "learning_rate": 1.4101428571428572e-06, + "loss": 0.813, + "step": 4130 + }, + { + "epoch": 0.600195715994346, + "grad_norm": 2.765625, + "learning_rate": 1.4087142857142857e-06, + "loss": 0.8393, + "step": 4140 + }, + { + "epoch": 0.6016454641006125, + "grad_norm": 2.828125, + "learning_rate": 1.4072857142857143e-06, + "loss": 0.8205, + "step": 4150 + }, + { + "epoch": 0.603095212206879, + "grad_norm": 2.8125, + "learning_rate": 1.4058571428571428e-06, + "loss": 0.8258, + "step": 4160 + }, + { + "epoch": 0.6045449603131456, + "grad_norm": 2.859375, + "learning_rate": 1.4044285714285714e-06, + "loss": 0.8534, + "step": 4170 + }, + { + "epoch": 0.6059947084194122, + "grad_norm": 3.1875, + "learning_rate": 1.403e-06, + "loss": 0.8111, + "step": 4180 + }, + { + "epoch": 0.6074444565256787, + "grad_norm": 2.765625, + "learning_rate": 1.4015714285714285e-06, + "loss": 0.7963, + "step": 4190 + }, + { + "epoch": 0.6088942046319452, + "grad_norm": 2.671875, + "learning_rate": 1.400142857142857e-06, + "loss": 0.8372, + "step": 4200 + }, + { + "epoch": 0.6103439527382117, + "grad_norm": 2.6875, + "learning_rate": 1.3987142857142858e-06, + "loss": 0.8316, + "step": 4210 + }, + { + "epoch": 0.6117937008444783, + "grad_norm": 2.75, + "learning_rate": 1.3972857142857142e-06, + "loss": 0.802, + "step": 4220 + }, + { + "epoch": 0.6132434489507448, + "grad_norm": 2.78125, + "learning_rate": 1.395857142857143e-06, + "loss": 0.8376, + "step": 4230 + }, + { + "epoch": 0.6146931970570113, + "grad_norm": 2.453125, + "learning_rate": 1.3944285714285713e-06, + "loss": 0.8275, + "step": 4240 + }, + { + "epoch": 0.6161429451632778, + "grad_norm": 2.796875, + "learning_rate": 1.393e-06, + "loss": 0.8129, + "step": 4250 + }, + { + "epoch": 0.6175926932695445, + "grad_norm": 2.75, + "learning_rate": 1.3915714285714286e-06, + "loss": 0.8223, + "step": 4260 + }, + { + "epoch": 0.619042441375811, + "grad_norm": 2.734375, + "learning_rate": 1.3901428571428572e-06, + "loss": 0.7829, + "step": 4270 + }, + { + "epoch": 0.6204921894820775, + "grad_norm": 2.375, + "learning_rate": 1.3887142857142857e-06, + "loss": 0.8008, + "step": 4280 + }, + { + "epoch": 0.621941937588344, + "grad_norm": 2.640625, + "learning_rate": 1.3872857142857143e-06, + "loss": 0.7598, + "step": 4290 + }, + { + "epoch": 0.6233916856946106, + "grad_norm": 2.75, + "learning_rate": 1.3858571428571428e-06, + "loss": 0.8197, + "step": 4300 + }, + { + "epoch": 0.6248414338008771, + "grad_norm": 2.84375, + "learning_rate": 1.3844285714285714e-06, + "loss": 0.8299, + "step": 4310 + }, + { + "epoch": 0.6262911819071436, + "grad_norm": 2.46875, + "learning_rate": 1.383e-06, + "loss": 0.801, + "step": 4320 + }, + { + "epoch": 0.6277409300134101, + "grad_norm": 2.171875, + "learning_rate": 1.3815714285714287e-06, + "loss": 0.8014, + "step": 4330 + }, + { + "epoch": 0.6291906781196767, + "grad_norm": 2.546875, + "learning_rate": 1.380142857142857e-06, + "loss": 0.7462, + "step": 4340 + }, + { + "epoch": 0.6306404262259433, + "grad_norm": 2.65625, + "learning_rate": 1.3787142857142858e-06, + "loss": 0.7883, + "step": 4350 + }, + { + "epoch": 0.6320901743322098, + "grad_norm": 3.015625, + "learning_rate": 1.3772857142857141e-06, + "loss": 0.8084, + "step": 4360 + }, + { + "epoch": 0.6335399224384763, + "grad_norm": 2.859375, + "learning_rate": 1.375857142857143e-06, + "loss": 0.8145, + "step": 4370 + }, + { + "epoch": 0.6349896705447429, + "grad_norm": 2.609375, + "learning_rate": 1.3744285714285713e-06, + "loss": 0.8432, + "step": 4380 + }, + { + "epoch": 0.6364394186510094, + "grad_norm": 3.09375, + "learning_rate": 1.373e-06, + "loss": 0.8193, + "step": 4390 + }, + { + "epoch": 0.6378891667572759, + "grad_norm": 2.609375, + "learning_rate": 1.3715714285714286e-06, + "loss": 0.8262, + "step": 4400 + }, + { + "epoch": 0.6393389148635424, + "grad_norm": 2.734375, + "learning_rate": 1.3701428571428571e-06, + "loss": 0.8625, + "step": 4410 + }, + { + "epoch": 0.640788662969809, + "grad_norm": 2.984375, + "learning_rate": 1.3687142857142857e-06, + "loss": 0.8308, + "step": 4420 + }, + { + "epoch": 0.6422384110760755, + "grad_norm": 3.46875, + "learning_rate": 1.3672857142857142e-06, + "loss": 0.8108, + "step": 4430 + }, + { + "epoch": 0.6436881591823421, + "grad_norm": 2.71875, + "learning_rate": 1.3658571428571428e-06, + "loss": 0.8205, + "step": 4440 + }, + { + "epoch": 0.6451379072886086, + "grad_norm": 3.21875, + "learning_rate": 1.3644285714285714e-06, + "loss": 0.8054, + "step": 4450 + }, + { + "epoch": 0.6465876553948752, + "grad_norm": 2.359375, + "learning_rate": 1.363e-06, + "loss": 0.8891, + "step": 4460 + }, + { + "epoch": 0.6480374035011417, + "grad_norm": 2.578125, + "learning_rate": 1.3615714285714287e-06, + "loss": 0.7841, + "step": 4470 + }, + { + "epoch": 0.6494871516074082, + "grad_norm": 2.375, + "learning_rate": 1.360142857142857e-06, + "loss": 0.8368, + "step": 4480 + }, + { + "epoch": 0.6509368997136747, + "grad_norm": 2.859375, + "learning_rate": 1.3587142857142858e-06, + "loss": 0.879, + "step": 4490 + }, + { + "epoch": 0.6523866478199413, + "grad_norm": 2.59375, + "learning_rate": 1.3572857142857141e-06, + "loss": 0.8111, + "step": 4500 + }, + { + "epoch": 0.6538363959262078, + "grad_norm": 2.828125, + "learning_rate": 1.355857142857143e-06, + "loss": 0.8273, + "step": 4510 + }, + { + "epoch": 0.6552861440324743, + "grad_norm": 2.75, + "learning_rate": 1.3544285714285715e-06, + "loss": 0.7983, + "step": 4520 + }, + { + "epoch": 0.6567358921387408, + "grad_norm": 2.46875, + "learning_rate": 1.353e-06, + "loss": 0.8547, + "step": 4530 + }, + { + "epoch": 0.6581856402450075, + "grad_norm": 2.1875, + "learning_rate": 1.3515714285714286e-06, + "loss": 0.788, + "step": 4540 + }, + { + "epoch": 0.659635388351274, + "grad_norm": 2.6875, + "learning_rate": 1.3501428571428571e-06, + "loss": 0.8447, + "step": 4550 + }, + { + "epoch": 0.6610851364575405, + "grad_norm": 2.625, + "learning_rate": 1.3487142857142857e-06, + "loss": 0.8145, + "step": 4560 + }, + { + "epoch": 0.662534884563807, + "grad_norm": 2.890625, + "learning_rate": 1.3472857142857142e-06, + "loss": 0.8199, + "step": 4570 + }, + { + "epoch": 0.6639846326700736, + "grad_norm": 3.359375, + "learning_rate": 1.3458571428571428e-06, + "loss": 0.7918, + "step": 4580 + }, + { + "epoch": 0.6654343807763401, + "grad_norm": 2.265625, + "learning_rate": 1.3444285714285715e-06, + "loss": 0.7996, + "step": 4590 + }, + { + "epoch": 0.6668841288826066, + "grad_norm": 2.5, + "learning_rate": 1.3429999999999999e-06, + "loss": 0.8048, + "step": 4600 + }, + { + "epoch": 0.6683338769888731, + "grad_norm": 2.8125, + "learning_rate": 1.3415714285714287e-06, + "loss": 0.8284, + "step": 4610 + }, + { + "epoch": 0.6697836250951397, + "grad_norm": 2.328125, + "learning_rate": 1.340142857142857e-06, + "loss": 0.7885, + "step": 4620 + }, + { + "epoch": 0.6712333732014063, + "grad_norm": 2.203125, + "learning_rate": 1.3387142857142858e-06, + "loss": 0.7696, + "step": 4630 + }, + { + "epoch": 0.6726831213076728, + "grad_norm": 2.453125, + "learning_rate": 1.3372857142857141e-06, + "loss": 0.8068, + "step": 4640 + }, + { + "epoch": 0.6741328694139394, + "grad_norm": 3.21875, + "learning_rate": 1.3358571428571429e-06, + "loss": 0.8056, + "step": 4650 + }, + { + "epoch": 0.6755826175202059, + "grad_norm": 2.546875, + "learning_rate": 1.3344285714285714e-06, + "loss": 0.8488, + "step": 4660 + }, + { + "epoch": 0.6770323656264724, + "grad_norm": 3.09375, + "learning_rate": 1.333e-06, + "loss": 0.8022, + "step": 4670 + }, + { + "epoch": 0.6784821137327389, + "grad_norm": 2.75, + "learning_rate": 1.3315714285714285e-06, + "loss": 0.8154, + "step": 4680 + }, + { + "epoch": 0.6799318618390054, + "grad_norm": 2.390625, + "learning_rate": 1.330142857142857e-06, + "loss": 0.8111, + "step": 4690 + }, + { + "epoch": 0.681381609945272, + "grad_norm": 2.890625, + "learning_rate": 1.3287142857142856e-06, + "loss": 0.8737, + "step": 4700 + }, + { + "epoch": 0.6828313580515385, + "grad_norm": 2.796875, + "learning_rate": 1.3272857142857142e-06, + "loss": 0.804, + "step": 4710 + }, + { + "epoch": 0.6842811061578051, + "grad_norm": 2.484375, + "learning_rate": 1.3258571428571428e-06, + "loss": 0.7909, + "step": 4720 + }, + { + "epoch": 0.6857308542640717, + "grad_norm": 2.75, + "learning_rate": 1.3244285714285715e-06, + "loss": 0.809, + "step": 4730 + }, + { + "epoch": 0.6871806023703382, + "grad_norm": 2.421875, + "learning_rate": 1.3229999999999999e-06, + "loss": 0.7717, + "step": 4740 + }, + { + "epoch": 0.6886303504766047, + "grad_norm": 2.6875, + "learning_rate": 1.3215714285714286e-06, + "loss": 0.8329, + "step": 4750 + }, + { + "epoch": 0.6900800985828712, + "grad_norm": 2.578125, + "learning_rate": 1.320142857142857e-06, + "loss": 0.8021, + "step": 4760 + }, + { + "epoch": 0.6915298466891378, + "grad_norm": 2.84375, + "learning_rate": 1.3187142857142857e-06, + "loss": 0.8835, + "step": 4770 + }, + { + "epoch": 0.6929795947954043, + "grad_norm": 2.5625, + "learning_rate": 1.3172857142857143e-06, + "loss": 0.8135, + "step": 4780 + }, + { + "epoch": 0.6944293429016708, + "grad_norm": 2.859375, + "learning_rate": 1.3158571428571429e-06, + "loss": 0.82, + "step": 4790 + }, + { + "epoch": 0.6958790910079373, + "grad_norm": 2.640625, + "learning_rate": 1.3144285714285714e-06, + "loss": 0.8141, + "step": 4800 + }, + { + "epoch": 0.697328839114204, + "grad_norm": 2.984375, + "learning_rate": 1.313e-06, + "loss": 0.8873, + "step": 4810 + }, + { + "epoch": 0.6987785872204705, + "grad_norm": 3.203125, + "learning_rate": 1.3115714285714285e-06, + "loss": 0.9005, + "step": 4820 + }, + { + "epoch": 0.700228335326737, + "grad_norm": 3.3125, + "learning_rate": 1.310142857142857e-06, + "loss": 0.774, + "step": 4830 + }, + { + "epoch": 0.7016780834330035, + "grad_norm": 2.453125, + "learning_rate": 1.3087142857142856e-06, + "loss": 0.8386, + "step": 4840 + }, + { + "epoch": 0.70312783153927, + "grad_norm": 2.84375, + "learning_rate": 1.3072857142857144e-06, + "loss": 0.8088, + "step": 4850 + }, + { + "epoch": 0.7045775796455366, + "grad_norm": 2.796875, + "learning_rate": 1.3058571428571427e-06, + "loss": 0.8616, + "step": 4860 + }, + { + "epoch": 0.7060273277518031, + "grad_norm": 2.5625, + "learning_rate": 1.3044285714285715e-06, + "loss": 0.8544, + "step": 4870 + }, + { + "epoch": 0.7074770758580696, + "grad_norm": 2.40625, + "learning_rate": 1.3029999999999998e-06, + "loss": 0.8872, + "step": 4880 + }, + { + "epoch": 0.7089268239643362, + "grad_norm": 3.296875, + "learning_rate": 1.3015714285714286e-06, + "loss": 0.8138, + "step": 4890 + }, + { + "epoch": 0.7103765720706028, + "grad_norm": 2.484375, + "learning_rate": 1.300142857142857e-06, + "loss": 0.7948, + "step": 4900 + }, + { + "epoch": 0.7118263201768693, + "grad_norm": 2.46875, + "learning_rate": 1.2987142857142857e-06, + "loss": 0.8441, + "step": 4910 + }, + { + "epoch": 0.7132760682831358, + "grad_norm": 2.265625, + "learning_rate": 1.2972857142857143e-06, + "loss": 0.7882, + "step": 4920 + }, + { + "epoch": 0.7147258163894024, + "grad_norm": 2.6875, + "learning_rate": 1.2958571428571428e-06, + "loss": 0.8114, + "step": 4930 + }, + { + "epoch": 0.7161755644956689, + "grad_norm": 3.109375, + "learning_rate": 1.2944285714285714e-06, + "loss": 0.8575, + "step": 4940 + }, + { + "epoch": 0.7176253126019354, + "grad_norm": 2.796875, + "learning_rate": 1.293e-06, + "loss": 0.8013, + "step": 4950 + }, + { + "epoch": 0.7190750607082019, + "grad_norm": 2.671875, + "learning_rate": 1.2915714285714285e-06, + "loss": 0.823, + "step": 4960 + }, + { + "epoch": 0.7205248088144685, + "grad_norm": 2.734375, + "learning_rate": 1.290142857142857e-06, + "loss": 0.8143, + "step": 4970 + }, + { + "epoch": 0.721974556920735, + "grad_norm": 2.609375, + "learning_rate": 1.2887142857142856e-06, + "loss": 0.7749, + "step": 4980 + }, + { + "epoch": 0.7234243050270016, + "grad_norm": 2.828125, + "learning_rate": 1.2872857142857144e-06, + "loss": 0.8161, + "step": 4990 + }, + { + "epoch": 0.7248740531332681, + "grad_norm": 2.359375, + "learning_rate": 1.2858571428571427e-06, + "loss": 0.8073, + "step": 5000 + }, + { + "epoch": 0.7263238012395347, + "grad_norm": 2.625, + "learning_rate": 1.2844285714285715e-06, + "loss": 0.8291, + "step": 5010 + }, + { + "epoch": 0.7277735493458012, + "grad_norm": 2.9375, + "learning_rate": 1.2829999999999998e-06, + "loss": 0.8544, + "step": 5020 + }, + { + "epoch": 0.7292232974520677, + "grad_norm": 2.8125, + "learning_rate": 1.2815714285714286e-06, + "loss": 0.8162, + "step": 5030 + }, + { + "epoch": 0.7306730455583342, + "grad_norm": 2.59375, + "learning_rate": 1.2801428571428572e-06, + "loss": 0.8541, + "step": 5040 + }, + { + "epoch": 0.7321227936646008, + "grad_norm": 2.40625, + "learning_rate": 1.2787142857142857e-06, + "loss": 0.8604, + "step": 5050 + }, + { + "epoch": 0.7335725417708673, + "grad_norm": 2.375, + "learning_rate": 1.2772857142857143e-06, + "loss": 0.772, + "step": 5060 + }, + { + "epoch": 0.7350222898771338, + "grad_norm": 3.21875, + "learning_rate": 1.2758571428571428e-06, + "loss": 0.8248, + "step": 5070 + }, + { + "epoch": 0.7364720379834003, + "grad_norm": 2.609375, + "learning_rate": 1.2744285714285714e-06, + "loss": 0.7764, + "step": 5080 + }, + { + "epoch": 0.737921786089667, + "grad_norm": 3.234375, + "learning_rate": 1.273e-06, + "loss": 0.8238, + "step": 5090 + }, + { + "epoch": 0.7393715341959335, + "grad_norm": 2.59375, + "learning_rate": 1.2715714285714285e-06, + "loss": 0.7802, + "step": 5100 + }, + { + "epoch": 0.7408212823022, + "grad_norm": 2.71875, + "learning_rate": 1.2701428571428572e-06, + "loss": 0.8374, + "step": 5110 + }, + { + "epoch": 0.7422710304084665, + "grad_norm": 2.75, + "learning_rate": 1.2687142857142856e-06, + "loss": 0.7997, + "step": 5120 + }, + { + "epoch": 0.7437207785147331, + "grad_norm": 2.3125, + "learning_rate": 1.2672857142857144e-06, + "loss": 0.77, + "step": 5130 + }, + { + "epoch": 0.7451705266209996, + "grad_norm": 2.71875, + "learning_rate": 1.2658571428571427e-06, + "loss": 0.7958, + "step": 5140 + }, + { + "epoch": 0.7466202747272661, + "grad_norm": 2.5625, + "learning_rate": 1.2644285714285715e-06, + "loss": 0.7644, + "step": 5150 + }, + { + "epoch": 0.7480700228335326, + "grad_norm": 2.40625, + "learning_rate": 1.2629999999999998e-06, + "loss": 0.7814, + "step": 5160 + }, + { + "epoch": 0.7495197709397992, + "grad_norm": 2.734375, + "learning_rate": 1.2615714285714286e-06, + "loss": 0.8337, + "step": 5170 + }, + { + "epoch": 0.7509695190460658, + "grad_norm": 2.59375, + "learning_rate": 1.2601428571428571e-06, + "loss": 0.8868, + "step": 5180 + }, + { + "epoch": 0.7524192671523323, + "grad_norm": 2.546875, + "learning_rate": 1.2587142857142857e-06, + "loss": 0.8678, + "step": 5190 + }, + { + "epoch": 0.7538690152585988, + "grad_norm": 2.25, + "learning_rate": 1.2572857142857142e-06, + "loss": 0.7822, + "step": 5200 + }, + { + "epoch": 0.7553187633648654, + "grad_norm": 3.203125, + "learning_rate": 1.2558571428571428e-06, + "loss": 0.8111, + "step": 5210 + }, + { + "epoch": 0.7567685114711319, + "grad_norm": 2.484375, + "learning_rate": 1.2544285714285713e-06, + "loss": 0.7549, + "step": 5220 + }, + { + "epoch": 0.7582182595773984, + "grad_norm": 2.8125, + "learning_rate": 1.253e-06, + "loss": 0.8656, + "step": 5230 + }, + { + "epoch": 0.7596680076836649, + "grad_norm": 2.875, + "learning_rate": 1.2515714285714285e-06, + "loss": 0.8771, + "step": 5240 + }, + { + "epoch": 0.7611177557899315, + "grad_norm": 2.5625, + "learning_rate": 1.2501428571428572e-06, + "loss": 0.7582, + "step": 5250 + }, + { + "epoch": 0.762567503896198, + "grad_norm": 2.609375, + "learning_rate": 1.2487142857142856e-06, + "loss": 0.8208, + "step": 5260 + }, + { + "epoch": 0.7640172520024646, + "grad_norm": 2.71875, + "learning_rate": 1.2472857142857143e-06, + "loss": 0.8268, + "step": 5270 + }, + { + "epoch": 0.7654670001087311, + "grad_norm": 2.640625, + "learning_rate": 1.2458571428571427e-06, + "loss": 0.8339, + "step": 5280 + }, + { + "epoch": 0.7669167482149977, + "grad_norm": 2.921875, + "learning_rate": 1.2444285714285714e-06, + "loss": 0.8306, + "step": 5290 + }, + { + "epoch": 0.7683664963212642, + "grad_norm": 2.96875, + "learning_rate": 1.243e-06, + "loss": 0.8365, + "step": 5300 + }, + { + "epoch": 0.7698162444275307, + "grad_norm": 2.984375, + "learning_rate": 1.2415714285714286e-06, + "loss": 0.8504, + "step": 5310 + }, + { + "epoch": 0.7712659925337972, + "grad_norm": 2.84375, + "learning_rate": 1.2401428571428571e-06, + "loss": 0.8195, + "step": 5320 + }, + { + "epoch": 0.7727157406400638, + "grad_norm": 2.75, + "learning_rate": 1.2387142857142857e-06, + "loss": 0.8476, + "step": 5330 + }, + { + "epoch": 0.7741654887463303, + "grad_norm": 2.625, + "learning_rate": 1.2372857142857142e-06, + "loss": 0.8107, + "step": 5340 + }, + { + "epoch": 0.7756152368525968, + "grad_norm": 2.671875, + "learning_rate": 1.2358571428571428e-06, + "loss": 0.8816, + "step": 5350 + }, + { + "epoch": 0.7770649849588634, + "grad_norm": 2.578125, + "learning_rate": 1.2344285714285713e-06, + "loss": 0.792, + "step": 5360 + }, + { + "epoch": 0.77851473306513, + "grad_norm": 2.765625, + "learning_rate": 1.233e-06, + "loss": 0.8008, + "step": 5370 + }, + { + "epoch": 0.7799644811713965, + "grad_norm": 2.625, + "learning_rate": 1.2315714285714284e-06, + "loss": 0.8397, + "step": 5380 + }, + { + "epoch": 0.781414229277663, + "grad_norm": 2.3125, + "learning_rate": 1.2301428571428572e-06, + "loss": 0.7887, + "step": 5390 + }, + { + "epoch": 0.7828639773839295, + "grad_norm": 2.5, + "learning_rate": 1.2287142857142855e-06, + "loss": 0.8641, + "step": 5400 + }, + { + "epoch": 0.7843137254901961, + "grad_norm": 3.0625, + "learning_rate": 1.2272857142857143e-06, + "loss": 0.8502, + "step": 5410 + }, + { + "epoch": 0.7857634735964626, + "grad_norm": 2.84375, + "learning_rate": 1.2258571428571427e-06, + "loss": 0.8217, + "step": 5420 + }, + { + "epoch": 0.7872132217027291, + "grad_norm": 2.921875, + "learning_rate": 1.2244285714285714e-06, + "loss": 0.7995, + "step": 5430 + }, + { + "epoch": 0.7886629698089956, + "grad_norm": 2.390625, + "learning_rate": 1.223e-06, + "loss": 0.8229, + "step": 5440 + }, + { + "epoch": 0.7901127179152623, + "grad_norm": 2.359375, + "learning_rate": 1.2215714285714285e-06, + "loss": 0.7973, + "step": 5450 + }, + { + "epoch": 0.7915624660215288, + "grad_norm": 3.1875, + "learning_rate": 1.220142857142857e-06, + "loss": 0.7657, + "step": 5460 + }, + { + "epoch": 0.7930122141277953, + "grad_norm": 2.71875, + "learning_rate": 1.2187142857142856e-06, + "loss": 0.831, + "step": 5470 + }, + { + "epoch": 0.7944619622340618, + "grad_norm": 2.59375, + "learning_rate": 1.2172857142857142e-06, + "loss": 0.796, + "step": 5480 + }, + { + "epoch": 0.7959117103403284, + "grad_norm": 2.515625, + "learning_rate": 1.2158571428571428e-06, + "loss": 0.8035, + "step": 5490 + }, + { + "epoch": 0.7973614584465949, + "grad_norm": 3.046875, + "learning_rate": 1.2144285714285713e-06, + "loss": 0.8123, + "step": 5500 + }, + { + "epoch": 0.7988112065528614, + "grad_norm": 2.75, + "learning_rate": 1.213e-06, + "loss": 0.8356, + "step": 5510 + }, + { + "epoch": 0.8002609546591279, + "grad_norm": 2.734375, + "learning_rate": 1.2115714285714284e-06, + "loss": 0.8358, + "step": 5520 + }, + { + "epoch": 0.8017107027653945, + "grad_norm": 5.65625, + "learning_rate": 1.2101428571428572e-06, + "loss": 0.7915, + "step": 5530 + }, + { + "epoch": 0.8031604508716611, + "grad_norm": 2.90625, + "learning_rate": 1.2087142857142855e-06, + "loss": 0.7561, + "step": 5540 + }, + { + "epoch": 0.8046101989779276, + "grad_norm": 2.25, + "learning_rate": 1.2072857142857143e-06, + "loss": 0.8569, + "step": 5550 + }, + { + "epoch": 0.8060599470841942, + "grad_norm": 2.640625, + "learning_rate": 1.2058571428571428e-06, + "loss": 0.8104, + "step": 5560 + }, + { + "epoch": 0.8075096951904607, + "grad_norm": 2.53125, + "learning_rate": 1.2044285714285714e-06, + "loss": 0.8399, + "step": 5570 + }, + { + "epoch": 0.8089594432967272, + "grad_norm": 2.421875, + "learning_rate": 1.203e-06, + "loss": 0.7867, + "step": 5580 + }, + { + "epoch": 0.8104091914029937, + "grad_norm": 2.96875, + "learning_rate": 1.2015714285714285e-06, + "loss": 0.8196, + "step": 5590 + }, + { + "epoch": 0.8118589395092602, + "grad_norm": 3.125, + "learning_rate": 1.200142857142857e-06, + "loss": 0.8408, + "step": 5600 + }, + { + "epoch": 0.8133086876155268, + "grad_norm": 2.578125, + "learning_rate": 1.1987142857142856e-06, + "loss": 0.7773, + "step": 5610 + }, + { + "epoch": 0.8147584357217933, + "grad_norm": 2.9375, + "learning_rate": 1.1972857142857142e-06, + "loss": 0.8397, + "step": 5620 + }, + { + "epoch": 0.8162081838280599, + "grad_norm": 2.46875, + "learning_rate": 1.195857142857143e-06, + "loss": 0.8039, + "step": 5630 + }, + { + "epoch": 0.8176579319343265, + "grad_norm": 2.5, + "learning_rate": 1.1944285714285713e-06, + "loss": 0.8339, + "step": 5640 + }, + { + "epoch": 0.819107680040593, + "grad_norm": 2.65625, + "learning_rate": 1.193e-06, + "loss": 0.7921, + "step": 5650 + }, + { + "epoch": 0.8205574281468595, + "grad_norm": 2.515625, + "learning_rate": 1.1915714285714284e-06, + "loss": 0.8002, + "step": 5660 + }, + { + "epoch": 0.822007176253126, + "grad_norm": 2.578125, + "learning_rate": 1.1901428571428572e-06, + "loss": 0.7698, + "step": 5670 + }, + { + "epoch": 0.8234569243593926, + "grad_norm": 3.015625, + "learning_rate": 1.1887142857142855e-06, + "loss": 0.7568, + "step": 5680 + }, + { + "epoch": 0.8249066724656591, + "grad_norm": 2.265625, + "learning_rate": 1.1872857142857143e-06, + "loss": 0.7839, + "step": 5690 + }, + { + "epoch": 0.8263564205719256, + "grad_norm": 2.34375, + "learning_rate": 1.1858571428571428e-06, + "loss": 0.7954, + "step": 5700 + }, + { + "epoch": 0.8278061686781921, + "grad_norm": 3.265625, + "learning_rate": 1.1844285714285714e-06, + "loss": 0.8033, + "step": 5710 + }, + { + "epoch": 0.8292559167844586, + "grad_norm": 2.359375, + "learning_rate": 1.183e-06, + "loss": 0.8199, + "step": 5720 + }, + { + "epoch": 0.8307056648907253, + "grad_norm": 2.78125, + "learning_rate": 1.1815714285714285e-06, + "loss": 0.806, + "step": 5730 + }, + { + "epoch": 0.8321554129969918, + "grad_norm": 2.359375, + "learning_rate": 1.180142857142857e-06, + "loss": 0.8187, + "step": 5740 + }, + { + "epoch": 0.8336051611032583, + "grad_norm": 2.75, + "learning_rate": 1.1787142857142856e-06, + "loss": 0.8368, + "step": 5750 + }, + { + "epoch": 0.8350549092095249, + "grad_norm": 2.6875, + "learning_rate": 1.1772857142857142e-06, + "loss": 0.8022, + "step": 5760 + }, + { + "epoch": 0.8365046573157914, + "grad_norm": 2.359375, + "learning_rate": 1.175857142857143e-06, + "loss": 0.8377, + "step": 5770 + }, + { + "epoch": 0.8379544054220579, + "grad_norm": 11.25, + "learning_rate": 1.1744285714285713e-06, + "loss": 0.8403, + "step": 5780 + }, + { + "epoch": 0.8394041535283244, + "grad_norm": 2.453125, + "learning_rate": 1.173e-06, + "loss": 0.7995, + "step": 5790 + }, + { + "epoch": 0.840853901634591, + "grad_norm": 2.5, + "learning_rate": 1.1715714285714284e-06, + "loss": 0.7676, + "step": 5800 + }, + { + "epoch": 0.8423036497408575, + "grad_norm": 2.828125, + "learning_rate": 1.1701428571428571e-06, + "loss": 0.811, + "step": 5810 + }, + { + "epoch": 0.8437533978471241, + "grad_norm": 2.765625, + "learning_rate": 1.1687142857142857e-06, + "loss": 0.7798, + "step": 5820 + }, + { + "epoch": 0.8452031459533906, + "grad_norm": 2.765625, + "learning_rate": 1.1672857142857143e-06, + "loss": 0.8597, + "step": 5830 + }, + { + "epoch": 0.8466528940596572, + "grad_norm": 2.875, + "learning_rate": 1.1658571428571428e-06, + "loss": 0.8837, + "step": 5840 + }, + { + "epoch": 0.8481026421659237, + "grad_norm": 2.609375, + "learning_rate": 1.1644285714285714e-06, + "loss": 0.7881, + "step": 5850 + }, + { + "epoch": 0.8495523902721902, + "grad_norm": 2.640625, + "learning_rate": 1.163e-06, + "loss": 0.7114, + "step": 5860 + }, + { + "epoch": 0.8510021383784567, + "grad_norm": 2.546875, + "learning_rate": 1.1615714285714285e-06, + "loss": 0.8134, + "step": 5870 + }, + { + "epoch": 0.8524518864847233, + "grad_norm": 2.359375, + "learning_rate": 1.160142857142857e-06, + "loss": 0.7709, + "step": 5880 + }, + { + "epoch": 0.8539016345909898, + "grad_norm": 2.6875, + "learning_rate": 1.1587142857142858e-06, + "loss": 0.7973, + "step": 5890 + }, + { + "epoch": 0.8553513826972563, + "grad_norm": 3.359375, + "learning_rate": 1.1572857142857141e-06, + "loss": 0.837, + "step": 5900 + }, + { + "epoch": 0.8568011308035229, + "grad_norm": 3.03125, + "learning_rate": 1.155857142857143e-06, + "loss": 0.7469, + "step": 5910 + }, + { + "epoch": 0.8582508789097895, + "grad_norm": 2.65625, + "learning_rate": 1.1544285714285712e-06, + "loss": 0.8283, + "step": 5920 + }, + { + "epoch": 0.859700627016056, + "grad_norm": 2.3125, + "learning_rate": 1.153e-06, + "loss": 0.7947, + "step": 5930 + }, + { + "epoch": 0.8611503751223225, + "grad_norm": 2.59375, + "learning_rate": 1.1515714285714284e-06, + "loss": 0.8102, + "step": 5940 + }, + { + "epoch": 0.862600123228589, + "grad_norm": 2.625, + "learning_rate": 1.1501428571428571e-06, + "loss": 0.8496, + "step": 5950 + }, + { + "epoch": 0.8640498713348556, + "grad_norm": 2.078125, + "learning_rate": 1.1487142857142857e-06, + "loss": 0.7966, + "step": 5960 + }, + { + "epoch": 0.8654996194411221, + "grad_norm": 2.484375, + "learning_rate": 1.1472857142857142e-06, + "loss": 0.8099, + "step": 5970 + }, + { + "epoch": 0.8669493675473886, + "grad_norm": 3.515625, + "learning_rate": 1.1458571428571428e-06, + "loss": 0.7931, + "step": 5980 + }, + { + "epoch": 0.8683991156536551, + "grad_norm": 2.453125, + "learning_rate": 1.1444285714285713e-06, + "loss": 0.793, + "step": 5990 + }, + { + "epoch": 0.8698488637599218, + "grad_norm": 2.28125, + "learning_rate": 1.143e-06, + "loss": 0.8572, + "step": 6000 + }, + { + "epoch": 0.8712986118661883, + "grad_norm": 2.15625, + "learning_rate": 1.1415714285714287e-06, + "loss": 0.8288, + "step": 6010 + }, + { + "epoch": 0.8727483599724548, + "grad_norm": 2.9375, + "learning_rate": 1.140142857142857e-06, + "loss": 0.7878, + "step": 6020 + }, + { + "epoch": 0.8741981080787213, + "grad_norm": 2.515625, + "learning_rate": 1.1387142857142858e-06, + "loss": 0.7805, + "step": 6030 + }, + { + "epoch": 0.8756478561849879, + "grad_norm": 2.703125, + "learning_rate": 1.1372857142857141e-06, + "loss": 0.8231, + "step": 6040 + }, + { + "epoch": 0.8770976042912544, + "grad_norm": 3.09375, + "learning_rate": 1.1358571428571429e-06, + "loss": 0.7834, + "step": 6050 + }, + { + "epoch": 0.8785473523975209, + "grad_norm": 2.828125, + "learning_rate": 1.1344285714285712e-06, + "loss": 0.84, + "step": 6060 + }, + { + "epoch": 0.8799971005037874, + "grad_norm": 2.484375, + "learning_rate": 1.133e-06, + "loss": 0.8107, + "step": 6070 + }, + { + "epoch": 0.881446848610054, + "grad_norm": 2.890625, + "learning_rate": 1.1315714285714285e-06, + "loss": 0.8414, + "step": 6080 + }, + { + "epoch": 0.8828965967163206, + "grad_norm": 2.40625, + "learning_rate": 1.130142857142857e-06, + "loss": 0.807, + "step": 6090 + }, + { + "epoch": 0.8843463448225871, + "grad_norm": 2.859375, + "learning_rate": 1.1287142857142857e-06, + "loss": 0.8326, + "step": 6100 + }, + { + "epoch": 0.8857960929288536, + "grad_norm": 2.359375, + "learning_rate": 1.1272857142857142e-06, + "loss": 0.811, + "step": 6110 + }, + { + "epoch": 0.8872458410351202, + "grad_norm": 2.546875, + "learning_rate": 1.1258571428571428e-06, + "loss": 0.8088, + "step": 6120 + }, + { + "epoch": 0.8886955891413867, + "grad_norm": 2.296875, + "learning_rate": 1.1244285714285713e-06, + "loss": 0.7912, + "step": 6130 + }, + { + "epoch": 0.8901453372476532, + "grad_norm": 3.046875, + "learning_rate": 1.1229999999999999e-06, + "loss": 0.8032, + "step": 6140 + }, + { + "epoch": 0.8915950853539197, + "grad_norm": 3.3125, + "learning_rate": 1.1215714285714286e-06, + "loss": 0.83, + "step": 6150 + }, + { + "epoch": 0.8930448334601863, + "grad_norm": 2.796875, + "learning_rate": 1.120142857142857e-06, + "loss": 0.8355, + "step": 6160 + }, + { + "epoch": 0.8944945815664528, + "grad_norm": 2.84375, + "learning_rate": 1.1187142857142858e-06, + "loss": 0.8224, + "step": 6170 + }, + { + "epoch": 0.8959443296727194, + "grad_norm": 2.390625, + "learning_rate": 1.117285714285714e-06, + "loss": 0.7877, + "step": 6180 + }, + { + "epoch": 0.8973940777789859, + "grad_norm": 2.59375, + "learning_rate": 1.1158571428571429e-06, + "loss": 0.8915, + "step": 6190 + }, + { + "epoch": 0.8988438258852525, + "grad_norm": 2.734375, + "learning_rate": 1.1144285714285712e-06, + "loss": 0.8118, + "step": 6200 + }, + { + "epoch": 0.900293573991519, + "grad_norm": 2.8125, + "learning_rate": 1.113e-06, + "loss": 0.8176, + "step": 6210 + }, + { + "epoch": 0.9017433220977855, + "grad_norm": 3.0, + "learning_rate": 1.1115714285714285e-06, + "loss": 0.9135, + "step": 6220 + }, + { + "epoch": 0.903193070204052, + "grad_norm": 3.34375, + "learning_rate": 1.110142857142857e-06, + "loss": 0.8326, + "step": 6230 + }, + { + "epoch": 0.9046428183103186, + "grad_norm": 2.78125, + "learning_rate": 1.1087142857142856e-06, + "loss": 0.814, + "step": 6240 + }, + { + "epoch": 0.9060925664165851, + "grad_norm": 3.421875, + "learning_rate": 1.1072857142857142e-06, + "loss": 0.8268, + "step": 6250 + }, + { + "epoch": 0.9075423145228516, + "grad_norm": 2.609375, + "learning_rate": 1.1058571428571427e-06, + "loss": 0.7706, + "step": 6260 + }, + { + "epoch": 0.9089920626291181, + "grad_norm": 2.890625, + "learning_rate": 1.1044285714285715e-06, + "loss": 0.7912, + "step": 6270 + }, + { + "epoch": 0.9104418107353848, + "grad_norm": 2.421875, + "learning_rate": 1.1029999999999999e-06, + "loss": 0.8289, + "step": 6280 + }, + { + "epoch": 0.9118915588416513, + "grad_norm": 2.296875, + "learning_rate": 1.1015714285714286e-06, + "loss": 0.7749, + "step": 6290 + }, + { + "epoch": 0.9133413069479178, + "grad_norm": 1.9765625, + "learning_rate": 1.100142857142857e-06, + "loss": 0.8111, + "step": 6300 + }, + { + "epoch": 0.9147910550541843, + "grad_norm": 2.765625, + "learning_rate": 1.0987142857142857e-06, + "loss": 0.8275, + "step": 6310 + }, + { + "epoch": 0.9162408031604509, + "grad_norm": 2.234375, + "learning_rate": 1.097285714285714e-06, + "loss": 0.8241, + "step": 6320 + }, + { + "epoch": 0.9176905512667174, + "grad_norm": 2.71875, + "learning_rate": 1.0958571428571428e-06, + "loss": 0.8487, + "step": 6330 + }, + { + "epoch": 0.9191402993729839, + "grad_norm": 2.59375, + "learning_rate": 1.0944285714285714e-06, + "loss": 0.8626, + "step": 6340 + }, + { + "epoch": 0.9205900474792504, + "grad_norm": 2.25, + "learning_rate": 1.093e-06, + "loss": 0.8114, + "step": 6350 + }, + { + "epoch": 0.922039795585517, + "grad_norm": 3.328125, + "learning_rate": 1.0915714285714285e-06, + "loss": 0.8226, + "step": 6360 + }, + { + "epoch": 0.9234895436917836, + "grad_norm": 2.859375, + "learning_rate": 1.090142857142857e-06, + "loss": 0.8169, + "step": 6370 + }, + { + "epoch": 0.9249392917980501, + "grad_norm": 2.484375, + "learning_rate": 1.0887142857142856e-06, + "loss": 0.7731, + "step": 6380 + }, + { + "epoch": 0.9263890399043166, + "grad_norm": 2.609375, + "learning_rate": 1.0872857142857142e-06, + "loss": 0.8196, + "step": 6390 + }, + { + "epoch": 0.9278387880105832, + "grad_norm": 2.984375, + "learning_rate": 1.0858571428571427e-06, + "loss": 0.8986, + "step": 6400 + }, + { + "epoch": 0.9292885361168497, + "grad_norm": 2.609375, + "learning_rate": 1.0844285714285715e-06, + "loss": 0.8374, + "step": 6410 + }, + { + "epoch": 0.9307382842231162, + "grad_norm": 2.296875, + "learning_rate": 1.0829999999999998e-06, + "loss": 0.7657, + "step": 6420 + }, + { + "epoch": 0.9321880323293827, + "grad_norm": 2.84375, + "learning_rate": 1.0815714285714286e-06, + "loss": 0.8308, + "step": 6430 + }, + { + "epoch": 0.9336377804356493, + "grad_norm": 2.453125, + "learning_rate": 1.080142857142857e-06, + "loss": 0.8192, + "step": 6440 + }, + { + "epoch": 0.9350875285419158, + "grad_norm": 2.921875, + "learning_rate": 1.0787142857142857e-06, + "loss": 0.8386, + "step": 6450 + }, + { + "epoch": 0.9365372766481824, + "grad_norm": 2.5, + "learning_rate": 1.077285714285714e-06, + "loss": 0.8318, + "step": 6460 + }, + { + "epoch": 0.937987024754449, + "grad_norm": 2.953125, + "learning_rate": 1.0758571428571428e-06, + "loss": 0.8701, + "step": 6470 + }, + { + "epoch": 0.9394367728607155, + "grad_norm": 2.53125, + "learning_rate": 1.0744285714285714e-06, + "loss": 0.8149, + "step": 6480 + }, + { + "epoch": 0.940886520966982, + "grad_norm": 2.296875, + "learning_rate": 1.073e-06, + "loss": 0.7941, + "step": 6490 + }, + { + "epoch": 0.9423362690732485, + "grad_norm": 2.75, + "learning_rate": 1.0715714285714285e-06, + "loss": 0.8047, + "step": 6500 + }, + { + "epoch": 0.943786017179515, + "grad_norm": 2.34375, + "learning_rate": 1.070142857142857e-06, + "loss": 0.7644, + "step": 6510 + }, + { + "epoch": 0.9452357652857816, + "grad_norm": 2.890625, + "learning_rate": 1.0687142857142856e-06, + "loss": 0.7944, + "step": 6520 + }, + { + "epoch": 0.9466855133920481, + "grad_norm": 2.453125, + "learning_rate": 1.0672857142857144e-06, + "loss": 0.8198, + "step": 6530 + }, + { + "epoch": 0.9481352614983146, + "grad_norm": 2.9375, + "learning_rate": 1.0658571428571427e-06, + "loss": 0.8423, + "step": 6540 + }, + { + "epoch": 0.9495850096045813, + "grad_norm": 2.28125, + "learning_rate": 1.0644285714285715e-06, + "loss": 0.8561, + "step": 6550 + }, + { + "epoch": 0.9510347577108478, + "grad_norm": 2.34375, + "learning_rate": 1.0629999999999998e-06, + "loss": 0.8211, + "step": 6560 + }, + { + "epoch": 0.9524845058171143, + "grad_norm": 2.578125, + "learning_rate": 1.0615714285714286e-06, + "loss": 0.7708, + "step": 6570 + }, + { + "epoch": 0.9539342539233808, + "grad_norm": 2.734375, + "learning_rate": 1.060142857142857e-06, + "loss": 0.7889, + "step": 6580 + }, + { + "epoch": 0.9553840020296474, + "grad_norm": 2.46875, + "learning_rate": 1.0587142857142857e-06, + "loss": 0.7942, + "step": 6590 + }, + { + "epoch": 0.9568337501359139, + "grad_norm": 2.359375, + "learning_rate": 1.0572857142857142e-06, + "loss": 0.7732, + "step": 6600 + }, + { + "epoch": 0.9582834982421804, + "grad_norm": 2.796875, + "learning_rate": 1.0558571428571428e-06, + "loss": 0.8532, + "step": 6610 + }, + { + "epoch": 0.9597332463484469, + "grad_norm": 2.453125, + "learning_rate": 1.0544285714285714e-06, + "loss": 0.8059, + "step": 6620 + }, + { + "epoch": 0.9611829944547134, + "grad_norm": 2.375, + "learning_rate": 1.053e-06, + "loss": 0.83, + "step": 6630 + }, + { + "epoch": 0.9626327425609801, + "grad_norm": 2.734375, + "learning_rate": 1.0515714285714285e-06, + "loss": 0.785, + "step": 6640 + }, + { + "epoch": 0.9640824906672466, + "grad_norm": 2.546875, + "learning_rate": 1.050142857142857e-06, + "loss": 0.8163, + "step": 6650 + }, + { + "epoch": 0.9655322387735131, + "grad_norm": 2.75, + "learning_rate": 1.0487142857142856e-06, + "loss": 0.7694, + "step": 6660 + }, + { + "epoch": 0.9669819868797797, + "grad_norm": 2.546875, + "learning_rate": 1.0472857142857143e-06, + "loss": 0.8601, + "step": 6670 + }, + { + "epoch": 0.9684317349860462, + "grad_norm": 2.265625, + "learning_rate": 1.0458571428571427e-06, + "loss": 0.8126, + "step": 6680 + }, + { + "epoch": 0.9698814830923127, + "grad_norm": 3.109375, + "learning_rate": 1.0444285714285715e-06, + "loss": 0.8167, + "step": 6690 + }, + { + "epoch": 0.9713312311985792, + "grad_norm": 2.875, + "learning_rate": 1.0429999999999998e-06, + "loss": 0.8244, + "step": 6700 + }, + { + "epoch": 0.9727809793048457, + "grad_norm": 2.78125, + "learning_rate": 1.0415714285714286e-06, + "loss": 0.8159, + "step": 6710 + }, + { + "epoch": 0.9742307274111123, + "grad_norm": 2.5, + "learning_rate": 1.040142857142857e-06, + "loss": 0.7883, + "step": 6720 + }, + { + "epoch": 0.9756804755173789, + "grad_norm": 2.671875, + "learning_rate": 1.0387142857142857e-06, + "loss": 0.819, + "step": 6730 + }, + { + "epoch": 0.9771302236236454, + "grad_norm": 2.359375, + "learning_rate": 1.0372857142857142e-06, + "loss": 0.7793, + "step": 6740 + }, + { + "epoch": 0.978579971729912, + "grad_norm": 2.65625, + "learning_rate": 1.0358571428571428e-06, + "loss": 0.788, + "step": 6750 + }, + { + "epoch": 0.9800297198361785, + "grad_norm": 2.78125, + "learning_rate": 1.0344285714285713e-06, + "loss": 0.8251, + "step": 6760 + }, + { + "epoch": 0.981479467942445, + "grad_norm": 3.359375, + "learning_rate": 1.0329999999999999e-06, + "loss": 0.785, + "step": 6770 + }, + { + "epoch": 0.9829292160487115, + "grad_norm": 2.453125, + "learning_rate": 1.0315714285714284e-06, + "loss": 0.8662, + "step": 6780 + }, + { + "epoch": 0.984378964154978, + "grad_norm": 2.515625, + "learning_rate": 1.0301428571428572e-06, + "loss": 0.8003, + "step": 6790 + }, + { + "epoch": 0.9858287122612446, + "grad_norm": 2.734375, + "learning_rate": 1.0287142857142856e-06, + "loss": 0.8702, + "step": 6800 + }, + { + "epoch": 0.9872784603675111, + "grad_norm": 2.53125, + "learning_rate": 1.0272857142857143e-06, + "loss": 0.7776, + "step": 6810 + }, + { + "epoch": 0.9887282084737777, + "grad_norm": 2.71875, + "learning_rate": 1.0258571428571427e-06, + "loss": 0.8028, + "step": 6820 + }, + { + "epoch": 0.9901779565800443, + "grad_norm": 2.421875, + "learning_rate": 1.0244285714285714e-06, + "loss": 0.8109, + "step": 6830 + }, + { + "epoch": 0.9916277046863108, + "grad_norm": 2.375, + "learning_rate": 1.0229999999999998e-06, + "loss": 0.8002, + "step": 6840 + }, + { + "epoch": 0.9930774527925773, + "grad_norm": 2.828125, + "learning_rate": 1.0215714285714285e-06, + "loss": 0.848, + "step": 6850 + }, + { + "epoch": 0.9945272008988438, + "grad_norm": 2.71875, + "learning_rate": 1.020142857142857e-06, + "loss": 0.8068, + "step": 6860 + }, + { + "epoch": 0.9959769490051104, + "grad_norm": 3.40625, + "learning_rate": 1.0187142857142857e-06, + "loss": 0.8292, + "step": 6870 + }, + { + "epoch": 0.9974266971113769, + "grad_norm": 2.375, + "learning_rate": 1.0172857142857142e-06, + "loss": 0.8513, + "step": 6880 + }, + { + "epoch": 0.9988764452176434, + "grad_norm": 2.953125, + "learning_rate": 1.0158571428571428e-06, + "loss": 0.7935, + "step": 6890 + }, + { + "epoch": 1.0002899496212534, + "grad_norm": 2.8125, + "learning_rate": 1.0144285714285713e-06, + "loss": 0.8169, + "step": 6900 + }, + { + "epoch": 1.00173969772752, + "grad_norm": 2.65625, + "learning_rate": 1.0129999999999999e-06, + "loss": 0.8039, + "step": 6910 + }, + { + "epoch": 1.0031894458337864, + "grad_norm": 3.25, + "learning_rate": 1.0115714285714284e-06, + "loss": 0.8597, + "step": 6920 + }, + { + "epoch": 1.004639193940053, + "grad_norm": 2.546875, + "learning_rate": 1.0101428571428572e-06, + "loss": 0.8097, + "step": 6930 + }, + { + "epoch": 1.0060889420463195, + "grad_norm": 3.328125, + "learning_rate": 1.0087142857142855e-06, + "loss": 0.8027, + "step": 6940 + }, + { + "epoch": 1.007538690152586, + "grad_norm": 2.578125, + "learning_rate": 1.0072857142857143e-06, + "loss": 0.7875, + "step": 6950 + }, + { + "epoch": 1.0089884382588525, + "grad_norm": 2.453125, + "learning_rate": 1.0058571428571426e-06, + "loss": 0.8417, + "step": 6960 + }, + { + "epoch": 1.010438186365119, + "grad_norm": 2.828125, + "learning_rate": 1.0044285714285714e-06, + "loss": 0.8548, + "step": 6970 + }, + { + "epoch": 1.0118879344713856, + "grad_norm": 2.34375, + "learning_rate": 1.0029999999999998e-06, + "loss": 0.7403, + "step": 6980 + }, + { + "epoch": 1.013337682577652, + "grad_norm": 2.625, + "learning_rate": 1.0015714285714285e-06, + "loss": 0.8183, + "step": 6990 + }, + { + "epoch": 1.0147874306839186, + "grad_norm": 3.8125, + "learning_rate": 1.000142857142857e-06, + "loss": 0.8604, + "step": 7000 + }, + { + "epoch": 1.0162371787901852, + "grad_norm": 2.734375, + "learning_rate": 9.987142857142856e-07, + "loss": 0.8375, + "step": 7010 + }, + { + "epoch": 1.0176869268964517, + "grad_norm": 2.9375, + "learning_rate": 9.972857142857142e-07, + "loss": 0.8061, + "step": 7020 + }, + { + "epoch": 1.0191366750027182, + "grad_norm": 2.25, + "learning_rate": 9.958571428571427e-07, + "loss": 0.7926, + "step": 7030 + }, + { + "epoch": 1.0205864231089847, + "grad_norm": 2.5625, + "learning_rate": 9.944285714285713e-07, + "loss": 0.7749, + "step": 7040 + }, + { + "epoch": 1.0220361712152513, + "grad_norm": 2.265625, + "learning_rate": 9.929999999999999e-07, + "loss": 0.7936, + "step": 7050 + }, + { + "epoch": 1.0234859193215178, + "grad_norm": 2.71875, + "learning_rate": 9.915714285714284e-07, + "loss": 0.787, + "step": 7060 + }, + { + "epoch": 1.0249356674277845, + "grad_norm": 2.5, + "learning_rate": 9.90142857142857e-07, + "loss": 0.8331, + "step": 7070 + }, + { + "epoch": 1.026385415534051, + "grad_norm": 2.75, + "learning_rate": 9.887142857142857e-07, + "loss": 0.8546, + "step": 7080 + }, + { + "epoch": 1.0278351636403176, + "grad_norm": 2.546875, + "learning_rate": 9.872857142857143e-07, + "loss": 0.8231, + "step": 7090 + }, + { + "epoch": 1.029284911746584, + "grad_norm": 2.703125, + "learning_rate": 9.858571428571428e-07, + "loss": 0.7965, + "step": 7100 + }, + { + "epoch": 1.0307346598528506, + "grad_norm": 2.75, + "learning_rate": 9.844285714285714e-07, + "loss": 0.8174, + "step": 7110 + }, + { + "epoch": 1.0321844079591171, + "grad_norm": 2.6875, + "learning_rate": 9.83e-07, + "loss": 0.8387, + "step": 7120 + }, + { + "epoch": 1.0336341560653837, + "grad_norm": 3.078125, + "learning_rate": 9.815714285714285e-07, + "loss": 0.8345, + "step": 7130 + }, + { + "epoch": 1.0350839041716502, + "grad_norm": 2.59375, + "learning_rate": 9.80142857142857e-07, + "loss": 0.7387, + "step": 7140 + }, + { + "epoch": 1.0365336522779167, + "grad_norm": 2.640625, + "learning_rate": 9.787142857142856e-07, + "loss": 0.7833, + "step": 7150 + }, + { + "epoch": 1.0379834003841832, + "grad_norm": 2.671875, + "learning_rate": 9.772857142857142e-07, + "loss": 0.7895, + "step": 7160 + }, + { + "epoch": 1.0394331484904498, + "grad_norm": 2.234375, + "learning_rate": 9.758571428571427e-07, + "loss": 0.7966, + "step": 7170 + }, + { + "epoch": 1.0408828965967163, + "grad_norm": 2.84375, + "learning_rate": 9.744285714285713e-07, + "loss": 0.8072, + "step": 7180 + }, + { + "epoch": 1.0423326447029828, + "grad_norm": 2.515625, + "learning_rate": 9.729999999999998e-07, + "loss": 0.775, + "step": 7190 + }, + { + "epoch": 1.0437823928092493, + "grad_norm": 2.828125, + "learning_rate": 9.715714285714284e-07, + "loss": 0.816, + "step": 7200 + }, + { + "epoch": 1.0452321409155159, + "grad_norm": 2.890625, + "learning_rate": 9.701428571428572e-07, + "loss": 0.8463, + "step": 7210 + }, + { + "epoch": 1.0466818890217824, + "grad_norm": 2.578125, + "learning_rate": 9.687142857142857e-07, + "loss": 0.8073, + "step": 7220 + }, + { + "epoch": 1.048131637128049, + "grad_norm": 2.140625, + "learning_rate": 9.672857142857143e-07, + "loss": 0.8127, + "step": 7230 + }, + { + "epoch": 1.0495813852343154, + "grad_norm": 2.6875, + "learning_rate": 9.658571428571428e-07, + "loss": 0.7778, + "step": 7240 + }, + { + "epoch": 1.0510311333405822, + "grad_norm": 2.71875, + "learning_rate": 9.644285714285714e-07, + "loss": 0.7895, + "step": 7250 + }, + { + "epoch": 1.0524808814468487, + "grad_norm": 2.5625, + "learning_rate": 9.63e-07, + "loss": 0.8125, + "step": 7260 + }, + { + "epoch": 1.0539306295531152, + "grad_norm": 2.625, + "learning_rate": 9.615714285714285e-07, + "loss": 0.8123, + "step": 7270 + }, + { + "epoch": 1.0553803776593818, + "grad_norm": 2.609375, + "learning_rate": 9.60142857142857e-07, + "loss": 0.8417, + "step": 7280 + }, + { + "epoch": 1.0568301257656483, + "grad_norm": 2.453125, + "learning_rate": 9.587142857142856e-07, + "loss": 0.8222, + "step": 7290 + }, + { + "epoch": 1.0582798738719148, + "grad_norm": 2.34375, + "learning_rate": 9.572857142857141e-07, + "loss": 0.85, + "step": 7300 + }, + { + "epoch": 1.0597296219781813, + "grad_norm": 2.6875, + "learning_rate": 9.558571428571427e-07, + "loss": 0.839, + "step": 7310 + }, + { + "epoch": 1.0611793700844478, + "grad_norm": 2.21875, + "learning_rate": 9.544285714285713e-07, + "loss": 0.8236, + "step": 7320 + }, + { + "epoch": 1.0626291181907144, + "grad_norm": 2.390625, + "learning_rate": 9.529999999999999e-07, + "loss": 0.7733, + "step": 7330 + }, + { + "epoch": 1.064078866296981, + "grad_norm": 2.421875, + "learning_rate": 9.515714285714286e-07, + "loss": 0.8245, + "step": 7340 + }, + { + "epoch": 1.0655286144032474, + "grad_norm": 2.5, + "learning_rate": 9.501428571428571e-07, + "loss": 0.8654, + "step": 7350 + }, + { + "epoch": 1.066978362509514, + "grad_norm": 2.6875, + "learning_rate": 9.487142857142857e-07, + "loss": 0.8297, + "step": 7360 + }, + { + "epoch": 1.0684281106157805, + "grad_norm": 2.890625, + "learning_rate": 9.472857142857142e-07, + "loss": 0.8302, + "step": 7370 + }, + { + "epoch": 1.069877858722047, + "grad_norm": 3.890625, + "learning_rate": 9.458571428571428e-07, + "loss": 0.8513, + "step": 7380 + }, + { + "epoch": 1.0713276068283135, + "grad_norm": 2.5625, + "learning_rate": 9.444285714285714e-07, + "loss": 0.8206, + "step": 7390 + }, + { + "epoch": 1.07277735493458, + "grad_norm": 2.515625, + "learning_rate": 9.429999999999999e-07, + "loss": 0.806, + "step": 7400 + }, + { + "epoch": 1.0742271030408466, + "grad_norm": 2.75, + "learning_rate": 9.415714285714286e-07, + "loss": 0.7656, + "step": 7410 + }, + { + "epoch": 1.075676851147113, + "grad_norm": 2.46875, + "learning_rate": 9.401428571428571e-07, + "loss": 0.8391, + "step": 7420 + }, + { + "epoch": 1.0771265992533796, + "grad_norm": 2.546875, + "learning_rate": 9.387142857142857e-07, + "loss": 0.8278, + "step": 7430 + }, + { + "epoch": 1.0785763473596464, + "grad_norm": 3.375, + "learning_rate": 9.372857142857142e-07, + "loss": 0.8012, + "step": 7440 + }, + { + "epoch": 1.0800260954659129, + "grad_norm": 2.9375, + "learning_rate": 9.358571428571428e-07, + "loss": 0.8127, + "step": 7450 + }, + { + "epoch": 1.0814758435721794, + "grad_norm": 2.1875, + "learning_rate": 9.344285714285713e-07, + "loss": 0.7833, + "step": 7460 + }, + { + "epoch": 1.082925591678446, + "grad_norm": 3.296875, + "learning_rate": 9.33e-07, + "loss": 0.8328, + "step": 7470 + }, + { + "epoch": 1.0843753397847125, + "grad_norm": 2.625, + "learning_rate": 9.315714285714286e-07, + "loss": 0.801, + "step": 7480 + }, + { + "epoch": 1.085825087890979, + "grad_norm": 2.34375, + "learning_rate": 9.301428571428571e-07, + "loss": 0.8318, + "step": 7490 + }, + { + "epoch": 1.0872748359972455, + "grad_norm": 2.9375, + "learning_rate": 9.287142857142857e-07, + "loss": 0.8233, + "step": 7500 + }, + { + "epoch": 1.088724584103512, + "grad_norm": 2.84375, + "learning_rate": 9.272857142857142e-07, + "loss": 0.7974, + "step": 7510 + }, + { + "epoch": 1.0901743322097786, + "grad_norm": 2.8125, + "learning_rate": 9.258571428571428e-07, + "loss": 0.8123, + "step": 7520 + }, + { + "epoch": 1.091624080316045, + "grad_norm": 2.828125, + "learning_rate": 9.244285714285713e-07, + "loss": 0.8341, + "step": 7530 + }, + { + "epoch": 1.0930738284223116, + "grad_norm": 3.125, + "learning_rate": 9.23e-07, + "loss": 0.8099, + "step": 7540 + }, + { + "epoch": 1.0945235765285781, + "grad_norm": 2.765625, + "learning_rate": 9.215714285714285e-07, + "loss": 0.8577, + "step": 7550 + }, + { + "epoch": 1.0959733246348446, + "grad_norm": 2.625, + "learning_rate": 9.201428571428571e-07, + "loss": 0.8267, + "step": 7560 + }, + { + "epoch": 1.0974230727411112, + "grad_norm": 2.546875, + "learning_rate": 9.187142857142857e-07, + "loss": 0.8578, + "step": 7570 + }, + { + "epoch": 1.0988728208473777, + "grad_norm": 2.265625, + "learning_rate": 9.172857142857142e-07, + "loss": 0.7729, + "step": 7580 + }, + { + "epoch": 1.1003225689536442, + "grad_norm": 2.890625, + "learning_rate": 9.158571428571428e-07, + "loss": 0.8445, + "step": 7590 + }, + { + "epoch": 1.1017723170599107, + "grad_norm": 2.9375, + "learning_rate": 9.144285714285714e-07, + "loss": 0.8264, + "step": 7600 + }, + { + "epoch": 1.1032220651661775, + "grad_norm": 2.9375, + "learning_rate": 9.13e-07, + "loss": 0.8362, + "step": 7610 + }, + { + "epoch": 1.104671813272444, + "grad_norm": 2.421875, + "learning_rate": 9.115714285714285e-07, + "loss": 0.8318, + "step": 7620 + }, + { + "epoch": 1.1061215613787105, + "grad_norm": 3.140625, + "learning_rate": 9.101428571428571e-07, + "loss": 0.7697, + "step": 7630 + }, + { + "epoch": 1.107571309484977, + "grad_norm": 3.328125, + "learning_rate": 9.087142857142856e-07, + "loss": 0.8144, + "step": 7640 + }, + { + "epoch": 1.1090210575912436, + "grad_norm": 2.25, + "learning_rate": 9.072857142857142e-07, + "loss": 0.786, + "step": 7650 + }, + { + "epoch": 1.11047080569751, + "grad_norm": 2.546875, + "learning_rate": 9.058571428571428e-07, + "loss": 0.8382, + "step": 7660 + }, + { + "epoch": 1.1119205538037766, + "grad_norm": 2.671875, + "learning_rate": 9.044285714285714e-07, + "loss": 0.8317, + "step": 7670 + }, + { + "epoch": 1.1133703019100432, + "grad_norm": 2.578125, + "learning_rate": 9.03e-07, + "loss": 0.7785, + "step": 7680 + }, + { + "epoch": 1.1148200500163097, + "grad_norm": 2.890625, + "learning_rate": 9.015714285714285e-07, + "loss": 0.84, + "step": 7690 + }, + { + "epoch": 1.1162697981225762, + "grad_norm": 2.359375, + "learning_rate": 9.001428571428571e-07, + "loss": 0.7565, + "step": 7700 + }, + { + "epoch": 1.1177195462288427, + "grad_norm": 2.625, + "learning_rate": 8.987142857142856e-07, + "loss": 0.777, + "step": 7710 + }, + { + "epoch": 1.1191692943351093, + "grad_norm": 3.078125, + "learning_rate": 8.972857142857142e-07, + "loss": 0.8414, + "step": 7720 + }, + { + "epoch": 1.1206190424413758, + "grad_norm": 2.734375, + "learning_rate": 8.958571428571429e-07, + "loss": 0.8034, + "step": 7730 + }, + { + "epoch": 1.1220687905476423, + "grad_norm": 2.734375, + "learning_rate": 8.944285714285714e-07, + "loss": 0.8034, + "step": 7740 + }, + { + "epoch": 1.1235185386539088, + "grad_norm": 2.46875, + "learning_rate": 8.93e-07, + "loss": 0.7982, + "step": 7750 + }, + { + "epoch": 1.1249682867601754, + "grad_norm": 2.734375, + "learning_rate": 8.915714285714285e-07, + "loss": 0.8353, + "step": 7760 + }, + { + "epoch": 1.1264180348664419, + "grad_norm": 2.125, + "learning_rate": 8.901428571428571e-07, + "loss": 0.8218, + "step": 7770 + }, + { + "epoch": 1.1278677829727084, + "grad_norm": 2.90625, + "learning_rate": 8.887142857142856e-07, + "loss": 0.8566, + "step": 7780 + }, + { + "epoch": 1.129317531078975, + "grad_norm": 2.578125, + "learning_rate": 8.872857142857142e-07, + "loss": 0.7618, + "step": 7790 + }, + { + "epoch": 1.1307672791852417, + "grad_norm": 2.890625, + "learning_rate": 8.858571428571428e-07, + "loss": 0.81, + "step": 7800 + }, + { + "epoch": 1.1322170272915082, + "grad_norm": 2.765625, + "learning_rate": 8.844285714285714e-07, + "loss": 0.776, + "step": 7810 + }, + { + "epoch": 1.1336667753977747, + "grad_norm": 2.546875, + "learning_rate": 8.83e-07, + "loss": 0.8359, + "step": 7820 + }, + { + "epoch": 1.1351165235040412, + "grad_norm": 2.59375, + "learning_rate": 8.815714285714285e-07, + "loss": 0.8449, + "step": 7830 + }, + { + "epoch": 1.1365662716103078, + "grad_norm": 2.46875, + "learning_rate": 8.801428571428571e-07, + "loss": 0.7953, + "step": 7840 + }, + { + "epoch": 1.1380160197165743, + "grad_norm": 2.546875, + "learning_rate": 8.787142857142856e-07, + "loss": 0.8203, + "step": 7850 + }, + { + "epoch": 1.1394657678228408, + "grad_norm": 2.484375, + "learning_rate": 8.772857142857143e-07, + "loss": 0.8192, + "step": 7860 + }, + { + "epoch": 1.1409155159291073, + "grad_norm": 2.59375, + "learning_rate": 8.758571428571428e-07, + "loss": 0.8367, + "step": 7870 + }, + { + "epoch": 1.1423652640353739, + "grad_norm": 2.875, + "learning_rate": 8.744285714285714e-07, + "loss": 0.7605, + "step": 7880 + }, + { + "epoch": 1.1438150121416404, + "grad_norm": 2.71875, + "learning_rate": 8.729999999999999e-07, + "loss": 0.7939, + "step": 7890 + }, + { + "epoch": 1.145264760247907, + "grad_norm": 2.765625, + "learning_rate": 8.715714285714285e-07, + "loss": 0.767, + "step": 7900 + }, + { + "epoch": 1.1467145083541734, + "grad_norm": 2.75, + "learning_rate": 8.70142857142857e-07, + "loss": 0.7972, + "step": 7910 + }, + { + "epoch": 1.14816425646044, + "grad_norm": 2.71875, + "learning_rate": 8.687142857142856e-07, + "loss": 0.8528, + "step": 7920 + }, + { + "epoch": 1.1496140045667065, + "grad_norm": 3.125, + "learning_rate": 8.672857142857143e-07, + "loss": 0.8309, + "step": 7930 + }, + { + "epoch": 1.151063752672973, + "grad_norm": 2.328125, + "learning_rate": 8.658571428571428e-07, + "loss": 0.8447, + "step": 7940 + }, + { + "epoch": 1.1525135007792395, + "grad_norm": 2.53125, + "learning_rate": 8.644285714285714e-07, + "loss": 0.7869, + "step": 7950 + }, + { + "epoch": 1.153963248885506, + "grad_norm": 2.3125, + "learning_rate": 8.629999999999999e-07, + "loss": 0.7844, + "step": 7960 + }, + { + "epoch": 1.1554129969917728, + "grad_norm": 2.8125, + "learning_rate": 8.615714285714285e-07, + "loss": 0.8669, + "step": 7970 + }, + { + "epoch": 1.156862745098039, + "grad_norm": 2.546875, + "learning_rate": 8.60142857142857e-07, + "loss": 0.7851, + "step": 7980 + }, + { + "epoch": 1.1583124932043058, + "grad_norm": 3.34375, + "learning_rate": 8.587142857142857e-07, + "loss": 0.7851, + "step": 7990 + }, + { + "epoch": 1.1597622413105724, + "grad_norm": 3.0625, + "learning_rate": 8.572857142857143e-07, + "loss": 0.8579, + "step": 8000 + }, + { + "epoch": 1.161211989416839, + "grad_norm": 2.9375, + "learning_rate": 8.558571428571428e-07, + "loss": 0.8479, + "step": 8010 + }, + { + "epoch": 1.1626617375231054, + "grad_norm": 2.578125, + "learning_rate": 8.544285714285714e-07, + "loss": 0.7908, + "step": 8020 + }, + { + "epoch": 1.164111485629372, + "grad_norm": 2.53125, + "learning_rate": 8.529999999999999e-07, + "loss": 0.8326, + "step": 8030 + }, + { + "epoch": 1.1655612337356385, + "grad_norm": 2.46875, + "learning_rate": 8.515714285714285e-07, + "loss": 0.8238, + "step": 8040 + }, + { + "epoch": 1.167010981841905, + "grad_norm": 2.765625, + "learning_rate": 8.50142857142857e-07, + "loss": 0.7843, + "step": 8050 + }, + { + "epoch": 1.1684607299481715, + "grad_norm": 2.625, + "learning_rate": 8.487142857142857e-07, + "loss": 0.8403, + "step": 8060 + }, + { + "epoch": 1.169910478054438, + "grad_norm": 2.296875, + "learning_rate": 8.472857142857142e-07, + "loss": 0.8133, + "step": 8070 + }, + { + "epoch": 1.1713602261607046, + "grad_norm": 2.703125, + "learning_rate": 8.458571428571428e-07, + "loss": 0.7944, + "step": 8080 + }, + { + "epoch": 1.172809974266971, + "grad_norm": 3.40625, + "learning_rate": 8.444285714285714e-07, + "loss": 0.8315, + "step": 8090 + }, + { + "epoch": 1.1742597223732376, + "grad_norm": 2.40625, + "learning_rate": 8.429999999999999e-07, + "loss": 0.7871, + "step": 8100 + }, + { + "epoch": 1.1757094704795041, + "grad_norm": 3.015625, + "learning_rate": 8.415714285714285e-07, + "loss": 0.875, + "step": 8110 + }, + { + "epoch": 1.1771592185857707, + "grad_norm": 3.3125, + "learning_rate": 8.401428571428571e-07, + "loss": 0.8113, + "step": 8120 + }, + { + "epoch": 1.1786089666920372, + "grad_norm": 2.625, + "learning_rate": 8.387142857142857e-07, + "loss": 0.755, + "step": 8130 + }, + { + "epoch": 1.1800587147983037, + "grad_norm": 2.296875, + "learning_rate": 8.372857142857142e-07, + "loss": 0.771, + "step": 8140 + }, + { + "epoch": 1.1815084629045702, + "grad_norm": 2.78125, + "learning_rate": 8.358571428571428e-07, + "loss": 0.8604, + "step": 8150 + }, + { + "epoch": 1.182958211010837, + "grad_norm": 2.734375, + "learning_rate": 8.344285714285713e-07, + "loss": 0.8214, + "step": 8160 + }, + { + "epoch": 1.1844079591171033, + "grad_norm": 2.765625, + "learning_rate": 8.329999999999999e-07, + "loss": 0.8521, + "step": 8170 + }, + { + "epoch": 1.18585770722337, + "grad_norm": 2.703125, + "learning_rate": 8.315714285714285e-07, + "loss": 0.8362, + "step": 8180 + }, + { + "epoch": 1.1873074553296366, + "grad_norm": 2.90625, + "learning_rate": 8.301428571428571e-07, + "loss": 0.7912, + "step": 8190 + }, + { + "epoch": 1.188757203435903, + "grad_norm": 2.5625, + "learning_rate": 8.287142857142857e-07, + "loss": 0.869, + "step": 8200 + }, + { + "epoch": 1.1902069515421696, + "grad_norm": 2.5, + "learning_rate": 8.272857142857142e-07, + "loss": 0.8159, + "step": 8210 + }, + { + "epoch": 1.1916566996484361, + "grad_norm": 2.40625, + "learning_rate": 8.258571428571428e-07, + "loss": 0.8613, + "step": 8220 + }, + { + "epoch": 1.1931064477547026, + "grad_norm": 2.484375, + "learning_rate": 8.244285714285713e-07, + "loss": 0.799, + "step": 8230 + }, + { + "epoch": 1.1945561958609692, + "grad_norm": 2.515625, + "learning_rate": 8.229999999999999e-07, + "loss": 0.7832, + "step": 8240 + }, + { + "epoch": 1.1960059439672357, + "grad_norm": 2.796875, + "learning_rate": 8.215714285714286e-07, + "loss": 0.8954, + "step": 8250 + }, + { + "epoch": 1.1974556920735022, + "grad_norm": 3.09375, + "learning_rate": 8.201428571428571e-07, + "loss": 0.8415, + "step": 8260 + }, + { + "epoch": 1.1989054401797687, + "grad_norm": 2.796875, + "learning_rate": 8.187142857142857e-07, + "loss": 0.8306, + "step": 8270 + }, + { + "epoch": 1.2003551882860353, + "grad_norm": 2.71875, + "learning_rate": 8.172857142857142e-07, + "loss": 0.8596, + "step": 8280 + }, + { + "epoch": 1.2018049363923018, + "grad_norm": 3.5, + "learning_rate": 8.158571428571428e-07, + "loss": 0.8486, + "step": 8290 + }, + { + "epoch": 1.2032546844985683, + "grad_norm": 2.328125, + "learning_rate": 8.144285714285713e-07, + "loss": 0.8356, + "step": 8300 + }, + { + "epoch": 1.2047044326048348, + "grad_norm": 2.484375, + "learning_rate": 8.129999999999999e-07, + "loss": 0.8169, + "step": 8310 + }, + { + "epoch": 1.2061541807111014, + "grad_norm": 2.703125, + "learning_rate": 8.115714285714285e-07, + "loss": 0.7965, + "step": 8320 + }, + { + "epoch": 1.2076039288173679, + "grad_norm": 2.921875, + "learning_rate": 8.101428571428571e-07, + "loss": 0.8163, + "step": 8330 + }, + { + "epoch": 1.2090536769236344, + "grad_norm": 2.75, + "learning_rate": 8.087142857142857e-07, + "loss": 0.7917, + "step": 8340 + }, + { + "epoch": 1.2105034250299012, + "grad_norm": 2.65625, + "learning_rate": 8.072857142857142e-07, + "loss": 0.8344, + "step": 8350 + }, + { + "epoch": 1.2119531731361677, + "grad_norm": 2.640625, + "learning_rate": 8.058571428571428e-07, + "loss": 0.8486, + "step": 8360 + }, + { + "epoch": 1.2134029212424342, + "grad_norm": 2.625, + "learning_rate": 8.044285714285713e-07, + "loss": 0.8488, + "step": 8370 + }, + { + "epoch": 1.2148526693487007, + "grad_norm": 3.1875, + "learning_rate": 8.03e-07, + "loss": 0.7769, + "step": 8380 + }, + { + "epoch": 1.2163024174549673, + "grad_norm": 2.46875, + "learning_rate": 8.015714285714285e-07, + "loss": 0.8475, + "step": 8390 + }, + { + "epoch": 1.2177521655612338, + "grad_norm": 2.390625, + "learning_rate": 8.001428571428571e-07, + "loss": 0.8258, + "step": 8400 + }, + { + "epoch": 1.2192019136675003, + "grad_norm": 2.34375, + "learning_rate": 7.987142857142856e-07, + "loss": 0.8467, + "step": 8410 + }, + { + "epoch": 1.2206516617737668, + "grad_norm": 2.90625, + "learning_rate": 7.972857142857142e-07, + "loss": 0.788, + "step": 8420 + }, + { + "epoch": 1.2221014098800334, + "grad_norm": 2.515625, + "learning_rate": 7.958571428571428e-07, + "loss": 0.8408, + "step": 8430 + }, + { + "epoch": 1.2235511579862999, + "grad_norm": 2.328125, + "learning_rate": 7.944285714285714e-07, + "loss": 0.7919, + "step": 8440 + }, + { + "epoch": 1.2250009060925664, + "grad_norm": 3.125, + "learning_rate": 7.93e-07, + "loss": 0.7887, + "step": 8450 + }, + { + "epoch": 1.226450654198833, + "grad_norm": 2.921875, + "learning_rate": 7.915714285714285e-07, + "loss": 0.7824, + "step": 8460 + }, + { + "epoch": 1.2279004023050994, + "grad_norm": 2.84375, + "learning_rate": 7.901428571428571e-07, + "loss": 0.8372, + "step": 8470 + }, + { + "epoch": 1.229350150411366, + "grad_norm": 2.328125, + "learning_rate": 7.887142857142856e-07, + "loss": 0.809, + "step": 8480 + }, + { + "epoch": 1.2307998985176325, + "grad_norm": 3.046875, + "learning_rate": 7.872857142857142e-07, + "loss": 0.7948, + "step": 8490 + }, + { + "epoch": 1.232249646623899, + "grad_norm": 2.640625, + "learning_rate": 7.858571428571427e-07, + "loss": 0.8212, + "step": 8500 + }, + { + "epoch": 1.2336993947301655, + "grad_norm": 2.84375, + "learning_rate": 7.844285714285714e-07, + "loss": 0.8151, + "step": 8510 + }, + { + "epoch": 1.2351491428364323, + "grad_norm": 2.46875, + "learning_rate": 7.83e-07, + "loss": 0.8008, + "step": 8520 + }, + { + "epoch": 1.2365988909426986, + "grad_norm": 2.890625, + "learning_rate": 7.815714285714285e-07, + "loss": 0.8103, + "step": 8530 + }, + { + "epoch": 1.2380486390489653, + "grad_norm": 2.328125, + "learning_rate": 7.801428571428571e-07, + "loss": 0.8686, + "step": 8540 + }, + { + "epoch": 1.2394983871552319, + "grad_norm": 2.734375, + "learning_rate": 7.787142857142856e-07, + "loss": 0.8127, + "step": 8550 + }, + { + "epoch": 1.2409481352614984, + "grad_norm": 2.171875, + "learning_rate": 7.772857142857142e-07, + "loss": 0.8065, + "step": 8560 + }, + { + "epoch": 1.242397883367765, + "grad_norm": 2.546875, + "learning_rate": 7.758571428571428e-07, + "loss": 0.776, + "step": 8570 + }, + { + "epoch": 1.2438476314740314, + "grad_norm": 3.171875, + "learning_rate": 7.744285714285714e-07, + "loss": 0.7866, + "step": 8580 + }, + { + "epoch": 1.245297379580298, + "grad_norm": 3.15625, + "learning_rate": 7.729999999999999e-07, + "loss": 0.8777, + "step": 8590 + }, + { + "epoch": 1.2467471276865645, + "grad_norm": 2.59375, + "learning_rate": 7.715714285714285e-07, + "loss": 0.7489, + "step": 8600 + }, + { + "epoch": 1.248196875792831, + "grad_norm": 2.46875, + "learning_rate": 7.701428571428571e-07, + "loss": 0.7945, + "step": 8610 + }, + { + "epoch": 1.2496466238990975, + "grad_norm": 2.875, + "learning_rate": 7.687142857142856e-07, + "loss": 0.7861, + "step": 8620 + }, + { + "epoch": 1.251096372005364, + "grad_norm": 2.5, + "learning_rate": 7.672857142857142e-07, + "loss": 0.7944, + "step": 8630 + }, + { + "epoch": 1.2525461201116306, + "grad_norm": 2.328125, + "learning_rate": 7.658571428571428e-07, + "loss": 0.8378, + "step": 8640 + }, + { + "epoch": 1.253995868217897, + "grad_norm": 3.484375, + "learning_rate": 7.644285714285714e-07, + "loss": 0.8523, + "step": 8650 + }, + { + "epoch": 1.2554456163241636, + "grad_norm": 2.265625, + "learning_rate": 7.629999999999999e-07, + "loss": 0.7892, + "step": 8660 + }, + { + "epoch": 1.2568953644304302, + "grad_norm": 2.71875, + "learning_rate": 7.615714285714285e-07, + "loss": 0.8642, + "step": 8670 + }, + { + "epoch": 1.2583451125366967, + "grad_norm": 2.671875, + "learning_rate": 7.60142857142857e-07, + "loss": 0.8872, + "step": 8680 + }, + { + "epoch": 1.2597948606429634, + "grad_norm": 3.140625, + "learning_rate": 7.587142857142856e-07, + "loss": 0.851, + "step": 8690 + }, + { + "epoch": 1.2612446087492297, + "grad_norm": 2.546875, + "learning_rate": 7.572857142857143e-07, + "loss": 0.7818, + "step": 8700 + }, + { + "epoch": 1.2626943568554965, + "grad_norm": 2.578125, + "learning_rate": 7.558571428571428e-07, + "loss": 0.7804, + "step": 8710 + }, + { + "epoch": 1.2641441049617628, + "grad_norm": 2.671875, + "learning_rate": 7.544285714285714e-07, + "loss": 0.7989, + "step": 8720 + }, + { + "epoch": 1.2655938530680295, + "grad_norm": 2.828125, + "learning_rate": 7.529999999999999e-07, + "loss": 0.8602, + "step": 8730 + }, + { + "epoch": 1.267043601174296, + "grad_norm": 2.90625, + "learning_rate": 7.515714285714285e-07, + "loss": 0.801, + "step": 8740 + }, + { + "epoch": 1.2684933492805626, + "grad_norm": 2.6875, + "learning_rate": 7.50142857142857e-07, + "loss": 0.869, + "step": 8750 + }, + { + "epoch": 1.269943097386829, + "grad_norm": 2.609375, + "learning_rate": 7.487142857142856e-07, + "loss": 0.7826, + "step": 8760 + }, + { + "epoch": 1.2713928454930956, + "grad_norm": 2.609375, + "learning_rate": 7.472857142857143e-07, + "loss": 0.8348, + "step": 8770 + }, + { + "epoch": 1.2728425935993621, + "grad_norm": 2.53125, + "learning_rate": 7.458571428571428e-07, + "loss": 0.8554, + "step": 8780 + }, + { + "epoch": 1.2742923417056287, + "grad_norm": 2.515625, + "learning_rate": 7.444285714285714e-07, + "loss": 0.8375, + "step": 8790 + }, + { + "epoch": 1.2757420898118952, + "grad_norm": 2.4375, + "learning_rate": 7.429999999999999e-07, + "loss": 0.8277, + "step": 8800 + }, + { + "epoch": 1.2771918379181617, + "grad_norm": 2.53125, + "learning_rate": 7.415714285714285e-07, + "loss": 0.8808, + "step": 8810 + }, + { + "epoch": 1.2786415860244282, + "grad_norm": 2.375, + "learning_rate": 7.40142857142857e-07, + "loss": 0.8384, + "step": 8820 + }, + { + "epoch": 1.2800913341306948, + "grad_norm": 2.390625, + "learning_rate": 7.387142857142857e-07, + "loss": 0.8229, + "step": 8830 + }, + { + "epoch": 1.2815410822369613, + "grad_norm": 2.46875, + "learning_rate": 7.372857142857142e-07, + "loss": 0.7984, + "step": 8840 + }, + { + "epoch": 1.2829908303432278, + "grad_norm": 2.875, + "learning_rate": 7.358571428571428e-07, + "loss": 0.9042, + "step": 8850 + }, + { + "epoch": 1.2844405784494943, + "grad_norm": 2.515625, + "learning_rate": 7.344285714285714e-07, + "loss": 0.8243, + "step": 8860 + }, + { + "epoch": 1.2858903265557609, + "grad_norm": 2.328125, + "learning_rate": 7.329999999999999e-07, + "loss": 0.8595, + "step": 8870 + }, + { + "epoch": 1.2873400746620276, + "grad_norm": 2.609375, + "learning_rate": 7.315714285714285e-07, + "loss": 0.8284, + "step": 8880 + }, + { + "epoch": 1.288789822768294, + "grad_norm": 2.6875, + "learning_rate": 7.30142857142857e-07, + "loss": 0.8163, + "step": 8890 + }, + { + "epoch": 1.2902395708745606, + "grad_norm": 3.25, + "learning_rate": 7.287142857142857e-07, + "loss": 0.8181, + "step": 8900 + }, + { + "epoch": 1.291689318980827, + "grad_norm": 2.359375, + "learning_rate": 7.272857142857142e-07, + "loss": 0.7856, + "step": 8910 + }, + { + "epoch": 1.2931390670870937, + "grad_norm": 2.34375, + "learning_rate": 7.258571428571428e-07, + "loss": 0.8072, + "step": 8920 + }, + { + "epoch": 1.2945888151933602, + "grad_norm": 2.28125, + "learning_rate": 7.244285714285713e-07, + "loss": 0.8645, + "step": 8930 + }, + { + "epoch": 1.2960385632996267, + "grad_norm": 2.78125, + "learning_rate": 7.229999999999999e-07, + "loss": 0.791, + "step": 8940 + }, + { + "epoch": 1.2974883114058933, + "grad_norm": 2.21875, + "learning_rate": 7.215714285714285e-07, + "loss": 0.7609, + "step": 8950 + }, + { + "epoch": 1.2989380595121598, + "grad_norm": 3.015625, + "learning_rate": 7.201428571428571e-07, + "loss": 0.7667, + "step": 8960 + }, + { + "epoch": 1.3003878076184263, + "grad_norm": 2.390625, + "learning_rate": 7.187142857142857e-07, + "loss": 0.7787, + "step": 8970 + }, + { + "epoch": 1.3018375557246928, + "grad_norm": 3.515625, + "learning_rate": 7.172857142857142e-07, + "loss": 0.8065, + "step": 8980 + }, + { + "epoch": 1.3032873038309594, + "grad_norm": 2.609375, + "learning_rate": 7.158571428571428e-07, + "loss": 0.8168, + "step": 8990 + }, + { + "epoch": 1.304737051937226, + "grad_norm": 2.453125, + "learning_rate": 7.144285714285713e-07, + "loss": 0.798, + "step": 9000 + }, + { + "epoch": 1.3061868000434924, + "grad_norm": 3.09375, + "learning_rate": 7.129999999999999e-07, + "loss": 0.8575, + "step": 9010 + }, + { + "epoch": 1.307636548149759, + "grad_norm": 3.203125, + "learning_rate": 7.115714285714284e-07, + "loss": 0.8598, + "step": 9020 + }, + { + "epoch": 1.3090862962560255, + "grad_norm": 2.390625, + "learning_rate": 7.101428571428571e-07, + "loss": 0.8568, + "step": 9030 + }, + { + "epoch": 1.310536044362292, + "grad_norm": 3.0625, + "learning_rate": 7.087142857142857e-07, + "loss": 0.8302, + "step": 9040 + }, + { + "epoch": 1.3119857924685585, + "grad_norm": 2.359375, + "learning_rate": 7.072857142857142e-07, + "loss": 0.8502, + "step": 9050 + }, + { + "epoch": 1.313435540574825, + "grad_norm": 2.640625, + "learning_rate": 7.058571428571428e-07, + "loss": 0.8371, + "step": 9060 + }, + { + "epoch": 1.3148852886810918, + "grad_norm": 2.296875, + "learning_rate": 7.044285714285713e-07, + "loss": 0.7493, + "step": 9070 + }, + { + "epoch": 1.316335036787358, + "grad_norm": 2.90625, + "learning_rate": 7.029999999999999e-07, + "loss": 0.8249, + "step": 9080 + }, + { + "epoch": 1.3177847848936248, + "grad_norm": 3.0, + "learning_rate": 7.015714285714286e-07, + "loss": 0.8279, + "step": 9090 + }, + { + "epoch": 1.3192345329998914, + "grad_norm": 2.21875, + "learning_rate": 7.001428571428572e-07, + "loss": 0.8713, + "step": 9100 + }, + { + "epoch": 1.3206842811061579, + "grad_norm": 2.9375, + "learning_rate": 6.987142857142858e-07, + "loss": 0.8276, + "step": 9110 + }, + { + "epoch": 1.3221340292124244, + "grad_norm": 2.609375, + "learning_rate": 6.972857142857143e-07, + "loss": 0.8324, + "step": 9120 + }, + { + "epoch": 1.323583777318691, + "grad_norm": 2.5625, + "learning_rate": 6.958571428571429e-07, + "loss": 0.826, + "step": 9130 + }, + { + "epoch": 1.3250335254249574, + "grad_norm": 2.640625, + "learning_rate": 6.944285714285713e-07, + "loss": 0.8038, + "step": 9140 + }, + { + "epoch": 1.326483273531224, + "grad_norm": 2.421875, + "learning_rate": 6.929999999999999e-07, + "loss": 0.8039, + "step": 9150 + }, + { + "epoch": 1.3279330216374905, + "grad_norm": 2.5625, + "learning_rate": 6.915714285714286e-07, + "loss": 0.7518, + "step": 9160 + }, + { + "epoch": 1.329382769743757, + "grad_norm": 2.421875, + "learning_rate": 6.901428571428572e-07, + "loss": 0.7778, + "step": 9170 + }, + { + "epoch": 1.3308325178500235, + "grad_norm": 3.0, + "learning_rate": 6.887142857142857e-07, + "loss": 0.8179, + "step": 9180 + }, + { + "epoch": 1.33228226595629, + "grad_norm": 2.90625, + "learning_rate": 6.872857142857143e-07, + "loss": 0.7926, + "step": 9190 + }, + { + "epoch": 1.3337320140625566, + "grad_norm": 2.75, + "learning_rate": 6.858571428571429e-07, + "loss": 0.8338, + "step": 9200 + }, + { + "epoch": 1.3351817621688231, + "grad_norm": 2.625, + "learning_rate": 6.844285714285714e-07, + "loss": 0.8783, + "step": 9210 + }, + { + "epoch": 1.3366315102750896, + "grad_norm": 2.546875, + "learning_rate": 6.830000000000001e-07, + "loss": 0.8047, + "step": 9220 + }, + { + "epoch": 1.3380812583813562, + "grad_norm": 2.765625, + "learning_rate": 6.815714285714286e-07, + "loss": 0.8026, + "step": 9230 + }, + { + "epoch": 1.339531006487623, + "grad_norm": 2.46875, + "learning_rate": 6.801428571428572e-07, + "loss": 0.8361, + "step": 9240 + }, + { + "epoch": 1.3409807545938892, + "grad_norm": 2.859375, + "learning_rate": 6.787142857142857e-07, + "loss": 0.8153, + "step": 9250 + }, + { + "epoch": 1.342430502700156, + "grad_norm": 2.4375, + "learning_rate": 6.772857142857143e-07, + "loss": 0.8409, + "step": 9260 + }, + { + "epoch": 1.3438802508064223, + "grad_norm": 2.703125, + "learning_rate": 6.758571428571428e-07, + "loss": 0.7971, + "step": 9270 + }, + { + "epoch": 1.345329998912689, + "grad_norm": 2.703125, + "learning_rate": 6.744285714285714e-07, + "loss": 0.8543, + "step": 9280 + }, + { + "epoch": 1.3467797470189555, + "grad_norm": 2.234375, + "learning_rate": 6.730000000000001e-07, + "loss": 0.7822, + "step": 9290 + }, + { + "epoch": 1.348229495125222, + "grad_norm": 2.578125, + "learning_rate": 6.715714285714286e-07, + "loss": 0.7853, + "step": 9300 + }, + { + "epoch": 1.3496792432314886, + "grad_norm": 2.859375, + "learning_rate": 6.701428571428572e-07, + "loss": 0.7807, + "step": 9310 + }, + { + "epoch": 1.351128991337755, + "grad_norm": 2.21875, + "learning_rate": 6.687142857142857e-07, + "loss": 0.7968, + "step": 9320 + }, + { + "epoch": 1.3525787394440216, + "grad_norm": 3.453125, + "learning_rate": 6.672857142857143e-07, + "loss": 0.8087, + "step": 9330 + }, + { + "epoch": 1.3540284875502882, + "grad_norm": 2.90625, + "learning_rate": 6.658571428571428e-07, + "loss": 0.8257, + "step": 9340 + }, + { + "epoch": 1.3554782356565547, + "grad_norm": 2.53125, + "learning_rate": 6.644285714285715e-07, + "loss": 0.8315, + "step": 9350 + }, + { + "epoch": 1.3569279837628212, + "grad_norm": 3.046875, + "learning_rate": 6.63e-07, + "loss": 0.7963, + "step": 9360 + }, + { + "epoch": 1.3583777318690877, + "grad_norm": 2.453125, + "learning_rate": 6.615714285714286e-07, + "loss": 0.7861, + "step": 9370 + }, + { + "epoch": 1.3598274799753542, + "grad_norm": 3.375, + "learning_rate": 6.601428571428572e-07, + "loss": 0.8209, + "step": 9380 + }, + { + "epoch": 1.3612772280816208, + "grad_norm": 3.296875, + "learning_rate": 6.587142857142857e-07, + "loss": 0.8583, + "step": 9390 + }, + { + "epoch": 1.3627269761878873, + "grad_norm": 3.0625, + "learning_rate": 6.572857142857143e-07, + "loss": 0.8324, + "step": 9400 + }, + { + "epoch": 1.3641767242941538, + "grad_norm": 3.234375, + "learning_rate": 6.558571428571428e-07, + "loss": 0.897, + "step": 9410 + }, + { + "epoch": 1.3656264724004203, + "grad_norm": 2.9375, + "learning_rate": 6.544285714285715e-07, + "loss": 0.7979, + "step": 9420 + }, + { + "epoch": 1.367076220506687, + "grad_norm": 2.5, + "learning_rate": 6.53e-07, + "loss": 0.8206, + "step": 9430 + }, + { + "epoch": 1.3685259686129534, + "grad_norm": 2.65625, + "learning_rate": 6.515714285714286e-07, + "loss": 0.85, + "step": 9440 + }, + { + "epoch": 1.3699757167192201, + "grad_norm": 3.5, + "learning_rate": 6.501428571428571e-07, + "loss": 0.8944, + "step": 9450 + }, + { + "epoch": 1.3714254648254864, + "grad_norm": 2.875, + "learning_rate": 6.487142857142857e-07, + "loss": 0.8246, + "step": 9460 + }, + { + "epoch": 1.3728752129317532, + "grad_norm": 2.421875, + "learning_rate": 6.472857142857143e-07, + "loss": 0.8597, + "step": 9470 + }, + { + "epoch": 1.3743249610380197, + "grad_norm": 2.796875, + "learning_rate": 6.458571428571429e-07, + "loss": 0.8071, + "step": 9480 + }, + { + "epoch": 1.3757747091442862, + "grad_norm": 2.46875, + "learning_rate": 6.444285714285715e-07, + "loss": 0.8467, + "step": 9490 + }, + { + "epoch": 1.3772244572505528, + "grad_norm": 3.421875, + "learning_rate": 6.43e-07, + "loss": 0.8617, + "step": 9500 + }, + { + "epoch": 1.3786742053568193, + "grad_norm": 2.5625, + "learning_rate": 6.415714285714286e-07, + "loss": 0.8237, + "step": 9510 + }, + { + "epoch": 1.3801239534630858, + "grad_norm": 2.734375, + "learning_rate": 6.401428571428571e-07, + "loss": 0.8327, + "step": 9520 + }, + { + "epoch": 1.3815737015693523, + "grad_norm": 2.453125, + "learning_rate": 6.387142857142857e-07, + "loss": 0.8409, + "step": 9530 + }, + { + "epoch": 1.3830234496756189, + "grad_norm": 2.625, + "learning_rate": 6.372857142857142e-07, + "loss": 0.8074, + "step": 9540 + }, + { + "epoch": 1.3844731977818854, + "grad_norm": 2.5, + "learning_rate": 6.358571428571429e-07, + "loss": 0.7674, + "step": 9550 + }, + { + "epoch": 1.385922945888152, + "grad_norm": 2.46875, + "learning_rate": 6.344285714285715e-07, + "loss": 0.8206, + "step": 9560 + }, + { + "epoch": 1.3873726939944184, + "grad_norm": 2.390625, + "learning_rate": 6.33e-07, + "loss": 0.7873, + "step": 9570 + }, + { + "epoch": 1.388822442100685, + "grad_norm": 2.4375, + "learning_rate": 6.315714285714286e-07, + "loss": 0.8226, + "step": 9580 + }, + { + "epoch": 1.3902721902069515, + "grad_norm": 2.609375, + "learning_rate": 6.301428571428571e-07, + "loss": 0.8494, + "step": 9590 + }, + { + "epoch": 1.391721938313218, + "grad_norm": 3.578125, + "learning_rate": 6.287142857142857e-07, + "loss": 0.7998, + "step": 9600 + }, + { + "epoch": 1.3931716864194845, + "grad_norm": 2.890625, + "learning_rate": 6.272857142857143e-07, + "loss": 0.8484, + "step": 9610 + }, + { + "epoch": 1.3946214345257513, + "grad_norm": 2.84375, + "learning_rate": 6.258571428571429e-07, + "loss": 0.8299, + "step": 9620 + }, + { + "epoch": 1.3960711826320176, + "grad_norm": 3.0, + "learning_rate": 6.244285714285715e-07, + "loss": 0.7854, + "step": 9630 + }, + { + "epoch": 1.3975209307382843, + "grad_norm": 2.875, + "learning_rate": 6.23e-07, + "loss": 0.8419, + "step": 9640 + }, + { + "epoch": 1.3989706788445508, + "grad_norm": 3.0, + "learning_rate": 6.215714285714286e-07, + "loss": 0.8291, + "step": 9650 + }, + { + "epoch": 1.4004204269508174, + "grad_norm": 2.625, + "learning_rate": 6.201428571428571e-07, + "loss": 0.8063, + "step": 9660 + }, + { + "epoch": 1.401870175057084, + "grad_norm": 2.984375, + "learning_rate": 6.187142857142857e-07, + "loss": 0.8419, + "step": 9670 + }, + { + "epoch": 1.4033199231633504, + "grad_norm": 3.234375, + "learning_rate": 6.172857142857143e-07, + "loss": 0.8078, + "step": 9680 + }, + { + "epoch": 1.404769671269617, + "grad_norm": 2.921875, + "learning_rate": 6.158571428571429e-07, + "loss": 0.8663, + "step": 9690 + }, + { + "epoch": 1.4062194193758835, + "grad_norm": 2.984375, + "learning_rate": 6.144285714285714e-07, + "loss": 0.8129, + "step": 9700 + }, + { + "epoch": 1.40766916748215, + "grad_norm": 3.1875, + "learning_rate": 6.13e-07, + "loss": 0.814, + "step": 9710 + }, + { + "epoch": 1.4091189155884165, + "grad_norm": 2.421875, + "learning_rate": 6.115714285714286e-07, + "loss": 0.7925, + "step": 9720 + }, + { + "epoch": 1.410568663694683, + "grad_norm": 2.734375, + "learning_rate": 6.101428571428571e-07, + "loss": 0.8169, + "step": 9730 + }, + { + "epoch": 1.4120184118009496, + "grad_norm": 3.046875, + "learning_rate": 6.087142857142858e-07, + "loss": 0.8716, + "step": 9740 + }, + { + "epoch": 1.413468159907216, + "grad_norm": 3.1875, + "learning_rate": 6.072857142857143e-07, + "loss": 0.841, + "step": 9750 + }, + { + "epoch": 1.4149179080134826, + "grad_norm": 2.59375, + "learning_rate": 6.058571428571429e-07, + "loss": 0.8767, + "step": 9760 + }, + { + "epoch": 1.4163676561197491, + "grad_norm": 2.40625, + "learning_rate": 6.044285714285714e-07, + "loss": 0.7778, + "step": 9770 + }, + { + "epoch": 1.4178174042260157, + "grad_norm": 2.71875, + "learning_rate": 6.03e-07, + "loss": 0.8804, + "step": 9780 + }, + { + "epoch": 1.4192671523322824, + "grad_norm": 2.734375, + "learning_rate": 6.015714285714285e-07, + "loss": 0.7711, + "step": 9790 + }, + { + "epoch": 1.4207169004385487, + "grad_norm": 2.546875, + "learning_rate": 6.001428571428571e-07, + "loss": 0.8202, + "step": 9800 + }, + { + "epoch": 1.4221666485448154, + "grad_norm": 2.59375, + "learning_rate": 5.987142857142858e-07, + "loss": 0.8305, + "step": 9810 + }, + { + "epoch": 1.4236163966510818, + "grad_norm": 2.5625, + "learning_rate": 5.972857142857143e-07, + "loss": 0.8357, + "step": 9820 + }, + { + "epoch": 1.4250661447573485, + "grad_norm": 2.671875, + "learning_rate": 5.958571428571429e-07, + "loss": 0.8539, + "step": 9830 + }, + { + "epoch": 1.426515892863615, + "grad_norm": 3.015625, + "learning_rate": 5.944285714285714e-07, + "loss": 0.7294, + "step": 9840 + }, + { + "epoch": 1.4279656409698815, + "grad_norm": 2.65625, + "learning_rate": 5.93e-07, + "loss": 0.7772, + "step": 9850 + }, + { + "epoch": 1.429415389076148, + "grad_norm": 2.90625, + "learning_rate": 5.915714285714285e-07, + "loss": 0.8769, + "step": 9860 + }, + { + "epoch": 1.4308651371824146, + "grad_norm": 3.046875, + "learning_rate": 5.901428571428572e-07, + "loss": 0.7308, + "step": 9870 + }, + { + "epoch": 1.4323148852886811, + "grad_norm": 3.109375, + "learning_rate": 5.887142857142857e-07, + "loss": 0.7947, + "step": 9880 + }, + { + "epoch": 1.4337646333949476, + "grad_norm": 2.703125, + "learning_rate": 5.872857142857143e-07, + "loss": 0.8534, + "step": 9890 + }, + { + "epoch": 1.4352143815012142, + "grad_norm": 2.53125, + "learning_rate": 5.858571428571429e-07, + "loss": 0.8411, + "step": 9900 + }, + { + "epoch": 1.4366641296074807, + "grad_norm": 2.4375, + "learning_rate": 5.844285714285714e-07, + "loss": 0.7942, + "step": 9910 + }, + { + "epoch": 1.4381138777137472, + "grad_norm": 2.984375, + "learning_rate": 5.83e-07, + "loss": 0.8053, + "step": 9920 + }, + { + "epoch": 1.4395636258200137, + "grad_norm": 2.609375, + "learning_rate": 5.815714285714285e-07, + "loss": 0.8855, + "step": 9930 + }, + { + "epoch": 1.4410133739262803, + "grad_norm": 2.71875, + "learning_rate": 5.801428571428572e-07, + "loss": 0.8172, + "step": 9940 + }, + { + "epoch": 1.4424631220325468, + "grad_norm": 2.71875, + "learning_rate": 5.787142857142857e-07, + "loss": 0.8262, + "step": 9950 + }, + { + "epoch": 1.4439128701388133, + "grad_norm": 2.578125, + "learning_rate": 5.772857142857143e-07, + "loss": 0.8097, + "step": 9960 + }, + { + "epoch": 1.4453626182450798, + "grad_norm": 2.78125, + "learning_rate": 5.758571428571428e-07, + "loss": 0.834, + "step": 9970 + }, + { + "epoch": 1.4468123663513466, + "grad_norm": 2.375, + "learning_rate": 5.744285714285714e-07, + "loss": 0.8054, + "step": 9980 + }, + { + "epoch": 1.4482621144576129, + "grad_norm": 2.921875, + "learning_rate": 5.73e-07, + "loss": 0.8374, + "step": 9990 + }, + { + "epoch": 1.4497118625638796, + "grad_norm": 2.828125, + "learning_rate": 5.715714285714286e-07, + "loss": 0.831, + "step": 10000 + }, + { + "epoch": 1.451161610670146, + "grad_norm": 2.5, + "learning_rate": 5.701428571428572e-07, + "loss": 0.8216, + "step": 10010 + }, + { + "epoch": 1.4526113587764127, + "grad_norm": 2.578125, + "learning_rate": 5.687142857142857e-07, + "loss": 0.7968, + "step": 10020 + }, + { + "epoch": 1.4540611068826792, + "grad_norm": 2.328125, + "learning_rate": 5.672857142857143e-07, + "loss": 0.7597, + "step": 10030 + }, + { + "epoch": 1.4555108549889457, + "grad_norm": 2.625, + "learning_rate": 5.658571428571428e-07, + "loss": 0.7882, + "step": 10040 + }, + { + "epoch": 1.4569606030952122, + "grad_norm": 2.421875, + "learning_rate": 5.644285714285714e-07, + "loss": 0.8038, + "step": 10050 + }, + { + "epoch": 1.4584103512014788, + "grad_norm": 2.5625, + "learning_rate": 5.629999999999999e-07, + "loss": 0.7929, + "step": 10060 + }, + { + "epoch": 1.4598600993077453, + "grad_norm": 2.765625, + "learning_rate": 5.615714285714286e-07, + "loss": 0.7532, + "step": 10070 + }, + { + "epoch": 1.4613098474140118, + "grad_norm": 2.84375, + "learning_rate": 5.601428571428572e-07, + "loss": 0.9015, + "step": 10080 + }, + { + "epoch": 1.4627595955202783, + "grad_norm": 2.703125, + "learning_rate": 5.587142857142857e-07, + "loss": 0.7989, + "step": 10090 + }, + { + "epoch": 1.4642093436265449, + "grad_norm": 2.890625, + "learning_rate": 5.572857142857143e-07, + "loss": 0.7932, + "step": 10100 + }, + { + "epoch": 1.4656590917328114, + "grad_norm": 2.875, + "learning_rate": 5.558571428571428e-07, + "loss": 0.7576, + "step": 10110 + }, + { + "epoch": 1.467108839839078, + "grad_norm": 2.984375, + "learning_rate": 5.544285714285714e-07, + "loss": 0.801, + "step": 10120 + }, + { + "epoch": 1.4685585879453444, + "grad_norm": 2.765625, + "learning_rate": 5.53e-07, + "loss": 0.7886, + "step": 10130 + }, + { + "epoch": 1.470008336051611, + "grad_norm": 2.59375, + "learning_rate": 5.515714285714286e-07, + "loss": 0.7866, + "step": 10140 + }, + { + "epoch": 1.4714580841578775, + "grad_norm": 2.578125, + "learning_rate": 5.501428571428572e-07, + "loss": 0.7927, + "step": 10150 + }, + { + "epoch": 1.472907832264144, + "grad_norm": 2.3125, + "learning_rate": 5.487142857142857e-07, + "loss": 0.7921, + "step": 10160 + }, + { + "epoch": 1.4743575803704108, + "grad_norm": 2.8125, + "learning_rate": 5.472857142857143e-07, + "loss": 0.885, + "step": 10170 + }, + { + "epoch": 1.475807328476677, + "grad_norm": 2.21875, + "learning_rate": 5.458571428571428e-07, + "loss": 0.6874, + "step": 10180 + }, + { + "epoch": 1.4772570765829438, + "grad_norm": 2.625, + "learning_rate": 5.444285714285715e-07, + "loss": 0.8245, + "step": 10190 + }, + { + "epoch": 1.4787068246892103, + "grad_norm": 2.640625, + "learning_rate": 5.43e-07, + "loss": 0.7855, + "step": 10200 + }, + { + "epoch": 1.4801565727954769, + "grad_norm": 3.171875, + "learning_rate": 5.415714285714286e-07, + "loss": 0.907, + "step": 10210 + }, + { + "epoch": 1.4816063209017434, + "grad_norm": 2.65625, + "learning_rate": 5.401428571428571e-07, + "loss": 0.8146, + "step": 10220 + }, + { + "epoch": 1.48305606900801, + "grad_norm": 2.703125, + "learning_rate": 5.387142857142857e-07, + "loss": 0.7752, + "step": 10230 + }, + { + "epoch": 1.4845058171142764, + "grad_norm": 2.46875, + "learning_rate": 5.372857142857143e-07, + "loss": 0.7439, + "step": 10240 + }, + { + "epoch": 1.485955565220543, + "grad_norm": 2.875, + "learning_rate": 5.358571428571428e-07, + "loss": 0.7902, + "step": 10250 + }, + { + "epoch": 1.4874053133268095, + "grad_norm": 2.46875, + "learning_rate": 5.344285714285715e-07, + "loss": 0.8357, + "step": 10260 + }, + { + "epoch": 1.488855061433076, + "grad_norm": 2.734375, + "learning_rate": 5.33e-07, + "loss": 0.8211, + "step": 10270 + }, + { + "epoch": 1.4903048095393425, + "grad_norm": 2.265625, + "learning_rate": 5.315714285714286e-07, + "loss": 0.8713, + "step": 10280 + }, + { + "epoch": 1.491754557645609, + "grad_norm": 2.5625, + "learning_rate": 5.301428571428571e-07, + "loss": 0.839, + "step": 10290 + }, + { + "epoch": 1.4932043057518756, + "grad_norm": 2.421875, + "learning_rate": 5.287142857142857e-07, + "loss": 0.7849, + "step": 10300 + }, + { + "epoch": 1.494654053858142, + "grad_norm": 3.171875, + "learning_rate": 5.272857142857142e-07, + "loss": 0.8072, + "step": 10310 + }, + { + "epoch": 1.4961038019644086, + "grad_norm": 3.390625, + "learning_rate": 5.258571428571429e-07, + "loss": 0.8188, + "step": 10320 + }, + { + "epoch": 1.4975535500706751, + "grad_norm": 2.78125, + "learning_rate": 5.244285714285715e-07, + "loss": 0.8151, + "step": 10330 + }, + { + "epoch": 1.499003298176942, + "grad_norm": 2.640625, + "learning_rate": 5.23e-07, + "loss": 0.838, + "step": 10340 + }, + { + "epoch": 1.5004530462832082, + "grad_norm": 2.796875, + "learning_rate": 5.215714285714286e-07, + "loss": 0.7786, + "step": 10350 + }, + { + "epoch": 1.501902794389475, + "grad_norm": 2.46875, + "learning_rate": 5.201428571428571e-07, + "loss": 0.7946, + "step": 10360 + }, + { + "epoch": 1.5033525424957412, + "grad_norm": 2.453125, + "learning_rate": 5.187142857142857e-07, + "loss": 0.8138, + "step": 10370 + }, + { + "epoch": 1.504802290602008, + "grad_norm": 2.359375, + "learning_rate": 5.172857142857142e-07, + "loss": 0.7467, + "step": 10380 + }, + { + "epoch": 1.5062520387082743, + "grad_norm": 2.765625, + "learning_rate": 5.158571428571429e-07, + "loss": 0.8926, + "step": 10390 + }, + { + "epoch": 1.507701786814541, + "grad_norm": 2.921875, + "learning_rate": 5.144285714285714e-07, + "loss": 0.782, + "step": 10400 + }, + { + "epoch": 1.5091515349208076, + "grad_norm": 2.296875, + "learning_rate": 5.13e-07, + "loss": 0.7712, + "step": 10410 + }, + { + "epoch": 1.510601283027074, + "grad_norm": 2.640625, + "learning_rate": 5.115714285714286e-07, + "loss": 0.8419, + "step": 10420 + }, + { + "epoch": 1.5120510311333406, + "grad_norm": 2.90625, + "learning_rate": 5.101428571428571e-07, + "loss": 0.7775, + "step": 10430 + }, + { + "epoch": 1.5135007792396071, + "grad_norm": 2.53125, + "learning_rate": 5.087142857142857e-07, + "loss": 0.8984, + "step": 10440 + }, + { + "epoch": 1.5149505273458737, + "grad_norm": 2.515625, + "learning_rate": 5.072857142857143e-07, + "loss": 0.8071, + "step": 10450 + }, + { + "epoch": 1.5164002754521402, + "grad_norm": 2.546875, + "learning_rate": 5.058571428571429e-07, + "loss": 0.7917, + "step": 10460 + }, + { + "epoch": 1.5178500235584067, + "grad_norm": 2.5625, + "learning_rate": 5.044285714285714e-07, + "loss": 0.8027, + "step": 10470 + }, + { + "epoch": 1.5192997716646732, + "grad_norm": 2.890625, + "learning_rate": 5.03e-07, + "loss": 0.7692, + "step": 10480 + }, + { + "epoch": 1.5207495197709398, + "grad_norm": 2.5625, + "learning_rate": 5.015714285714285e-07, + "loss": 0.8074, + "step": 10490 + }, + { + "epoch": 1.5221992678772063, + "grad_norm": 2.59375, + "learning_rate": 5.001428571428571e-07, + "loss": 0.8273, + "step": 10500 + }, + { + "epoch": 1.523649015983473, + "grad_norm": 3.75, + "learning_rate": 4.987142857142857e-07, + "loss": 0.7933, + "step": 10510 + }, + { + "epoch": 1.5250987640897393, + "grad_norm": 2.53125, + "learning_rate": 4.972857142857142e-07, + "loss": 0.8166, + "step": 10520 + }, + { + "epoch": 1.526548512196006, + "grad_norm": 2.515625, + "learning_rate": 4.958571428571428e-07, + "loss": 0.8589, + "step": 10530 + }, + { + "epoch": 1.5279982603022724, + "grad_norm": 3.03125, + "learning_rate": 4.944285714285714e-07, + "loss": 0.8492, + "step": 10540 + }, + { + "epoch": 1.5294480084085391, + "grad_norm": 2.6875, + "learning_rate": 4.93e-07, + "loss": 0.8393, + "step": 10550 + }, + { + "epoch": 1.5308977565148054, + "grad_norm": 2.25, + "learning_rate": 4.915714285714285e-07, + "loss": 0.8112, + "step": 10560 + }, + { + "epoch": 1.5323475046210722, + "grad_norm": 2.84375, + "learning_rate": 4.901428571428571e-07, + "loss": 0.821, + "step": 10570 + }, + { + "epoch": 1.5337972527273385, + "grad_norm": 2.875, + "learning_rate": 4.887142857142856e-07, + "loss": 0.8244, + "step": 10580 + }, + { + "epoch": 1.5352470008336052, + "grad_norm": 2.484375, + "learning_rate": 4.872857142857142e-07, + "loss": 0.8278, + "step": 10590 + }, + { + "epoch": 1.5366967489398717, + "grad_norm": 2.53125, + "learning_rate": 4.858571428571429e-07, + "loss": 0.8576, + "step": 10600 + }, + { + "epoch": 1.5381464970461383, + "grad_norm": 2.5625, + "learning_rate": 4.844285714285714e-07, + "loss": 0.8099, + "step": 10610 + }, + { + "epoch": 1.5395962451524048, + "grad_norm": 3.296875, + "learning_rate": 4.83e-07, + "loss": 0.8593, + "step": 10620 + }, + { + "epoch": 1.5410459932586713, + "grad_norm": 2.8125, + "learning_rate": 4.815714285714285e-07, + "loss": 0.8349, + "step": 10630 + }, + { + "epoch": 1.5424957413649378, + "grad_norm": 2.671875, + "learning_rate": 4.801428571428571e-07, + "loss": 0.7727, + "step": 10640 + }, + { + "epoch": 1.5439454894712044, + "grad_norm": 2.59375, + "learning_rate": 4.787142857142856e-07, + "loss": 0.7543, + "step": 10650 + }, + { + "epoch": 1.5453952375774709, + "grad_norm": 3.203125, + "learning_rate": 4.772857142857143e-07, + "loss": 0.8336, + "step": 10660 + }, + { + "epoch": 1.5468449856837374, + "grad_norm": 2.8125, + "learning_rate": 4.7585714285714285e-07, + "loss": 0.8209, + "step": 10670 + }, + { + "epoch": 1.5482947337900042, + "grad_norm": 2.734375, + "learning_rate": 4.744285714285714e-07, + "loss": 0.8028, + "step": 10680 + }, + { + "epoch": 1.5497444818962705, + "grad_norm": 2.4375, + "learning_rate": 4.7299999999999996e-07, + "loss": 0.8109, + "step": 10690 + }, + { + "epoch": 1.5511942300025372, + "grad_norm": 2.234375, + "learning_rate": 4.7157142857142857e-07, + "loss": 0.7753, + "step": 10700 + }, + { + "epoch": 1.5526439781088035, + "grad_norm": 2.53125, + "learning_rate": 4.701428571428571e-07, + "loss": 0.7938, + "step": 10710 + }, + { + "epoch": 1.5540937262150702, + "grad_norm": 2.75, + "learning_rate": 4.687142857142857e-07, + "loss": 0.7945, + "step": 10720 + }, + { + "epoch": 1.5555434743213366, + "grad_norm": 2.734375, + "learning_rate": 4.672857142857143e-07, + "loss": 0.8301, + "step": 10730 + }, + { + "epoch": 1.5569932224276033, + "grad_norm": 2.65625, + "learning_rate": 4.6585714285714284e-07, + "loss": 0.8297, + "step": 10740 + }, + { + "epoch": 1.5584429705338696, + "grad_norm": 2.9375, + "learning_rate": 4.644285714285714e-07, + "loss": 0.8426, + "step": 10750 + }, + { + "epoch": 1.5598927186401363, + "grad_norm": 2.25, + "learning_rate": 4.63e-07, + "loss": 0.7708, + "step": 10760 + }, + { + "epoch": 1.5613424667464026, + "grad_norm": 2.828125, + "learning_rate": 4.6157142857142856e-07, + "loss": 0.8196, + "step": 10770 + }, + { + "epoch": 1.5627922148526694, + "grad_norm": 2.296875, + "learning_rate": 4.601428571428571e-07, + "loss": 0.8621, + "step": 10780 + }, + { + "epoch": 1.564241962958936, + "grad_norm": 2.765625, + "learning_rate": 4.587142857142857e-07, + "loss": 0.8039, + "step": 10790 + }, + { + "epoch": 1.5656917110652024, + "grad_norm": 3.046875, + "learning_rate": 4.572857142857143e-07, + "loss": 0.8218, + "step": 10800 + }, + { + "epoch": 1.567141459171469, + "grad_norm": 3.703125, + "learning_rate": 4.5585714285714283e-07, + "loss": 0.7837, + "step": 10810 + }, + { + "epoch": 1.5685912072777355, + "grad_norm": 2.5625, + "learning_rate": 4.544285714285714e-07, + "loss": 0.8932, + "step": 10820 + }, + { + "epoch": 1.570040955384002, + "grad_norm": 2.140625, + "learning_rate": 4.53e-07, + "loss": 0.7924, + "step": 10830 + }, + { + "epoch": 1.5714907034902685, + "grad_norm": 3.515625, + "learning_rate": 4.5157142857142855e-07, + "loss": 0.7774, + "step": 10840 + }, + { + "epoch": 1.572940451596535, + "grad_norm": 2.703125, + "learning_rate": 4.501428571428571e-07, + "loss": 0.7837, + "step": 10850 + }, + { + "epoch": 1.5743901997028016, + "grad_norm": 2.609375, + "learning_rate": 4.487142857142857e-07, + "loss": 0.8699, + "step": 10860 + }, + { + "epoch": 1.5758399478090683, + "grad_norm": 2.328125, + "learning_rate": 4.4728571428571427e-07, + "loss": 0.7944, + "step": 10870 + }, + { + "epoch": 1.5772896959153346, + "grad_norm": 2.53125, + "learning_rate": 4.458571428571428e-07, + "loss": 0.8127, + "step": 10880 + }, + { + "epoch": 1.5787394440216014, + "grad_norm": 2.609375, + "learning_rate": 4.4442857142857143e-07, + "loss": 0.7946, + "step": 10890 + }, + { + "epoch": 1.5801891921278677, + "grad_norm": 2.515625, + "learning_rate": 4.43e-07, + "loss": 0.7753, + "step": 10900 + }, + { + "epoch": 1.5816389402341344, + "grad_norm": 2.4375, + "learning_rate": 4.4157142857142854e-07, + "loss": 0.8218, + "step": 10910 + }, + { + "epoch": 1.5830886883404007, + "grad_norm": 2.875, + "learning_rate": 4.4014285714285715e-07, + "loss": 0.837, + "step": 10920 + }, + { + "epoch": 1.5845384364466675, + "grad_norm": 3.125, + "learning_rate": 4.387142857142857e-07, + "loss": 0.815, + "step": 10930 + }, + { + "epoch": 1.5859881845529338, + "grad_norm": 2.703125, + "learning_rate": 4.3728571428571426e-07, + "loss": 0.8154, + "step": 10940 + }, + { + "epoch": 1.5874379326592005, + "grad_norm": 2.828125, + "learning_rate": 4.358571428571428e-07, + "loss": 0.8589, + "step": 10950 + }, + { + "epoch": 1.588887680765467, + "grad_norm": 2.53125, + "learning_rate": 4.344285714285714e-07, + "loss": 0.8204, + "step": 10960 + }, + { + "epoch": 1.5903374288717336, + "grad_norm": 2.5, + "learning_rate": 4.3299999999999997e-07, + "loss": 0.7635, + "step": 10970 + }, + { + "epoch": 1.591787176978, + "grad_norm": 2.8125, + "learning_rate": 4.3157142857142853e-07, + "loss": 0.8369, + "step": 10980 + }, + { + "epoch": 1.5932369250842666, + "grad_norm": 2.765625, + "learning_rate": 4.3014285714285714e-07, + "loss": 0.8181, + "step": 10990 + }, + { + "epoch": 1.5946866731905331, + "grad_norm": 2.171875, + "learning_rate": 4.287142857142857e-07, + "loss": 0.7581, + "step": 11000 + }, + { + "epoch": 1.5961364212967997, + "grad_norm": 2.3125, + "learning_rate": 4.2728571428571425e-07, + "loss": 0.8259, + "step": 11010 + }, + { + "epoch": 1.5975861694030662, + "grad_norm": 2.921875, + "learning_rate": 4.2585714285714285e-07, + "loss": 0.8612, + "step": 11020 + }, + { + "epoch": 1.5990359175093327, + "grad_norm": 2.421875, + "learning_rate": 4.244285714285714e-07, + "loss": 0.8416, + "step": 11030 + }, + { + "epoch": 1.6004856656155992, + "grad_norm": 3.0, + "learning_rate": 4.2299999999999996e-07, + "loss": 0.8612, + "step": 11040 + }, + { + "epoch": 1.6019354137218658, + "grad_norm": 3.3125, + "learning_rate": 4.2157142857142857e-07, + "loss": 0.8126, + "step": 11050 + }, + { + "epoch": 1.6033851618281325, + "grad_norm": 2.671875, + "learning_rate": 4.201428571428571e-07, + "loss": 0.8409, + "step": 11060 + }, + { + "epoch": 1.6048349099343988, + "grad_norm": 2.328125, + "learning_rate": 4.187142857142857e-07, + "loss": 0.8273, + "step": 11070 + }, + { + "epoch": 1.6062846580406656, + "grad_norm": 2.640625, + "learning_rate": 4.1728571428571424e-07, + "loss": 0.7948, + "step": 11080 + }, + { + "epoch": 1.6077344061469319, + "grad_norm": 2.703125, + "learning_rate": 4.1585714285714284e-07, + "loss": 0.8522, + "step": 11090 + }, + { + "epoch": 1.6091841542531986, + "grad_norm": 3.5, + "learning_rate": 4.144285714285714e-07, + "loss": 0.8145, + "step": 11100 + }, + { + "epoch": 1.610633902359465, + "grad_norm": 2.546875, + "learning_rate": 4.1299999999999995e-07, + "loss": 0.7976, + "step": 11110 + }, + { + "epoch": 1.6120836504657317, + "grad_norm": 2.90625, + "learning_rate": 4.1157142857142856e-07, + "loss": 0.8532, + "step": 11120 + }, + { + "epoch": 1.613533398571998, + "grad_norm": 2.515625, + "learning_rate": 4.101428571428571e-07, + "loss": 0.7275, + "step": 11130 + }, + { + "epoch": 1.6149831466782647, + "grad_norm": 2.734375, + "learning_rate": 4.0871428571428567e-07, + "loss": 0.8153, + "step": 11140 + }, + { + "epoch": 1.6164328947845312, + "grad_norm": 2.78125, + "learning_rate": 4.072857142857143e-07, + "loss": 0.7863, + "step": 11150 + }, + { + "epoch": 1.6178826428907978, + "grad_norm": 2.875, + "learning_rate": 4.0585714285714283e-07, + "loss": 0.8146, + "step": 11160 + }, + { + "epoch": 1.6193323909970643, + "grad_norm": 2.46875, + "learning_rate": 4.044285714285714e-07, + "loss": 0.7843, + "step": 11170 + }, + { + "epoch": 1.6207821391033308, + "grad_norm": 2.765625, + "learning_rate": 4.03e-07, + "loss": 0.7451, + "step": 11180 + }, + { + "epoch": 1.6222318872095973, + "grad_norm": 2.4375, + "learning_rate": 4.0157142857142855e-07, + "loss": 0.8424, + "step": 11190 + }, + { + "epoch": 1.6236816353158638, + "grad_norm": 3.140625, + "learning_rate": 4.001428571428571e-07, + "loss": 0.8205, + "step": 11200 + }, + { + "epoch": 1.6251313834221304, + "grad_norm": 2.6875, + "learning_rate": 3.9871428571428566e-07, + "loss": 0.852, + "step": 11210 + }, + { + "epoch": 1.626581131528397, + "grad_norm": 3.125, + "learning_rate": 3.9728571428571427e-07, + "loss": 0.905, + "step": 11220 + }, + { + "epoch": 1.6280308796346636, + "grad_norm": 2.5, + "learning_rate": 3.958571428571428e-07, + "loss": 0.8111, + "step": 11230 + }, + { + "epoch": 1.62948062774093, + "grad_norm": 2.609375, + "learning_rate": 3.944285714285714e-07, + "loss": 0.738, + "step": 11240 + }, + { + "epoch": 1.6309303758471967, + "grad_norm": 2.8125, + "learning_rate": 3.93e-07, + "loss": 0.8548, + "step": 11250 + }, + { + "epoch": 1.632380123953463, + "grad_norm": 2.828125, + "learning_rate": 3.9157142857142854e-07, + "loss": 0.784, + "step": 11260 + }, + { + "epoch": 1.6338298720597297, + "grad_norm": 3.875, + "learning_rate": 3.901428571428571e-07, + "loss": 0.811, + "step": 11270 + }, + { + "epoch": 1.635279620165996, + "grad_norm": 3.109375, + "learning_rate": 3.887142857142857e-07, + "loss": 0.8864, + "step": 11280 + }, + { + "epoch": 1.6367293682722628, + "grad_norm": 2.453125, + "learning_rate": 3.8728571428571426e-07, + "loss": 0.8207, + "step": 11290 + }, + { + "epoch": 1.638179116378529, + "grad_norm": 2.8125, + "learning_rate": 3.858571428571428e-07, + "loss": 0.8604, + "step": 11300 + }, + { + "epoch": 1.6396288644847958, + "grad_norm": 2.859375, + "learning_rate": 3.844285714285714e-07, + "loss": 0.8158, + "step": 11310 + }, + { + "epoch": 1.6410786125910624, + "grad_norm": 3.109375, + "learning_rate": 3.83e-07, + "loss": 0.8185, + "step": 11320 + }, + { + "epoch": 1.6425283606973289, + "grad_norm": 2.734375, + "learning_rate": 3.8157142857142853e-07, + "loss": 0.8559, + "step": 11330 + }, + { + "epoch": 1.6439781088035954, + "grad_norm": 2.84375, + "learning_rate": 3.801428571428571e-07, + "loss": 0.8295, + "step": 11340 + }, + { + "epoch": 1.645427856909862, + "grad_norm": 2.390625, + "learning_rate": 3.787142857142857e-07, + "loss": 0.8142, + "step": 11350 + }, + { + "epoch": 1.6468776050161285, + "grad_norm": 2.28125, + "learning_rate": 3.7728571428571425e-07, + "loss": 0.8304, + "step": 11360 + }, + { + "epoch": 1.648327353122395, + "grad_norm": 2.4375, + "learning_rate": 3.758571428571428e-07, + "loss": 0.8166, + "step": 11370 + }, + { + "epoch": 1.6497771012286615, + "grad_norm": 2.640625, + "learning_rate": 3.744285714285714e-07, + "loss": 0.8152, + "step": 11380 + }, + { + "epoch": 1.651226849334928, + "grad_norm": 2.59375, + "learning_rate": 3.7299999999999997e-07, + "loss": 0.7771, + "step": 11390 + }, + { + "epoch": 1.6526765974411946, + "grad_norm": 3.046875, + "learning_rate": 3.715714285714285e-07, + "loss": 0.8334, + "step": 11400 + }, + { + "epoch": 1.654126345547461, + "grad_norm": 2.3125, + "learning_rate": 3.7014285714285713e-07, + "loss": 0.7661, + "step": 11410 + }, + { + "epoch": 1.6555760936537278, + "grad_norm": 2.75, + "learning_rate": 3.687142857142857e-07, + "loss": 0.7939, + "step": 11420 + }, + { + "epoch": 1.6570258417599941, + "grad_norm": 2.296875, + "learning_rate": 3.6728571428571424e-07, + "loss": 0.8028, + "step": 11430 + }, + { + "epoch": 1.6584755898662609, + "grad_norm": 2.1875, + "learning_rate": 3.6585714285714285e-07, + "loss": 0.8186, + "step": 11440 + }, + { + "epoch": 1.6599253379725272, + "grad_norm": 2.984375, + "learning_rate": 3.644285714285714e-07, + "loss": 0.8467, + "step": 11450 + }, + { + "epoch": 1.661375086078794, + "grad_norm": 2.421875, + "learning_rate": 3.6299999999999995e-07, + "loss": 0.8373, + "step": 11460 + }, + { + "epoch": 1.6628248341850602, + "grad_norm": 2.5625, + "learning_rate": 3.6157142857142856e-07, + "loss": 0.801, + "step": 11470 + }, + { + "epoch": 1.664274582291327, + "grad_norm": 2.203125, + "learning_rate": 3.601428571428571e-07, + "loss": 0.8331, + "step": 11480 + }, + { + "epoch": 1.6657243303975933, + "grad_norm": 2.375, + "learning_rate": 3.5871428571428567e-07, + "loss": 0.7665, + "step": 11490 + }, + { + "epoch": 1.66717407850386, + "grad_norm": 2.359375, + "learning_rate": 3.5728571428571423e-07, + "loss": 0.8383, + "step": 11500 + }, + { + "epoch": 1.6686238266101265, + "grad_norm": 2.9375, + "learning_rate": 3.5585714285714284e-07, + "loss": 0.8566, + "step": 11510 + }, + { + "epoch": 1.670073574716393, + "grad_norm": 2.875, + "learning_rate": 3.544285714285714e-07, + "loss": 0.7652, + "step": 11520 + }, + { + "epoch": 1.6715233228226596, + "grad_norm": 2.78125, + "learning_rate": 3.5299999999999994e-07, + "loss": 0.8145, + "step": 11530 + }, + { + "epoch": 1.672973070928926, + "grad_norm": 2.734375, + "learning_rate": 3.5157142857142855e-07, + "loss": 0.7768, + "step": 11540 + }, + { + "epoch": 1.6744228190351926, + "grad_norm": 2.734375, + "learning_rate": 3.501428571428571e-07, + "loss": 0.7858, + "step": 11550 + }, + { + "epoch": 1.6758725671414592, + "grad_norm": 2.609375, + "learning_rate": 3.4871428571428566e-07, + "loss": 0.8163, + "step": 11560 + }, + { + "epoch": 1.6773223152477257, + "grad_norm": 2.75, + "learning_rate": 3.4728571428571427e-07, + "loss": 0.8221, + "step": 11570 + }, + { + "epoch": 1.6787720633539922, + "grad_norm": 2.65625, + "learning_rate": 3.458571428571428e-07, + "loss": 0.7725, + "step": 11580 + }, + { + "epoch": 1.6802218114602587, + "grad_norm": 6.84375, + "learning_rate": 3.444285714285714e-07, + "loss": 0.7476, + "step": 11590 + }, + { + "epoch": 1.6816715595665253, + "grad_norm": 2.703125, + "learning_rate": 3.43e-07, + "loss": 0.8204, + "step": 11600 + }, + { + "epoch": 1.683121307672792, + "grad_norm": 2.46875, + "learning_rate": 3.4157142857142854e-07, + "loss": 0.8318, + "step": 11610 + }, + { + "epoch": 1.6845710557790583, + "grad_norm": 2.703125, + "learning_rate": 3.401428571428571e-07, + "loss": 0.7973, + "step": 11620 + }, + { + "epoch": 1.686020803885325, + "grad_norm": 2.921875, + "learning_rate": 3.3871428571428565e-07, + "loss": 0.8654, + "step": 11630 + }, + { + "epoch": 1.6874705519915913, + "grad_norm": 2.4375, + "learning_rate": 3.3728571428571426e-07, + "loss": 0.7964, + "step": 11640 + }, + { + "epoch": 1.688920300097858, + "grad_norm": 2.6875, + "learning_rate": 3.358571428571428e-07, + "loss": 0.7809, + "step": 11650 + }, + { + "epoch": 1.6903700482041244, + "grad_norm": 3.359375, + "learning_rate": 3.3442857142857137e-07, + "loss": 0.8053, + "step": 11660 + }, + { + "epoch": 1.6918197963103911, + "grad_norm": 2.96875, + "learning_rate": 3.33e-07, + "loss": 0.8298, + "step": 11670 + }, + { + "epoch": 1.6932695444166574, + "grad_norm": 3.0, + "learning_rate": 3.3157142857142853e-07, + "loss": 0.8435, + "step": 11680 + }, + { + "epoch": 1.6947192925229242, + "grad_norm": 3.09375, + "learning_rate": 3.301428571428571e-07, + "loss": 0.8098, + "step": 11690 + }, + { + "epoch": 1.6961690406291907, + "grad_norm": 2.75, + "learning_rate": 3.2871428571428575e-07, + "loss": 0.8635, + "step": 11700 + }, + { + "epoch": 1.6976187887354572, + "grad_norm": 2.578125, + "learning_rate": 3.2728571428571425e-07, + "loss": 0.803, + "step": 11710 + }, + { + "epoch": 1.6990685368417238, + "grad_norm": 2.578125, + "learning_rate": 3.258571428571428e-07, + "loss": 0.7972, + "step": 11720 + }, + { + "epoch": 1.7005182849479903, + "grad_norm": 2.828125, + "learning_rate": 3.2442857142857147e-07, + "loss": 0.8319, + "step": 11730 + }, + { + "epoch": 1.7019680330542568, + "grad_norm": 2.75, + "learning_rate": 3.23e-07, + "loss": 0.8166, + "step": 11740 + }, + { + "epoch": 1.7034177811605233, + "grad_norm": 2.703125, + "learning_rate": 3.215714285714286e-07, + "loss": 0.8171, + "step": 11750 + }, + { + "epoch": 1.7048675292667899, + "grad_norm": 2.421875, + "learning_rate": 3.201428571428571e-07, + "loss": 0.8326, + "step": 11760 + }, + { + "epoch": 1.7063172773730564, + "grad_norm": 3.109375, + "learning_rate": 3.1871428571428574e-07, + "loss": 0.8635, + "step": 11770 + }, + { + "epoch": 1.7077670254793231, + "grad_norm": 2.734375, + "learning_rate": 3.172857142857143e-07, + "loss": 0.8039, + "step": 11780 + }, + { + "epoch": 1.7092167735855894, + "grad_norm": 2.609375, + "learning_rate": 3.1585714285714285e-07, + "loss": 0.7861, + "step": 11790 + }, + { + "epoch": 1.7106665216918562, + "grad_norm": 2.5, + "learning_rate": 3.1442857142857146e-07, + "loss": 0.8092, + "step": 11800 + }, + { + "epoch": 1.7121162697981225, + "grad_norm": 2.734375, + "learning_rate": 3.13e-07, + "loss": 0.7777, + "step": 11810 + }, + { + "epoch": 1.7135660179043892, + "grad_norm": 3.046875, + "learning_rate": 3.1157142857142857e-07, + "loss": 0.7597, + "step": 11820 + }, + { + "epoch": 1.7150157660106555, + "grad_norm": 2.84375, + "learning_rate": 3.1014285714285717e-07, + "loss": 0.803, + "step": 11830 + }, + { + "epoch": 1.7164655141169223, + "grad_norm": 2.78125, + "learning_rate": 3.0871428571428573e-07, + "loss": 0.79, + "step": 11840 + }, + { + "epoch": 1.7179152622231886, + "grad_norm": 2.46875, + "learning_rate": 3.072857142857143e-07, + "loss": 0.8834, + "step": 11850 + }, + { + "epoch": 1.7193650103294553, + "grad_norm": 2.625, + "learning_rate": 3.058571428571429e-07, + "loss": 0.7999, + "step": 11860 + }, + { + "epoch": 1.7208147584357218, + "grad_norm": 2.859375, + "learning_rate": 3.0442857142857145e-07, + "loss": 0.8894, + "step": 11870 + }, + { + "epoch": 1.7222645065419884, + "grad_norm": 2.34375, + "learning_rate": 3.03e-07, + "loss": 0.7917, + "step": 11880 + }, + { + "epoch": 1.723714254648255, + "grad_norm": 2.453125, + "learning_rate": 3.0157142857142855e-07, + "loss": 0.8188, + "step": 11890 + }, + { + "epoch": 1.7251640027545214, + "grad_norm": 2.6875, + "learning_rate": 3.0014285714285716e-07, + "loss": 0.7753, + "step": 11900 + }, + { + "epoch": 1.726613750860788, + "grad_norm": 3.0, + "learning_rate": 2.987142857142857e-07, + "loss": 0.807, + "step": 11910 + }, + { + "epoch": 1.7280634989670545, + "grad_norm": 2.5, + "learning_rate": 2.9728571428571427e-07, + "loss": 0.835, + "step": 11920 + }, + { + "epoch": 1.729513247073321, + "grad_norm": 2.578125, + "learning_rate": 2.958571428571429e-07, + "loss": 0.8339, + "step": 11930 + }, + { + "epoch": 1.7309629951795875, + "grad_norm": 2.671875, + "learning_rate": 2.9442857142857144e-07, + "loss": 0.845, + "step": 11940 + }, + { + "epoch": 1.732412743285854, + "grad_norm": 2.59375, + "learning_rate": 2.93e-07, + "loss": 0.8415, + "step": 11950 + }, + { + "epoch": 1.7338624913921206, + "grad_norm": 2.4375, + "learning_rate": 2.915714285714286e-07, + "loss": 0.8512, + "step": 11960 + }, + { + "epoch": 1.7353122394983873, + "grad_norm": 2.546875, + "learning_rate": 2.9014285714285715e-07, + "loss": 0.7983, + "step": 11970 + }, + { + "epoch": 1.7367619876046536, + "grad_norm": 2.796875, + "learning_rate": 2.887142857142857e-07, + "loss": 0.857, + "step": 11980 + }, + { + "epoch": 1.7382117357109204, + "grad_norm": 2.640625, + "learning_rate": 2.872857142857143e-07, + "loss": 0.8359, + "step": 11990 + }, + { + "epoch": 1.7396614838171867, + "grad_norm": 3.234375, + "learning_rate": 2.8585714285714287e-07, + "loss": 0.8621, + "step": 12000 + }, + { + "epoch": 1.7411112319234534, + "grad_norm": 2.765625, + "learning_rate": 2.844285714285714e-07, + "loss": 0.8128, + "step": 12010 + }, + { + "epoch": 1.7425609800297197, + "grad_norm": 2.625, + "learning_rate": 2.83e-07, + "loss": 0.8417, + "step": 12020 + }, + { + "epoch": 1.7440107281359865, + "grad_norm": 2.53125, + "learning_rate": 2.815714285714286e-07, + "loss": 0.7689, + "step": 12030 + }, + { + "epoch": 1.7454604762422528, + "grad_norm": 2.765625, + "learning_rate": 2.8014285714285714e-07, + "loss": 0.8465, + "step": 12040 + }, + { + "epoch": 1.7469102243485195, + "grad_norm": 2.671875, + "learning_rate": 2.787142857142857e-07, + "loss": 0.7888, + "step": 12050 + }, + { + "epoch": 1.748359972454786, + "grad_norm": 2.75, + "learning_rate": 2.772857142857143e-07, + "loss": 0.8822, + "step": 12060 + }, + { + "epoch": 1.7498097205610526, + "grad_norm": 2.90625, + "learning_rate": 2.7585714285714286e-07, + "loss": 0.7979, + "step": 12070 + }, + { + "epoch": 1.751259468667319, + "grad_norm": 2.609375, + "learning_rate": 2.744285714285714e-07, + "loss": 0.8022, + "step": 12080 + }, + { + "epoch": 1.7527092167735856, + "grad_norm": 2.796875, + "learning_rate": 2.73e-07, + "loss": 0.8221, + "step": 12090 + }, + { + "epoch": 1.7541589648798521, + "grad_norm": 2.5, + "learning_rate": 2.715714285714286e-07, + "loss": 0.8559, + "step": 12100 + }, + { + "epoch": 1.7556087129861186, + "grad_norm": 2.984375, + "learning_rate": 2.7014285714285713e-07, + "loss": 0.8455, + "step": 12110 + }, + { + "epoch": 1.7570584610923852, + "grad_norm": 2.671875, + "learning_rate": 2.6871428571428574e-07, + "loss": 0.8326, + "step": 12120 + }, + { + "epoch": 1.7585082091986517, + "grad_norm": 2.875, + "learning_rate": 2.672857142857143e-07, + "loss": 0.8721, + "step": 12130 + }, + { + "epoch": 1.7599579573049182, + "grad_norm": 2.71875, + "learning_rate": 2.6585714285714285e-07, + "loss": 0.8537, + "step": 12140 + }, + { + "epoch": 1.7614077054111847, + "grad_norm": 2.765625, + "learning_rate": 2.644285714285714e-07, + "loss": 0.7957, + "step": 12150 + }, + { + "epoch": 1.7628574535174515, + "grad_norm": 2.359375, + "learning_rate": 2.63e-07, + "loss": 0.8132, + "step": 12160 + }, + { + "epoch": 1.7643072016237178, + "grad_norm": 2.890625, + "learning_rate": 2.6157142857142857e-07, + "loss": 0.8447, + "step": 12170 + }, + { + "epoch": 1.7657569497299845, + "grad_norm": 2.703125, + "learning_rate": 2.601428571428571e-07, + "loss": 0.8393, + "step": 12180 + }, + { + "epoch": 1.7672066978362508, + "grad_norm": 2.6875, + "learning_rate": 2.5871428571428573e-07, + "loss": 0.8473, + "step": 12190 + }, + { + "epoch": 1.7686564459425176, + "grad_norm": 2.4375, + "learning_rate": 2.572857142857143e-07, + "loss": 0.768, + "step": 12200 + }, + { + "epoch": 1.7701061940487839, + "grad_norm": 2.640625, + "learning_rate": 2.5585714285714284e-07, + "loss": 0.8802, + "step": 12210 + }, + { + "epoch": 1.7715559421550506, + "grad_norm": 2.546875, + "learning_rate": 2.5442857142857145e-07, + "loss": 0.8095, + "step": 12220 + }, + { + "epoch": 1.773005690261317, + "grad_norm": 2.359375, + "learning_rate": 2.53e-07, + "loss": 0.7748, + "step": 12230 + }, + { + "epoch": 1.7744554383675837, + "grad_norm": 2.6875, + "learning_rate": 2.5157142857142856e-07, + "loss": 0.8208, + "step": 12240 + }, + { + "epoch": 1.7759051864738502, + "grad_norm": 2.21875, + "learning_rate": 2.5014285714285716e-07, + "loss": 0.8418, + "step": 12250 + }, + { + "epoch": 1.7773549345801167, + "grad_norm": 2.4375, + "learning_rate": 2.487142857142857e-07, + "loss": 0.8764, + "step": 12260 + }, + { + "epoch": 1.7788046826863833, + "grad_norm": 2.828125, + "learning_rate": 2.472857142857143e-07, + "loss": 0.7999, + "step": 12270 + }, + { + "epoch": 1.7802544307926498, + "grad_norm": 2.75, + "learning_rate": 2.4585714285714283e-07, + "loss": 0.807, + "step": 12280 + }, + { + "epoch": 1.7817041788989163, + "grad_norm": 2.609375, + "learning_rate": 2.4442857142857144e-07, + "loss": 0.7841, + "step": 12290 + }, + { + "epoch": 1.7831539270051828, + "grad_norm": 2.671875, + "learning_rate": 2.43e-07, + "loss": 0.7637, + "step": 12300 + }, + { + "epoch": 1.7846036751114494, + "grad_norm": 2.4375, + "learning_rate": 2.4157142857142855e-07, + "loss": 0.8731, + "step": 12310 + }, + { + "epoch": 1.7860534232177159, + "grad_norm": 2.53125, + "learning_rate": 2.401428571428571e-07, + "loss": 0.7878, + "step": 12320 + }, + { + "epoch": 1.7875031713239826, + "grad_norm": 2.421875, + "learning_rate": 2.387142857142857e-07, + "loss": 0.8474, + "step": 12330 + }, + { + "epoch": 1.788952919430249, + "grad_norm": 3.40625, + "learning_rate": 2.372857142857143e-07, + "loss": 0.8304, + "step": 12340 + }, + { + "epoch": 1.7904026675365157, + "grad_norm": 2.40625, + "learning_rate": 2.3585714285714285e-07, + "loss": 0.8093, + "step": 12350 + }, + { + "epoch": 1.791852415642782, + "grad_norm": 3.265625, + "learning_rate": 2.3442857142857143e-07, + "loss": 0.8044, + "step": 12360 + }, + { + "epoch": 1.7933021637490487, + "grad_norm": 3.109375, + "learning_rate": 2.33e-07, + "loss": 0.8266, + "step": 12370 + }, + { + "epoch": 1.794751911855315, + "grad_norm": 2.28125, + "learning_rate": 2.3157142857142856e-07, + "loss": 0.8795, + "step": 12380 + }, + { + "epoch": 1.7962016599615818, + "grad_norm": 2.890625, + "learning_rate": 2.3014285714285714e-07, + "loss": 0.8737, + "step": 12390 + }, + { + "epoch": 1.797651408067848, + "grad_norm": 2.4375, + "learning_rate": 2.287142857142857e-07, + "loss": 0.8285, + "step": 12400 + }, + { + "epoch": 1.7991011561741148, + "grad_norm": 2.640625, + "learning_rate": 2.2728571428571428e-07, + "loss": 0.8, + "step": 12410 + }, + { + "epoch": 1.8005509042803813, + "grad_norm": 2.90625, + "learning_rate": 2.2585714285714286e-07, + "loss": 0.8492, + "step": 12420 + }, + { + "epoch": 1.8020006523866479, + "grad_norm": 2.5625, + "learning_rate": 2.2442857142857142e-07, + "loss": 0.7745, + "step": 12430 + }, + { + "epoch": 1.8034504004929144, + "grad_norm": 2.59375, + "learning_rate": 2.23e-07, + "loss": 0.8089, + "step": 12440 + }, + { + "epoch": 1.804900148599181, + "grad_norm": 2.59375, + "learning_rate": 2.2157142857142855e-07, + "loss": 0.7944, + "step": 12450 + }, + { + "epoch": 1.8063498967054474, + "grad_norm": 2.703125, + "learning_rate": 2.2014285714285713e-07, + "loss": 0.815, + "step": 12460 + }, + { + "epoch": 1.807799644811714, + "grad_norm": 2.859375, + "learning_rate": 2.1871428571428572e-07, + "loss": 0.9026, + "step": 12470 + }, + { + "epoch": 1.8092493929179805, + "grad_norm": 3.203125, + "learning_rate": 2.1728571428571427e-07, + "loss": 0.856, + "step": 12480 + }, + { + "epoch": 1.810699141024247, + "grad_norm": 2.21875, + "learning_rate": 2.1585714285714285e-07, + "loss": 0.7871, + "step": 12490 + }, + { + "epoch": 1.8121488891305135, + "grad_norm": 2.3125, + "learning_rate": 2.1442857142857143e-07, + "loss": 0.777, + "step": 12500 + }, + { + "epoch": 1.81359863723678, + "grad_norm": 2.359375, + "learning_rate": 2.13e-07, + "loss": 0.784, + "step": 12510 + }, + { + "epoch": 1.8150483853430468, + "grad_norm": 2.5625, + "learning_rate": 2.1157142857142857e-07, + "loss": 0.8354, + "step": 12520 + }, + { + "epoch": 1.816498133449313, + "grad_norm": 2.40625, + "learning_rate": 2.1014285714285712e-07, + "loss": 0.8356, + "step": 12530 + }, + { + "epoch": 1.8179478815555798, + "grad_norm": 2.953125, + "learning_rate": 2.087142857142857e-07, + "loss": 0.8485, + "step": 12540 + }, + { + "epoch": 1.8193976296618461, + "grad_norm": 2.5, + "learning_rate": 2.072857142857143e-07, + "loss": 0.8491, + "step": 12550 + }, + { + "epoch": 1.820847377768113, + "grad_norm": 2.375, + "learning_rate": 2.0585714285714284e-07, + "loss": 0.8231, + "step": 12560 + }, + { + "epoch": 1.8222971258743792, + "grad_norm": 2.765625, + "learning_rate": 2.0442857142857142e-07, + "loss": 0.774, + "step": 12570 + }, + { + "epoch": 1.823746873980646, + "grad_norm": 2.28125, + "learning_rate": 2.03e-07, + "loss": 0.8164, + "step": 12580 + }, + { + "epoch": 1.8251966220869122, + "grad_norm": 2.25, + "learning_rate": 2.0157142857142856e-07, + "loss": 0.7683, + "step": 12590 + }, + { + "epoch": 1.826646370193179, + "grad_norm": 3.1875, + "learning_rate": 2.0014285714285714e-07, + "loss": 0.8301, + "step": 12600 + }, + { + "epoch": 1.8280961182994455, + "grad_norm": 2.5625, + "learning_rate": 1.987142857142857e-07, + "loss": 0.8543, + "step": 12610 + }, + { + "epoch": 1.829545866405712, + "grad_norm": 2.4375, + "learning_rate": 1.9728571428571428e-07, + "loss": 0.8467, + "step": 12620 + }, + { + "epoch": 1.8309956145119786, + "grad_norm": 2.40625, + "learning_rate": 1.9585714285714286e-07, + "loss": 0.7802, + "step": 12630 + }, + { + "epoch": 1.832445362618245, + "grad_norm": 2.34375, + "learning_rate": 1.944285714285714e-07, + "loss": 0.7613, + "step": 12640 + }, + { + "epoch": 1.8338951107245116, + "grad_norm": 2.609375, + "learning_rate": 1.93e-07, + "loss": 0.8436, + "step": 12650 + }, + { + "epoch": 1.8353448588307781, + "grad_norm": 3.125, + "learning_rate": 1.9157142857142855e-07, + "loss": 0.8326, + "step": 12660 + }, + { + "epoch": 1.8367946069370447, + "grad_norm": 3.09375, + "learning_rate": 1.9014285714285713e-07, + "loss": 0.7919, + "step": 12670 + }, + { + "epoch": 1.8382443550433112, + "grad_norm": 2.234375, + "learning_rate": 1.887142857142857e-07, + "loss": 0.727, + "step": 12680 + }, + { + "epoch": 1.8396941031495777, + "grad_norm": 2.921875, + "learning_rate": 1.8728571428571427e-07, + "loss": 0.8251, + "step": 12690 + }, + { + "epoch": 1.8411438512558442, + "grad_norm": 2.390625, + "learning_rate": 1.8585714285714285e-07, + "loss": 0.8646, + "step": 12700 + }, + { + "epoch": 1.842593599362111, + "grad_norm": 2.65625, + "learning_rate": 1.8442857142857143e-07, + "loss": 0.8493, + "step": 12710 + }, + { + "epoch": 1.8440433474683773, + "grad_norm": 2.90625, + "learning_rate": 1.8299999999999998e-07, + "loss": 0.8343, + "step": 12720 + }, + { + "epoch": 1.845493095574644, + "grad_norm": 2.9375, + "learning_rate": 1.8157142857142857e-07, + "loss": 0.8268, + "step": 12730 + }, + { + "epoch": 1.8469428436809103, + "grad_norm": 2.359375, + "learning_rate": 1.8014285714285712e-07, + "loss": 0.8756, + "step": 12740 + }, + { + "epoch": 1.848392591787177, + "grad_norm": 3.015625, + "learning_rate": 1.787142857142857e-07, + "loss": 0.8431, + "step": 12750 + }, + { + "epoch": 1.8498423398934434, + "grad_norm": 2.515625, + "learning_rate": 1.7728571428571428e-07, + "loss": 0.7936, + "step": 12760 + }, + { + "epoch": 1.8512920879997101, + "grad_norm": 2.921875, + "learning_rate": 1.7585714285714284e-07, + "loss": 0.7677, + "step": 12770 + }, + { + "epoch": 1.8527418361059764, + "grad_norm": 3.078125, + "learning_rate": 1.7442857142857142e-07, + "loss": 0.7348, + "step": 12780 + }, + { + "epoch": 1.8541915842122432, + "grad_norm": 2.53125, + "learning_rate": 1.7299999999999997e-07, + "loss": 0.8505, + "step": 12790 + }, + { + "epoch": 1.8556413323185097, + "grad_norm": 2.640625, + "learning_rate": 1.7157142857142856e-07, + "loss": 0.8892, + "step": 12800 + }, + { + "epoch": 1.8570910804247762, + "grad_norm": 2.984375, + "learning_rate": 1.7014285714285714e-07, + "loss": 0.7968, + "step": 12810 + }, + { + "epoch": 1.8585408285310427, + "grad_norm": 2.4375, + "learning_rate": 1.687142857142857e-07, + "loss": 0.8264, + "step": 12820 + }, + { + "epoch": 1.8599905766373093, + "grad_norm": 2.15625, + "learning_rate": 1.6728571428571427e-07, + "loss": 0.8244, + "step": 12830 + }, + { + "epoch": 1.8614403247435758, + "grad_norm": 2.296875, + "learning_rate": 1.6585714285714285e-07, + "loss": 0.8559, + "step": 12840 + }, + { + "epoch": 1.8628900728498423, + "grad_norm": 2.234375, + "learning_rate": 1.644285714285714e-07, + "loss": 0.8491, + "step": 12850 + }, + { + "epoch": 1.8643398209561088, + "grad_norm": 2.78125, + "learning_rate": 1.63e-07, + "loss": 0.8191, + "step": 12860 + }, + { + "epoch": 1.8657895690623754, + "grad_norm": 2.359375, + "learning_rate": 1.6157142857142855e-07, + "loss": 0.8598, + "step": 12870 + }, + { + "epoch": 1.867239317168642, + "grad_norm": 2.171875, + "learning_rate": 1.6014285714285713e-07, + "loss": 0.828, + "step": 12880 + }, + { + "epoch": 1.8686890652749084, + "grad_norm": 2.703125, + "learning_rate": 1.587142857142857e-07, + "loss": 0.8634, + "step": 12890 + }, + { + "epoch": 1.8701388133811752, + "grad_norm": 2.171875, + "learning_rate": 1.5728571428571426e-07, + "loss": 0.7831, + "step": 12900 + }, + { + "epoch": 1.8715885614874415, + "grad_norm": 2.859375, + "learning_rate": 1.5585714285714284e-07, + "loss": 0.7579, + "step": 12910 + }, + { + "epoch": 1.8730383095937082, + "grad_norm": 2.6875, + "learning_rate": 1.544285714285714e-07, + "loss": 0.8414, + "step": 12920 + }, + { + "epoch": 1.8744880576999745, + "grad_norm": 2.78125, + "learning_rate": 1.5299999999999998e-07, + "loss": 0.808, + "step": 12930 + }, + { + "epoch": 1.8759378058062413, + "grad_norm": 2.5625, + "learning_rate": 1.5157142857142856e-07, + "loss": 0.8435, + "step": 12940 + }, + { + "epoch": 1.8773875539125076, + "grad_norm": 3.046875, + "learning_rate": 1.5014285714285712e-07, + "loss": 0.8302, + "step": 12950 + }, + { + "epoch": 1.8788373020187743, + "grad_norm": 2.75, + "learning_rate": 1.487142857142857e-07, + "loss": 0.829, + "step": 12960 + }, + { + "epoch": 1.8802870501250408, + "grad_norm": 2.609375, + "learning_rate": 1.4728571428571428e-07, + "loss": 0.8003, + "step": 12970 + }, + { + "epoch": 1.8817367982313074, + "grad_norm": 2.03125, + "learning_rate": 1.4585714285714283e-07, + "loss": 0.8025, + "step": 12980 + }, + { + "epoch": 1.8831865463375739, + "grad_norm": 2.578125, + "learning_rate": 1.4442857142857142e-07, + "loss": 0.8619, + "step": 12990 + }, + { + "epoch": 1.8846362944438404, + "grad_norm": 2.421875, + "learning_rate": 1.4299999999999997e-07, + "loss": 0.8479, + "step": 13000 + }, + { + "epoch": 1.886086042550107, + "grad_norm": 2.40625, + "learning_rate": 1.4157142857142855e-07, + "loss": 0.7885, + "step": 13010 + }, + { + "epoch": 1.8875357906563734, + "grad_norm": 2.34375, + "learning_rate": 1.4014285714285713e-07, + "loss": 0.7909, + "step": 13020 + }, + { + "epoch": 1.88898553876264, + "grad_norm": 2.8125, + "learning_rate": 1.387142857142857e-07, + "loss": 0.8453, + "step": 13030 + }, + { + "epoch": 1.8904352868689065, + "grad_norm": 3.078125, + "learning_rate": 1.372857142857143e-07, + "loss": 0.798, + "step": 13040 + }, + { + "epoch": 1.891885034975173, + "grad_norm": 2.6875, + "learning_rate": 1.3585714285714288e-07, + "loss": 0.8134, + "step": 13050 + }, + { + "epoch": 1.8933347830814395, + "grad_norm": 2.8125, + "learning_rate": 1.3442857142857143e-07, + "loss": 0.8185, + "step": 13060 + }, + { + "epoch": 1.8947845311877063, + "grad_norm": 2.34375, + "learning_rate": 1.33e-07, + "loss": 0.7857, + "step": 13070 + }, + { + "epoch": 1.8962342792939726, + "grad_norm": 2.796875, + "learning_rate": 1.3157142857142857e-07, + "loss": 0.8502, + "step": 13080 + }, + { + "epoch": 1.8976840274002393, + "grad_norm": 2.953125, + "learning_rate": 1.3014285714285715e-07, + "loss": 0.762, + "step": 13090 + }, + { + "epoch": 1.8991337755065056, + "grad_norm": 2.296875, + "learning_rate": 1.2871428571428573e-07, + "loss": 0.7886, + "step": 13100 + }, + { + "epoch": 1.9005835236127724, + "grad_norm": 2.421875, + "learning_rate": 1.2728571428571429e-07, + "loss": 0.8484, + "step": 13110 + }, + { + "epoch": 1.9020332717190387, + "grad_norm": 2.453125, + "learning_rate": 1.2585714285714287e-07, + "loss": 0.8259, + "step": 13120 + }, + { + "epoch": 1.9034830198253054, + "grad_norm": 2.765625, + "learning_rate": 1.2442857142857142e-07, + "loss": 0.8271, + "step": 13130 + }, + { + "epoch": 1.9049327679315717, + "grad_norm": 2.5625, + "learning_rate": 1.23e-07, + "loss": 0.8196, + "step": 13140 + }, + { + "epoch": 1.9063825160378385, + "grad_norm": 3.046875, + "learning_rate": 1.2157142857142856e-07, + "loss": 0.8847, + "step": 13150 + }, + { + "epoch": 1.907832264144105, + "grad_norm": 2.640625, + "learning_rate": 1.2014285714285714e-07, + "loss": 0.8433, + "step": 13160 + }, + { + "epoch": 1.9092820122503715, + "grad_norm": 2.328125, + "learning_rate": 1.1871428571428571e-07, + "loss": 0.8433, + "step": 13170 + }, + { + "epoch": 1.910731760356638, + "grad_norm": 2.78125, + "learning_rate": 1.1728571428571428e-07, + "loss": 0.8407, + "step": 13180 + }, + { + "epoch": 1.9121815084629046, + "grad_norm": 2.65625, + "learning_rate": 1.1585714285714284e-07, + "loss": 0.7998, + "step": 13190 + }, + { + "epoch": 1.913631256569171, + "grad_norm": 2.484375, + "learning_rate": 1.1442857142857142e-07, + "loss": 0.8513, + "step": 13200 + }, + { + "epoch": 1.9150810046754376, + "grad_norm": 2.609375, + "learning_rate": 1.1299999999999999e-07, + "loss": 0.8452, + "step": 13210 + }, + { + "epoch": 1.9165307527817041, + "grad_norm": 2.8125, + "learning_rate": 1.1157142857142856e-07, + "loss": 0.7979, + "step": 13220 + }, + { + "epoch": 1.9179805008879707, + "grad_norm": 2.328125, + "learning_rate": 1.1014285714285713e-07, + "loss": 0.8147, + "step": 13230 + }, + { + "epoch": 1.9194302489942372, + "grad_norm": 2.453125, + "learning_rate": 1.087142857142857e-07, + "loss": 0.8008, + "step": 13240 + }, + { + "epoch": 1.9208799971005037, + "grad_norm": 2.828125, + "learning_rate": 1.0728571428571428e-07, + "loss": 0.8438, + "step": 13250 + }, + { + "epoch": 1.9223297452067705, + "grad_norm": 2.4375, + "learning_rate": 1.0585714285714286e-07, + "loss": 0.7993, + "step": 13260 + }, + { + "epoch": 1.9237794933130368, + "grad_norm": 2.265625, + "learning_rate": 1.0442857142857143e-07, + "loss": 0.7801, + "step": 13270 + }, + { + "epoch": 1.9252292414193035, + "grad_norm": 2.734375, + "learning_rate": 1.03e-07, + "loss": 0.8806, + "step": 13280 + }, + { + "epoch": 1.9266789895255698, + "grad_norm": 3.203125, + "learning_rate": 1.0157142857142858e-07, + "loss": 0.7843, + "step": 13290 + }, + { + "epoch": 1.9281287376318366, + "grad_norm": 2.375, + "learning_rate": 1.0014285714285715e-07, + "loss": 0.8979, + "step": 13300 + }, + { + "epoch": 1.9295784857381029, + "grad_norm": 3.0, + "learning_rate": 9.871428571428571e-08, + "loss": 0.8026, + "step": 13310 + }, + { + "epoch": 1.9310282338443696, + "grad_norm": 2.484375, + "learning_rate": 9.728571428571428e-08, + "loss": 0.8321, + "step": 13320 + }, + { + "epoch": 1.932477981950636, + "grad_norm": 2.90625, + "learning_rate": 9.585714285714286e-08, + "loss": 0.8682, + "step": 13330 + }, + { + "epoch": 1.9339277300569027, + "grad_norm": 2.421875, + "learning_rate": 9.442857142857143e-08, + "loss": 0.7917, + "step": 13340 + }, + { + "epoch": 1.9353774781631692, + "grad_norm": 2.8125, + "learning_rate": 9.3e-08, + "loss": 0.8086, + "step": 13350 + }, + { + "epoch": 1.9368272262694357, + "grad_norm": 2.453125, + "learning_rate": 9.157142857142857e-08, + "loss": 0.8097, + "step": 13360 + }, + { + "epoch": 1.9382769743757022, + "grad_norm": 2.359375, + "learning_rate": 9.014285714285715e-08, + "loss": 0.8072, + "step": 13370 + }, + { + "epoch": 1.9397267224819688, + "grad_norm": 2.96875, + "learning_rate": 8.871428571428572e-08, + "loss": 0.8249, + "step": 13380 + }, + { + "epoch": 1.9411764705882353, + "grad_norm": 2.453125, + "learning_rate": 8.728571428571428e-08, + "loss": 0.7796, + "step": 13390 + }, + { + "epoch": 1.9426262186945018, + "grad_norm": 2.421875, + "learning_rate": 8.585714285714285e-08, + "loss": 0.8064, + "step": 13400 + }, + { + "epoch": 1.9440759668007683, + "grad_norm": 2.671875, + "learning_rate": 8.442857142857142e-08, + "loss": 0.7801, + "step": 13410 + }, + { + "epoch": 1.9455257149070349, + "grad_norm": 3.28125, + "learning_rate": 8.3e-08, + "loss": 0.8028, + "step": 13420 + }, + { + "epoch": 1.9469754630133016, + "grad_norm": 2.890625, + "learning_rate": 8.157142857142857e-08, + "loss": 0.8497, + "step": 13430 + }, + { + "epoch": 1.948425211119568, + "grad_norm": 2.953125, + "learning_rate": 8.014285714285714e-08, + "loss": 0.7967, + "step": 13440 + }, + { + "epoch": 1.9498749592258346, + "grad_norm": 2.515625, + "learning_rate": 7.87142857142857e-08, + "loss": 0.8126, + "step": 13450 + }, + { + "epoch": 1.951324707332101, + "grad_norm": 2.671875, + "learning_rate": 7.728571428571429e-08, + "loss": 0.8411, + "step": 13460 + }, + { + "epoch": 1.9527744554383677, + "grad_norm": 2.421875, + "learning_rate": 7.585714285714286e-08, + "loss": 0.889, + "step": 13470 + }, + { + "epoch": 1.954224203544634, + "grad_norm": 3.125, + "learning_rate": 7.442857142857142e-08, + "loss": 0.8292, + "step": 13480 + }, + { + "epoch": 1.9556739516509007, + "grad_norm": 2.859375, + "learning_rate": 7.299999999999999e-08, + "loss": 0.8002, + "step": 13490 + }, + { + "epoch": 1.957123699757167, + "grad_norm": 2.8125, + "learning_rate": 7.157142857142857e-08, + "loss": 0.8194, + "step": 13500 + }, + { + "epoch": 1.9585734478634338, + "grad_norm": 2.96875, + "learning_rate": 7.014285714285714e-08, + "loss": 0.8333, + "step": 13510 + }, + { + "epoch": 1.9600231959697003, + "grad_norm": 3.1875, + "learning_rate": 6.871428571428571e-08, + "loss": 0.8607, + "step": 13520 + }, + { + "epoch": 1.9614729440759668, + "grad_norm": 2.78125, + "learning_rate": 6.728571428571428e-08, + "loss": 0.8509, + "step": 13530 + }, + { + "epoch": 1.9629226921822334, + "grad_norm": 3.234375, + "learning_rate": 6.585714285714285e-08, + "loss": 0.8044, + "step": 13540 + }, + { + "epoch": 1.9643724402884999, + "grad_norm": 2.171875, + "learning_rate": 6.442857142857143e-08, + "loss": 0.8042, + "step": 13550 + }, + { + "epoch": 1.9658221883947664, + "grad_norm": 2.4375, + "learning_rate": 6.3e-08, + "loss": 0.8003, + "step": 13560 + }, + { + "epoch": 1.967271936501033, + "grad_norm": 2.453125, + "learning_rate": 6.157142857142856e-08, + "loss": 0.8565, + "step": 13570 + }, + { + "epoch": 1.9687216846072995, + "grad_norm": 2.453125, + "learning_rate": 6.014285714285714e-08, + "loss": 0.8083, + "step": 13580 + }, + { + "epoch": 1.970171432713566, + "grad_norm": 2.828125, + "learning_rate": 5.871428571428571e-08, + "loss": 0.8022, + "step": 13590 + }, + { + "epoch": 1.9716211808198325, + "grad_norm": 2.71875, + "learning_rate": 5.728571428571429e-08, + "loss": 0.8345, + "step": 13600 + }, + { + "epoch": 1.973070928926099, + "grad_norm": 2.796875, + "learning_rate": 5.5857142857142855e-08, + "loss": 0.8072, + "step": 13610 + }, + { + "epoch": 1.9745206770323658, + "grad_norm": 2.734375, + "learning_rate": 5.442857142857143e-08, + "loss": 0.8244, + "step": 13620 + }, + { + "epoch": 1.975970425138632, + "grad_norm": 2.609375, + "learning_rate": 5.3e-08, + "loss": 0.8511, + "step": 13630 + }, + { + "epoch": 1.9774201732448988, + "grad_norm": 2.3125, + "learning_rate": 5.157142857142857e-08, + "loss": 0.7875, + "step": 13640 + }, + { + "epoch": 1.9788699213511651, + "grad_norm": 2.734375, + "learning_rate": 5.014285714285714e-08, + "loss": 0.8104, + "step": 13650 + }, + { + "epoch": 1.9803196694574319, + "grad_norm": 3.359375, + "learning_rate": 4.8714285714285716e-08, + "loss": 0.8374, + "step": 13660 + }, + { + "epoch": 1.9817694175636982, + "grad_norm": 2.390625, + "learning_rate": 4.7285714285714284e-08, + "loss": 0.7904, + "step": 13670 + }, + { + "epoch": 1.983219165669965, + "grad_norm": 2.953125, + "learning_rate": 4.585714285714286e-08, + "loss": 0.8194, + "step": 13680 + }, + { + "epoch": 1.9846689137762312, + "grad_norm": 3.09375, + "learning_rate": 4.4428571428571427e-08, + "loss": 0.8037, + "step": 13690 + }, + { + "epoch": 1.986118661882498, + "grad_norm": 2.953125, + "learning_rate": 4.2999999999999995e-08, + "loss": 0.7715, + "step": 13700 + }, + { + "epoch": 1.9875684099887645, + "grad_norm": 2.5625, + "learning_rate": 4.157142857142857e-08, + "loss": 0.7816, + "step": 13710 + }, + { + "epoch": 1.989018158095031, + "grad_norm": 2.28125, + "learning_rate": 4.014285714285714e-08, + "loss": 0.823, + "step": 13720 + }, + { + "epoch": 1.9904679062012975, + "grad_norm": 2.453125, + "learning_rate": 3.871428571428571e-08, + "loss": 0.8054, + "step": 13730 + }, + { + "epoch": 1.991917654307564, + "grad_norm": 2.953125, + "learning_rate": 3.728571428571428e-08, + "loss": 0.7687, + "step": 13740 + }, + { + "epoch": 1.9933674024138306, + "grad_norm": 2.609375, + "learning_rate": 3.5857142857142855e-08, + "loss": 0.7404, + "step": 13750 + }, + { + "epoch": 1.9948171505200971, + "grad_norm": 2.6875, + "learning_rate": 3.442857142857142e-08, + "loss": 0.8413, + "step": 13760 + }, + { + "epoch": 1.9962668986263636, + "grad_norm": 2.734375, + "learning_rate": 3.3e-08, + "loss": 0.8277, + "step": 13770 + }, + { + "epoch": 1.9977166467326302, + "grad_norm": 2.703125, + "learning_rate": 3.1571428571428566e-08, + "loss": 0.7982, + "step": 13780 + }, + { + "epoch": 1.9991663948388967, + "grad_norm": 2.734375, + "learning_rate": 3.014285714285714e-08, + "loss": 0.8439, + "step": 13790 + }, + { + "epoch": 2.000579899242507, + "grad_norm": 2.640625, + "learning_rate": 2.8714285714285712e-08, + "loss": 0.8232, + "step": 13800 + }, + { + "epoch": 2.002029647348773, + "grad_norm": 2.21875, + "learning_rate": 2.7285714285714283e-08, + "loss": 0.796, + "step": 13810 + }, + { + "epoch": 2.00347939545504, + "grad_norm": 2.90625, + "learning_rate": 2.5857142857142855e-08, + "loss": 0.8377, + "step": 13820 + }, + { + "epoch": 2.004929143561306, + "grad_norm": 3.140625, + "learning_rate": 2.4428571428571426e-08, + "loss": 0.7335, + "step": 13830 + }, + { + "epoch": 2.006378891667573, + "grad_norm": 2.515625, + "learning_rate": 2.2999999999999998e-08, + "loss": 0.8438, + "step": 13840 + }, + { + "epoch": 2.007828639773839, + "grad_norm": 2.46875, + "learning_rate": 2.157142857142857e-08, + "loss": 0.8081, + "step": 13850 + }, + { + "epoch": 2.009278387880106, + "grad_norm": 2.75, + "learning_rate": 2.0142857142857144e-08, + "loss": 0.7972, + "step": 13860 + }, + { + "epoch": 2.0107281359863722, + "grad_norm": 2.703125, + "learning_rate": 1.8714285714285715e-08, + "loss": 0.818, + "step": 13870 + }, + { + "epoch": 2.012177884092639, + "grad_norm": 2.375, + "learning_rate": 1.7285714285714286e-08, + "loss": 0.7841, + "step": 13880 + }, + { + "epoch": 2.0136276321989053, + "grad_norm": 2.59375, + "learning_rate": 1.5857142857142858e-08, + "loss": 0.8538, + "step": 13890 + }, + { + "epoch": 2.015077380305172, + "grad_norm": 2.3125, + "learning_rate": 1.4428571428571428e-08, + "loss": 0.8128, + "step": 13900 + }, + { + "epoch": 2.0165271284114383, + "grad_norm": 2.421875, + "learning_rate": 1.2999999999999999e-08, + "loss": 0.815, + "step": 13910 + }, + { + "epoch": 2.017976876517705, + "grad_norm": 2.96875, + "learning_rate": 1.157142857142857e-08, + "loss": 0.8338, + "step": 13920 + }, + { + "epoch": 2.0194266246239714, + "grad_norm": 3.109375, + "learning_rate": 1.0142857142857142e-08, + "loss": 0.802, + "step": 13930 + }, + { + "epoch": 2.020876372730238, + "grad_norm": 2.90625, + "learning_rate": 8.714285714285713e-09, + "loss": 0.8423, + "step": 13940 + }, + { + "epoch": 2.022326120836505, + "grad_norm": 4.3125, + "learning_rate": 7.285714285714285e-09, + "loss": 0.8071, + "step": 13950 + }, + { + "epoch": 2.023775868942771, + "grad_norm": 2.765625, + "learning_rate": 5.857142857142857e-09, + "loss": 0.7923, + "step": 13960 + }, + { + "epoch": 2.025225617049038, + "grad_norm": 2.828125, + "learning_rate": 4.428571428571428e-09, + "loss": 0.8269, + "step": 13970 + }, + { + "epoch": 2.026675365155304, + "grad_norm": 2.828125, + "learning_rate": 3e-09, + "loss": 0.8181, + "step": 13980 + }, + { + "epoch": 2.028125113261571, + "grad_norm": 3.15625, + "learning_rate": 1.5714285714285714e-09, + "loss": 0.8523, + "step": 13990 + }, + { + "epoch": 2.0295748613678373, + "grad_norm": 2.359375, + "learning_rate": 1.4285714285714285e-10, + "loss": 0.8096, + "step": 14000 + } + ], + "logging_steps": 10, + "max_steps": 14000, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.0063048273646844e+19, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}