| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 900, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.625, | |
| "learning_rate": 7.4074074074074075e-06, | |
| "loss": 0.4546, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.41796875, | |
| "learning_rate": 9.994529604032901e-06, | |
| "loss": 0.2781, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.38671875, | |
| "learning_rate": 9.964784918620284e-06, | |
| "loss": 0.2492, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.3984375, | |
| "learning_rate": 9.909333658705933e-06, | |
| "loss": 0.2379, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.400390625, | |
| "learning_rate": 9.82846293883108e-06, | |
| "loss": 0.2301, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.37109375, | |
| "learning_rate": 9.722591489961829e-06, | |
| "loss": 0.2255, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.388671875, | |
| "learning_rate": 9.59226749139145e-06, | |
| "loss": 0.2223, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.76171875, | |
| "learning_rate": 9.438165732387471e-06, | |
| "loss": 0.2171, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.39453125, | |
| "learning_rate": 9.261084118279846e-06, | |
| "loss": 0.2154, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.41015625, | |
| "learning_rate": 9.06193953908105e-06, | |
| "loss": 0.2126, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.37109375, | |
| "learning_rate": 8.84176312202936e-06, | |
| "loss": 0.2143, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.37890625, | |
| "learning_rate": 8.601694892636701e-06, | |
| "loss": 0.2082, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.369140625, | |
| "learning_rate": 8.34297787188496e-06, | |
| "loss": 0.2072, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.404296875, | |
| "learning_rate": 8.066951640134183e-06, | |
| "loss": 0.2069, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.404296875, | |
| "learning_rate": 7.77504540106735e-06, | |
| "loss": 0.2067, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "grad_norm": 0.431640625, | |
| "learning_rate": 7.468770581585147e-06, | |
| "loss": 0.1909, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "grad_norm": 0.39453125, | |
| "learning_rate": 7.149713005966784e-06, | |
| "loss": 0.1909, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "grad_norm": 0.400390625, | |
| "learning_rate": 6.819524684817439e-06, | |
| "loss": 0.1877, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "grad_norm": 0.396484375, | |
| "learning_rate": 6.479915261317299e-06, | |
| "loss": 0.1901, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "grad_norm": 0.404296875, | |
| "learning_rate": 6.132643159061707e-06, | |
| "loss": 0.1886, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "grad_norm": 0.376953125, | |
| "learning_rate": 5.779506477326933e-06, | |
| "loss": 0.1887, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "grad_norm": 0.404296875, | |
| "learning_rate": 5.4223336809039205e-06, | |
| "loss": 0.1866, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "grad_norm": 0.419921875, | |
| "learning_rate": 5.062974132706017e-06, | |
| "loss": 0.1887, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "grad_norm": 0.390625, | |
| "learning_rate": 4.703288518170774e-06, | |
| "loss": 0.1849, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "grad_norm": 0.390625, | |
| "learning_rate": 4.345139211036192e-06, | |
| "loss": 0.187, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "grad_norm": 0.421875, | |
| "learning_rate": 3.9903806303753036e-06, | |
| "loss": 0.1839, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "grad_norm": 0.40234375, | |
| "learning_rate": 3.6408496388182857e-06, | |
| "loss": 0.187, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "grad_norm": 0.390625, | |
| "learning_rate": 3.2983560316780104e-06, | |
| "loss": 0.1849, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "grad_norm": 0.412109375, | |
| "learning_rate": 2.9646731662242554e-06, | |
| "loss": 0.184, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.4140625, | |
| "learning_rate": 2.6415287796261707e-06, | |
| "loss": 0.1844, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "grad_norm": 0.412109375, | |
| "learning_rate": 2.330596043105683e-06, | |
| "loss": 0.1795, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "grad_norm": 0.4453125, | |
| "learning_rate": 2.03348489862149e-06, | |
| "loss": 0.1787, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "grad_norm": 0.39453125, | |
| "learning_rate": 1.7517337229403946e-06, | |
| "loss": 0.1781, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "grad_norm": 0.39453125, | |
| "learning_rate": 1.4868013622576138e-06, | |
| "loss": 0.1759, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "grad_norm": 0.404296875, | |
| "learning_rate": 1.240059578609054e-06, | |
| "loss": 0.178, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "grad_norm": 0.416015625, | |
| "learning_rate": 1.012785947186397e-06, | |
| "loss": 0.1785, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "grad_norm": 0.40234375, | |
| "learning_rate": 8.061572413311253e-07, | |
| "loss": 0.1778, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "grad_norm": 0.404296875, | |
| "learning_rate": 6.212433394585865e-07, | |
| "loss": 0.1771, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "grad_norm": 0.41796875, | |
| "learning_rate": 4.590016854606727e-07, | |
| "loss": 0.1756, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "grad_norm": 0.392578125, | |
| "learning_rate": 3.2027233126997405e-07, | |
| "loss": 0.1776, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "grad_norm": 0.400390625, | |
| "learning_rate": 2.057735872539157e-07, | |
| "loss": 0.1764, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "grad_norm": 0.412109375, | |
| "learning_rate": 1.1609830296019142e-07, | |
| "loss": 0.1815, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "grad_norm": 0.490234375, | |
| "learning_rate": 5.1710797470987393e-08, | |
| "loss": 0.1763, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "grad_norm": 0.400390625, | |
| "learning_rate": 1.2944455259944476e-08, | |
| "loss": 0.1788, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 0.41015625, | |
| "learning_rate": 0.0, | |
| "loss": 0.1766, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 900, | |
| "total_flos": 4.549145968838181e+18, | |
| "train_loss": 0.20135066880120173, | |
| "train_runtime": 71504.7155, | |
| "train_samples_per_second": 2.411, | |
| "train_steps_per_second": 0.013 | |
| } | |
| ], | |
| "logging_steps": 20, | |
| "max_steps": 900, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 10000, | |
| "total_flos": 4.549145968838181e+18, | |
| "train_batch_size": 24, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |