| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9985257985257985, | |
| "eval_steps": 500, | |
| "global_step": 508, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 562.125, | |
| "epoch": 0.0019656019656019656, | |
| "grad_norm": 0.0, | |
| "kl": 0.0, | |
| "learning_rate": 5e-08, | |
| "loss": 0.0, | |
| "num_tokens": 22458.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 1 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 382.84375, | |
| "epoch": 0.003931203931203931, | |
| "grad_norm": 0.0, | |
| "kl": 0.0, | |
| "learning_rate": 1e-07, | |
| "loss": 0.0, | |
| "num_tokens": 39539.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 2 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 296.84375, | |
| "epoch": 0.005896805896805897, | |
| "grad_norm": 0.0, | |
| "kl": 0.0, | |
| "learning_rate": 1.5e-07, | |
| "loss": 0.0, | |
| "num_tokens": 52892.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 3 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 562.84375, | |
| "epoch": 0.007862407862407862, | |
| "grad_norm": 0.0, | |
| "kl": 0.0, | |
| "learning_rate": 2e-07, | |
| "loss": 0.0, | |
| "num_tokens": 73799.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 4 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 454.9375, | |
| "epoch": 0.009828009828009828, | |
| "grad_norm": 0.0, | |
| "kl": 0.0, | |
| "learning_rate": 2.5e-07, | |
| "loss": 0.0, | |
| "num_tokens": 92389.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 5 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 401.5625, | |
| "epoch": 0.011793611793611793, | |
| "grad_norm": 0.8271026359027706, | |
| "kl": 0.0, | |
| "learning_rate": 3e-07, | |
| "loss": 0.0459, | |
| "num_tokens": 108305.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 6 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 390.625, | |
| "epoch": 0.013759213759213759, | |
| "grad_norm": 0.0003446225536382493, | |
| "kl": 0.0002499818801879883, | |
| "learning_rate": 3.5e-07, | |
| "loss": 0.0, | |
| "num_tokens": 125607.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 7 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 448.75, | |
| "epoch": 0.015724815724815724, | |
| "grad_norm": 0.0013026212429223525, | |
| "kl": 0.0004665851593017578, | |
| "learning_rate": 4e-07, | |
| "loss": 0.0, | |
| "num_tokens": 142955.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 8 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 461.90625, | |
| "epoch": 0.01769041769041769, | |
| "grad_norm": 0.0021171088963702245, | |
| "kl": 0.0006817877292633057, | |
| "learning_rate": 4.5e-07, | |
| "loss": 0.0, | |
| "num_tokens": 162956.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 9 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 330.6875, | |
| "epoch": 0.019656019656019656, | |
| "grad_norm": 1.3023849518044601, | |
| "kl": 0.0006229877471923828, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0444, | |
| "num_tokens": 177758.0, | |
| "reward": 0.0625, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.0625, | |
| "step": 10 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 513.1875, | |
| "epoch": 0.021621621621621623, | |
| "grad_norm": 0.00027502974416413834, | |
| "kl": 0.00018683075904846191, | |
| "learning_rate": 5.5e-07, | |
| "loss": 0.0, | |
| "num_tokens": 197364.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 11 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 424.625, | |
| "epoch": 0.023587223587223587, | |
| "grad_norm": 0.6170315097653724, | |
| "kl": 0.00029702484607696533, | |
| "learning_rate": 6e-07, | |
| "loss": -0.0489, | |
| "num_tokens": 214156.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 12 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 532.1875, | |
| "epoch": 0.025552825552825554, | |
| "grad_norm": 0.0002590285678817397, | |
| "kl": 0.00019553303718566895, | |
| "learning_rate": 6.5e-07, | |
| "loss": 0.0, | |
| "num_tokens": 234938.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 13 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 391.6875, | |
| "epoch": 0.027518427518427518, | |
| "grad_norm": 0.2468481109530296, | |
| "kl": 0.0002957582473754883, | |
| "learning_rate": 7e-07, | |
| "loss": -0.042, | |
| "num_tokens": 250970.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 14 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 526.75, | |
| "epoch": 0.029484029484029485, | |
| "grad_norm": 0.6658783590054981, | |
| "kl": 0.00033867359161376953, | |
| "learning_rate": 7.5e-07, | |
| "loss": -0.0001, | |
| "num_tokens": 271842.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 15 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 357.5, | |
| "epoch": 0.03144963144963145, | |
| "grad_norm": 0.001449294133896928, | |
| "kl": 0.00034862756729125977, | |
| "learning_rate": 8e-07, | |
| "loss": 0.0, | |
| "num_tokens": 288920.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 16 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 466.9375, | |
| "epoch": 0.03341523341523341, | |
| "grad_norm": 0.000550624611359764, | |
| "kl": 0.00036454200744628906, | |
| "learning_rate": 8.499999999999999e-07, | |
| "loss": 0.0, | |
| "num_tokens": 306990.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 17 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 425.15625, | |
| "epoch": 0.03538083538083538, | |
| "grad_norm": 0.4966139299418229, | |
| "kl": 0.0004194974899291992, | |
| "learning_rate": 9e-07, | |
| "loss": -0.059, | |
| "num_tokens": 324483.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 18 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 558.84375, | |
| "epoch": 0.03734643734643735, | |
| "grad_norm": 0.0005133453694281673, | |
| "kl": 0.000286102294921875, | |
| "learning_rate": 9.499999999999999e-07, | |
| "loss": 0.0, | |
| "num_tokens": 346726.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 19 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 462.96875, | |
| "epoch": 0.03931203931203931, | |
| "grad_norm": 0.7524081119290347, | |
| "kl": 0.0004723072052001953, | |
| "learning_rate": 1e-06, | |
| "loss": -0.051, | |
| "num_tokens": 364915.0, | |
| "reward": 0.0625, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.0625, | |
| "step": 20 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 442.84375, | |
| "epoch": 0.041277641277641275, | |
| "grad_norm": 0.002276209851987083, | |
| "kl": 0.00047028064727783203, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 382538.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 21 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 487.8125, | |
| "epoch": 0.043243243243243246, | |
| "grad_norm": 0.0009159776748314612, | |
| "kl": 0.0003075599670410156, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 402660.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 22 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 521.09375, | |
| "epoch": 0.04520884520884521, | |
| "grad_norm": 0.4431344945845042, | |
| "kl": 0.0008206367492675781, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0654, | |
| "num_tokens": 423169.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 23 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 424.9375, | |
| "epoch": 0.04717444717444717, | |
| "grad_norm": 0.00201223494239014, | |
| "kl": 0.0007357597351074219, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 439839.0, | |
| "reward": 0.0625, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0625, | |
| "step": 24 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 593.625, | |
| "epoch": 0.04914004914004914, | |
| "grad_norm": 0.18561111784701523, | |
| "kl": 0.0019412040710449219, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0044, | |
| "num_tokens": 463265.0, | |
| "reward": 0.09375, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.09375, | |
| "step": 25 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 280.375, | |
| "epoch": 0.05110565110565111, | |
| "grad_norm": 5.884242558131956, | |
| "kl": 0.0025135278701782227, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0569, | |
| "num_tokens": 475453.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 26 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 513.90625, | |
| "epoch": 0.05307125307125307, | |
| "grad_norm": 0.0019098622203968013, | |
| "kl": 0.0006109476089477539, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 495570.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 27 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 537.8125, | |
| "epoch": 0.055036855036855035, | |
| "grad_norm": 0.22075214354309136, | |
| "kl": 0.0009493827819824219, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0333, | |
| "num_tokens": 516232.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 28 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 485.6875, | |
| "epoch": 0.057002457002457, | |
| "grad_norm": 0.0027609914881789005, | |
| "kl": 0.0013211965560913086, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 534828.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 29 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 394.8125, | |
| "epoch": 0.05896805896805897, | |
| "grad_norm": 1035336.9697758019, | |
| "kl": 7488.001363635063, | |
| "learning_rate": 1e-06, | |
| "loss": 7.5002, | |
| "num_tokens": 550682.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 30 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 472.96875, | |
| "epoch": 0.060933660933660934, | |
| "grad_norm": 0.005787911590971138, | |
| "kl": 0.002418994903564453, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 568901.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 31 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 532.1875, | |
| "epoch": 0.0628992628992629, | |
| "grad_norm": 0.22606153101175416, | |
| "kl": 0.0019965171813964844, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 589439.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 32 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 443.78125, | |
| "epoch": 0.06486486486486487, | |
| "grad_norm": 0.35741687124698773, | |
| "kl": 0.002090930938720703, | |
| "learning_rate": 1e-06, | |
| "loss": 0.007, | |
| "num_tokens": 606976.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 33 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 582.25, | |
| "epoch": 0.06683046683046683, | |
| "grad_norm": 0.0012562903075668153, | |
| "kl": 0.0010848045349121094, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 629146.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 34 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 445.96875, | |
| "epoch": 0.0687960687960688, | |
| "grad_norm": 0.9035998758815753, | |
| "kl": 0.0037384033203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0705, | |
| "num_tokens": 646643.0, | |
| "reward": 0.125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.125, | |
| "step": 35 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 444.6875, | |
| "epoch": 0.07076167076167077, | |
| "grad_norm": 0.3422524583969719, | |
| "kl": 0.0064411163330078125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0749, | |
| "num_tokens": 666223.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 36 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 375.15625, | |
| "epoch": 0.07272727272727272, | |
| "grad_norm": 0.6543309959643355, | |
| "kl": 0.025150299072265625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.074, | |
| "num_tokens": 681430.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 37 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 465.40625, | |
| "epoch": 0.0746928746928747, | |
| "grad_norm": 0.4629932120108371, | |
| "kl": 0.006771087646484375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 699985.0, | |
| "reward": 0.09375, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.09375, | |
| "step": 38 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 386.46875, | |
| "epoch": 0.07665847665847667, | |
| "grad_norm": 0.2536795642942375, | |
| "kl": 0.010662078857421875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0128, | |
| "num_tokens": 715666.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 39 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 398.90625, | |
| "epoch": 0.07862407862407862, | |
| "grad_norm": 0.36055266191548657, | |
| "kl": 0.014057159423828125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0222, | |
| "num_tokens": 732739.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 40 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 497.78125, | |
| "epoch": 0.08058968058968059, | |
| "grad_norm": 0.5439508463238408, | |
| "kl": 0.011962890625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0736, | |
| "num_tokens": 752036.0, | |
| "reward": 0.0625, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.0625, | |
| "step": 41 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 553.9375, | |
| "epoch": 0.08255528255528255, | |
| "grad_norm": 0.2940020586639228, | |
| "kl": 0.01055145263671875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0098, | |
| "num_tokens": 774304.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 42 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 506.78125, | |
| "epoch": 0.08452088452088452, | |
| "grad_norm": 0.34987523935436315, | |
| "kl": 0.022228240966796875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0136, | |
| "num_tokens": 794573.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 43 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 534.71875, | |
| "epoch": 0.08648648648648649, | |
| "grad_norm": 0.28775109044810143, | |
| "kl": 0.016704559326171875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.022, | |
| "num_tokens": 814952.0, | |
| "reward": 0.0625, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.0625, | |
| "step": 44 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 551.125, | |
| "epoch": 0.08845208845208845, | |
| "grad_norm": 0.5294418172877363, | |
| "kl": 0.01171875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0127, | |
| "num_tokens": 837090.0, | |
| "reward": 0.0625, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.0625, | |
| "step": 45 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 424.09375, | |
| "epoch": 0.09041769041769042, | |
| "grad_norm": 0.7442762340799661, | |
| "kl": 0.0166015625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0527, | |
| "num_tokens": 855063.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 46 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 578.90625, | |
| "epoch": 0.09238329238329239, | |
| "grad_norm": 0.9964430347064654, | |
| "kl": 0.01439666748046875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0347, | |
| "num_tokens": 876960.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.1875, | |
| "step": 47 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 516.65625, | |
| "epoch": 0.09434889434889435, | |
| "grad_norm": 0.0029736329333483326, | |
| "kl": 0.015178680419921875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 897477.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 48 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 406.53125, | |
| "epoch": 0.09631449631449632, | |
| "grad_norm": 0.42366305164369583, | |
| "kl": 0.0214385986328125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.023, | |
| "num_tokens": 913848.0, | |
| "reward": 0.0625, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.0625, | |
| "step": 49 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 615.46875, | |
| "epoch": 0.09828009828009827, | |
| "grad_norm": 0.2894821606425995, | |
| "kl": 0.010334014892578125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0008, | |
| "num_tokens": 936513.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 50 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 484.40625, | |
| "epoch": 0.10024570024570024, | |
| "grad_norm": 0.8466055231288873, | |
| "kl": 0.01735687255859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0053, | |
| "num_tokens": 955662.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 51 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 603.40625, | |
| "epoch": 0.10221130221130222, | |
| "grad_norm": 0.4594988801570104, | |
| "kl": 0.01914215087890625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0157, | |
| "num_tokens": 978603.0, | |
| "reward": 0.0625, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.0625, | |
| "step": 52 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 571.53125, | |
| "epoch": 0.10417690417690417, | |
| "grad_norm": 0.6301487761037352, | |
| "kl": 0.01680755615234375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0518, | |
| "num_tokens": 1000414.0, | |
| "reward": 0.125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.125, | |
| "step": 53 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 480.625, | |
| "epoch": 0.10614250614250614, | |
| "grad_norm": 0.1705061403375656, | |
| "kl": 0.019412994384765625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0151, | |
| "num_tokens": 1019984.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 54 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 586.5625, | |
| "epoch": 0.10810810810810811, | |
| "grad_norm": 0.3563070736118851, | |
| "kl": 0.01416778564453125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0113, | |
| "num_tokens": 1043134.0, | |
| "reward": 0.03125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.03125, | |
| "step": 55 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 573.78125, | |
| "epoch": 0.11007371007371007, | |
| "grad_norm": 0.8170380900607339, | |
| "kl": 0.0185394287109375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0107, | |
| "num_tokens": 1065473.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 56 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 596.25, | |
| "epoch": 0.11203931203931204, | |
| "grad_norm": 0.6138181155074318, | |
| "kl": 0.01763916015625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0173, | |
| "num_tokens": 1088233.0, | |
| "reward": 0.0625, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.0625, | |
| "step": 57 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 579.96875, | |
| "epoch": 0.114004914004914, | |
| "grad_norm": 0.5592170560306685, | |
| "kl": 0.015392303466796875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.042, | |
| "num_tokens": 1111490.0, | |
| "reward": 0.125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.125, | |
| "step": 58 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 590.3125, | |
| "epoch": 0.11597051597051597, | |
| "grad_norm": 0.004534498568535779, | |
| "kl": 0.017578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 1134000.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.125, | |
| "step": 59 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 552.75, | |
| "epoch": 0.11793611793611794, | |
| "grad_norm": 0.26122008088795556, | |
| "kl": 0.019062042236328125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 1155236.0, | |
| "reward": 0.09375, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.09375, | |
| "step": 60 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 533.46875, | |
| "epoch": 0.1199017199017199, | |
| "grad_norm": 0.6246443042334695, | |
| "kl": 0.01920318603515625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0014, | |
| "num_tokens": 1175617.0, | |
| "reward": 0.09375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.09375, | |
| "step": 61 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 484.34375, | |
| "epoch": 0.12186732186732187, | |
| "grad_norm": 1.0553043786365885, | |
| "kl": 0.0269622802734375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0461, | |
| "num_tokens": 1194332.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 62 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 604.84375, | |
| "epoch": 0.12383292383292384, | |
| "grad_norm": 0.5614007392931322, | |
| "kl": 0.0162506103515625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0098, | |
| "num_tokens": 1217363.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.1875, | |
| "step": 63 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 579.59375, | |
| "epoch": 0.1257985257985258, | |
| "grad_norm": 0.001324981221462767, | |
| "kl": 0.0164794921875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 1239884.0, | |
| "reward": 0.0625, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0625, | |
| "step": 64 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 643.0, | |
| "epoch": 0.12776412776412777, | |
| "grad_norm": 0.6061011991202276, | |
| "kl": 0.01563262939453125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.065, | |
| "num_tokens": 1263644.0, | |
| "reward": 0.125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.125, | |
| "step": 65 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 487.875, | |
| "epoch": 0.12972972972972974, | |
| "grad_norm": 0.8366609324695296, | |
| "kl": 0.02077484130859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0478, | |
| "num_tokens": 1282820.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.15625, | |
| "step": 66 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 662.59375, | |
| "epoch": 0.1316953316953317, | |
| "grad_norm": 0.17490281761649612, | |
| "kl": 0.0147705078125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0204, | |
| "num_tokens": 1307457.0, | |
| "reward": 0.09375, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.09375, | |
| "step": 67 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 525.03125, | |
| "epoch": 0.13366093366093365, | |
| "grad_norm": 0.8215530702995869, | |
| "kl": 0.020721435546875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.064, | |
| "num_tokens": 1327798.0, | |
| "reward": 0.25, | |
| "reward_std": 0.3535533845424652, | |
| "rewards/reward_func": 0.25, | |
| "step": 68 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 567.1875, | |
| "epoch": 0.13562653562653562, | |
| "grad_norm": 0.65089823105939, | |
| "kl": 0.019927978515625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0026, | |
| "num_tokens": 1350334.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.15625, | |
| "step": 69 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 584.78125, | |
| "epoch": 0.1375921375921376, | |
| "grad_norm": 1.008422919649498, | |
| "kl": 0.01912689208984375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0025, | |
| "num_tokens": 1372311.0, | |
| "reward": 0.25, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.25, | |
| "step": 70 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 545.125, | |
| "epoch": 0.13955773955773956, | |
| "grad_norm": 0.6752475010590075, | |
| "kl": 0.01819610595703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0177, | |
| "num_tokens": 1393645.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 71 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 526.40625, | |
| "epoch": 0.14152334152334153, | |
| "grad_norm": 0.7868187210048034, | |
| "kl": 0.0257568359375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0076, | |
| "num_tokens": 1414260.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 72 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 465.46875, | |
| "epoch": 0.14348894348894348, | |
| "grad_norm": 0.40424658361331584, | |
| "kl": 0.02496337890625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0034, | |
| "num_tokens": 1432617.0, | |
| "reward": 0.09375, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.09375, | |
| "step": 73 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 523.0, | |
| "epoch": 0.14545454545454545, | |
| "grad_norm": 0.7217516854340651, | |
| "kl": 0.0191497802734375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0438, | |
| "num_tokens": 1452631.0, | |
| "reward": 0.125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.125, | |
| "step": 74 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 564.59375, | |
| "epoch": 0.14742014742014742, | |
| "grad_norm": 1.020886020177138, | |
| "kl": 0.02172088623046875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0443, | |
| "num_tokens": 1474180.0, | |
| "reward": 0.25, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.25, | |
| "step": 75 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 599.03125, | |
| "epoch": 0.1493857493857494, | |
| "grad_norm": 0.38779629766219476, | |
| "kl": 0.021942138671875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0253, | |
| "num_tokens": 1496597.0, | |
| "reward": 0.0625, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.0625, | |
| "step": 76 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 535.65625, | |
| "epoch": 0.15135135135135136, | |
| "grad_norm": 0.5231239100329028, | |
| "kl": 0.0213470458984375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0072, | |
| "num_tokens": 1518002.0, | |
| "reward": 0.09375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.09375, | |
| "step": 77 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 483.28125, | |
| "epoch": 0.15331695331695333, | |
| "grad_norm": 1.01964836669152, | |
| "kl": 0.02301025390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0253, | |
| "num_tokens": 1538251.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.1875, | |
| "step": 78 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 597.21875, | |
| "epoch": 0.15528255528255527, | |
| "grad_norm": 0.3894010167090417, | |
| "kl": 0.01868438720703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0049, | |
| "num_tokens": 1561066.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 79 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 494.21875, | |
| "epoch": 0.15724815724815724, | |
| "grad_norm": 0.4559857762776204, | |
| "kl": 0.02178955078125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0144, | |
| "num_tokens": 1579851.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 80 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 459.4375, | |
| "epoch": 0.15921375921375922, | |
| "grad_norm": 0.7408328801166559, | |
| "kl": 0.02642822265625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0034, | |
| "num_tokens": 1597675.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 81 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 601.90625, | |
| "epoch": 0.16117936117936119, | |
| "grad_norm": 0.4202464041477536, | |
| "kl": 0.01740264892578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0084, | |
| "num_tokens": 1621632.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 82 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 517.21875, | |
| "epoch": 0.16314496314496316, | |
| "grad_norm": 0.7785902853786559, | |
| "kl": 0.02386474609375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0036, | |
| "num_tokens": 1641873.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 83 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 390.6875, | |
| "epoch": 0.1651105651105651, | |
| "grad_norm": 0.6791841008394204, | |
| "kl": 0.034637451171875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0184, | |
| "num_tokens": 1658453.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.1875, | |
| "step": 84 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 485.78125, | |
| "epoch": 0.16707616707616707, | |
| "grad_norm": 0.4367719984766766, | |
| "kl": 0.0251617431640625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0156, | |
| "num_tokens": 1676992.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 85 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 528.875, | |
| "epoch": 0.16904176904176904, | |
| "grad_norm": 0.7222029349504244, | |
| "kl": 0.02223968505859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0459, | |
| "num_tokens": 1700584.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 86 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 658.125, | |
| "epoch": 0.171007371007371, | |
| "grad_norm": 0.34834776108087606, | |
| "kl": 0.01824951171875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0326, | |
| "num_tokens": 1725524.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 87 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 644.5625, | |
| "epoch": 0.17297297297297298, | |
| "grad_norm": 0.6750518830704131, | |
| "kl": 0.01828765869140625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0412, | |
| "num_tokens": 1749556.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.15625, | |
| "step": 88 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 547.15625, | |
| "epoch": 0.17493857493857493, | |
| "grad_norm": 0.31981853904702845, | |
| "kl": 0.01940155029296875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.012, | |
| "num_tokens": 1770175.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.1875, | |
| "step": 89 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 475.75, | |
| "epoch": 0.1769041769041769, | |
| "grad_norm": 1.176119812297408, | |
| "kl": 0.0265045166015625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0174, | |
| "num_tokens": 1789467.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.1875, | |
| "step": 90 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 448.625, | |
| "epoch": 0.17886977886977887, | |
| "grad_norm": 0.7247013137515755, | |
| "kl": 0.02973175048828125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0025, | |
| "num_tokens": 1807295.0, | |
| "reward": 0.09375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.09375, | |
| "step": 91 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 461.84375, | |
| "epoch": 0.18083538083538084, | |
| "grad_norm": 0.9576563539486035, | |
| "kl": 0.027984619140625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0572, | |
| "num_tokens": 1827104.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 92 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 549.375, | |
| "epoch": 0.1828009828009828, | |
| "grad_norm": 0.8080609769164848, | |
| "kl": 0.02375030517578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0641, | |
| "num_tokens": 1848272.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 93 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 673.53125, | |
| "epoch": 0.18476658476658478, | |
| "grad_norm": 0.6070813087390035, | |
| "kl": 0.0193634033203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0232, | |
| "num_tokens": 1873533.0, | |
| "reward": 0.125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.125, | |
| "step": 94 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 582.75, | |
| "epoch": 0.18673218673218672, | |
| "grad_norm": 0.6479257231531951, | |
| "kl": 0.0185394287109375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0017, | |
| "num_tokens": 1895601.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 95 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 443.5, | |
| "epoch": 0.1886977886977887, | |
| "grad_norm": 0.7375270346991556, | |
| "kl": 0.0291290283203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0343, | |
| "num_tokens": 1912993.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 96 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 499.125, | |
| "epoch": 0.19066339066339066, | |
| "grad_norm": 0.6426005454499798, | |
| "kl": 0.0281524658203125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0165, | |
| "num_tokens": 1933631.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 97 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 583.21875, | |
| "epoch": 0.19262899262899263, | |
| "grad_norm": 0.656305823795339, | |
| "kl": 0.0197601318359375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0098, | |
| "num_tokens": 1955588.0, | |
| "reward": 0.125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.125, | |
| "step": 98 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 604.5, | |
| "epoch": 0.1945945945945946, | |
| "grad_norm": 0.7159925904721693, | |
| "kl": 0.0202484130859375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0024, | |
| "num_tokens": 1978466.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.15625, | |
| "step": 99 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 623.8125, | |
| "epoch": 0.19656019656019655, | |
| "grad_norm": 0.6011532165345448, | |
| "kl": 0.0197296142578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.015, | |
| "num_tokens": 2001554.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.1875, | |
| "step": 100 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 487.84375, | |
| "epoch": 0.19852579852579852, | |
| "grad_norm": 0.6696267166180018, | |
| "kl": 0.02446746826171875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0323, | |
| "num_tokens": 2020681.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 101 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 512.90625, | |
| "epoch": 0.2004914004914005, | |
| "grad_norm": 0.5144645053944626, | |
| "kl": 0.0222320556640625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0308, | |
| "num_tokens": 2040058.0, | |
| "reward": 0.0625, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.0625, | |
| "step": 102 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 492.59375, | |
| "epoch": 0.20245700245700246, | |
| "grad_norm": 0.9792203746023931, | |
| "kl": 0.02777099609375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0185, | |
| "num_tokens": 2059481.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 103 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 540.5625, | |
| "epoch": 0.20442260442260443, | |
| "grad_norm": 0.6773524274596018, | |
| "kl": 0.0245513916015625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0115, | |
| "num_tokens": 2079795.0, | |
| "reward": 0.125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.125, | |
| "step": 104 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 629.75, | |
| "epoch": 0.20638820638820637, | |
| "grad_norm": 0.0014408090990997875, | |
| "kl": 0.020843505859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 2104281.0, | |
| "reward": 0.0, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.0, | |
| "step": 105 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 489.625, | |
| "epoch": 0.20835380835380835, | |
| "grad_norm": 0.43577632912535613, | |
| "kl": 0.02980804443359375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0265, | |
| "num_tokens": 2124275.0, | |
| "reward": 0.25, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.25, | |
| "step": 106 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 480.90625, | |
| "epoch": 0.21031941031941032, | |
| "grad_norm": 0.5136803725458018, | |
| "kl": 0.02611541748046875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0184, | |
| "num_tokens": 2143080.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 107 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 659.9375, | |
| "epoch": 0.2122850122850123, | |
| "grad_norm": 0.4897037723304078, | |
| "kl": 0.0196380615234375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0062, | |
| "num_tokens": 2167550.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 108 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 549.53125, | |
| "epoch": 0.21425061425061426, | |
| "grad_norm": 0.6823729897684875, | |
| "kl": 0.020660400390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0027, | |
| "num_tokens": 2188815.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.1875, | |
| "step": 109 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 492.4375, | |
| "epoch": 0.21621621621621623, | |
| "grad_norm": 0.8520053469926316, | |
| "kl": 0.02719879150390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0077, | |
| "num_tokens": 2207977.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 110 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 487.34375, | |
| "epoch": 0.21818181818181817, | |
| "grad_norm": 0.8983378201666078, | |
| "kl": 0.024871826171875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0207, | |
| "num_tokens": 2226818.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.40625, | |
| "step": 111 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 534.71875, | |
| "epoch": 0.22014742014742014, | |
| "grad_norm": 0.6341326642846564, | |
| "kl": 0.022857666015625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.005, | |
| "num_tokens": 2247351.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 112 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 509.96875, | |
| "epoch": 0.2221130221130221, | |
| "grad_norm": 0.5084014496191064, | |
| "kl": 0.02394866943359375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0121, | |
| "num_tokens": 2267646.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.1875, | |
| "step": 113 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 493.46875, | |
| "epoch": 0.22407862407862408, | |
| "grad_norm": 0.47081172119878306, | |
| "kl": 0.023681640625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.003, | |
| "num_tokens": 2287461.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 114 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 593.46875, | |
| "epoch": 0.22604422604422605, | |
| "grad_norm": 0.5402745390200868, | |
| "kl": 0.0226593017578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0254, | |
| "num_tokens": 2310300.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 115 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 659.34375, | |
| "epoch": 0.228009828009828, | |
| "grad_norm": 0.7384552819583179, | |
| "kl": 0.01863861083984375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.011, | |
| "num_tokens": 2336289.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 116 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 426.0625, | |
| "epoch": 0.22997542997542997, | |
| "grad_norm": 0.9856793371703316, | |
| "kl": 0.029998779296875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0107, | |
| "num_tokens": 2353069.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.1875, | |
| "step": 117 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 553.15625, | |
| "epoch": 0.23194103194103194, | |
| "grad_norm": 0.6524109632549928, | |
| "kl": 0.0211334228515625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0166, | |
| "num_tokens": 2374500.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 118 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 495.96875, | |
| "epoch": 0.2339066339066339, | |
| "grad_norm": 0.750164710998597, | |
| "kl": 0.0268707275390625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0291, | |
| "num_tokens": 2393455.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 119 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 519.375, | |
| "epoch": 0.23587223587223588, | |
| "grad_norm": 0.8136338780518924, | |
| "kl": 0.0233917236328125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0213, | |
| "num_tokens": 2413821.0, | |
| "reward": 0.25, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.25, | |
| "step": 120 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 492.5, | |
| "epoch": 0.23783783783783785, | |
| "grad_norm": 0.9472454617054751, | |
| "kl": 0.0245513916015625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0434, | |
| "num_tokens": 2432727.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.3535533845424652, | |
| "rewards/reward_func": 0.3125, | |
| "step": 121 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 650.125, | |
| "epoch": 0.2398034398034398, | |
| "grad_norm": 0.7941644508393335, | |
| "kl": 0.0188446044921875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0161, | |
| "num_tokens": 2457059.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 122 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 505.40625, | |
| "epoch": 0.24176904176904176, | |
| "grad_norm": 0.9756007077675318, | |
| "kl": 0.02396392822265625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0303, | |
| "num_tokens": 2477004.0, | |
| "reward": 0.53125, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.53125, | |
| "step": 123 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 483.96875, | |
| "epoch": 0.24373464373464374, | |
| "grad_norm": 0.6941208484996456, | |
| "kl": 0.027313232421875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0341, | |
| "num_tokens": 2495959.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.40625, | |
| "step": 124 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 561.3125, | |
| "epoch": 0.2457002457002457, | |
| "grad_norm": 0.5265733548592475, | |
| "kl": 0.01922607421875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0423, | |
| "num_tokens": 2517637.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.1875, | |
| "step": 125 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 492.96875, | |
| "epoch": 0.24766584766584768, | |
| "grad_norm": 0.8350574297158757, | |
| "kl": 0.023712158203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0469, | |
| "num_tokens": 2538338.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 126 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 573.5, | |
| "epoch": 0.24963144963144962, | |
| "grad_norm": 0.8216738883653201, | |
| "kl": 0.02204132080078125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0218, | |
| "num_tokens": 2560112.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.28125, | |
| "step": 127 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 529.3125, | |
| "epoch": 0.2515970515970516, | |
| "grad_norm": 0.78248285660239, | |
| "kl": 0.02326202392578125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.044, | |
| "num_tokens": 2580546.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 128 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 602.1875, | |
| "epoch": 0.25356265356265356, | |
| "grad_norm": 0.2772418236153871, | |
| "kl": 0.019317626953125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0027, | |
| "num_tokens": 2604460.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.15625, | |
| "step": 129 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 548.0625, | |
| "epoch": 0.25552825552825553, | |
| "grad_norm": 0.6625525706905956, | |
| "kl": 0.02095794677734375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0115, | |
| "num_tokens": 2624972.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 130 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 585.9375, | |
| "epoch": 0.2574938574938575, | |
| "grad_norm": 0.752169210277677, | |
| "kl": 0.021514892578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0359, | |
| "num_tokens": 2648150.0, | |
| "reward": 0.375, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.375, | |
| "step": 131 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 574.0625, | |
| "epoch": 0.2594594594594595, | |
| "grad_norm": 0.5002363204112915, | |
| "kl": 0.0212249755859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.023, | |
| "num_tokens": 2670338.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 132 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 526.625, | |
| "epoch": 0.26142506142506144, | |
| "grad_norm": 0.9674758127355131, | |
| "kl": 0.0261077880859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0159, | |
| "num_tokens": 2691642.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.34375, | |
| "step": 133 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 412.59375, | |
| "epoch": 0.2633906633906634, | |
| "grad_norm": 1.0324181511597994, | |
| "kl": 0.0365447998046875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0109, | |
| "num_tokens": 2708471.0, | |
| "reward": 0.53125, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.53125, | |
| "step": 134 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 501.1875, | |
| "epoch": 0.26535626535626533, | |
| "grad_norm": 0.646085784140391, | |
| "kl": 0.02829742431640625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0245, | |
| "num_tokens": 2728631.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 135 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 524.59375, | |
| "epoch": 0.2673218673218673, | |
| "grad_norm": 0.5827914647557386, | |
| "kl": 0.0233001708984375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0133, | |
| "num_tokens": 2748682.0, | |
| "reward": 0.375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.375, | |
| "step": 136 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 493.625, | |
| "epoch": 0.26928746928746927, | |
| "grad_norm": 0.5383283228577356, | |
| "kl": 0.02911376953125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.006, | |
| "num_tokens": 2768384.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.15625, | |
| "step": 137 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 563.34375, | |
| "epoch": 0.27125307125307124, | |
| "grad_norm": 0.44514691751088103, | |
| "kl": 0.02154541015625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.045, | |
| "num_tokens": 2790263.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 138 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 520.8125, | |
| "epoch": 0.2732186732186732, | |
| "grad_norm": 0.7190292298516152, | |
| "kl": 0.0236663818359375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0132, | |
| "num_tokens": 2810351.0, | |
| "reward": 0.375, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.375, | |
| "step": 139 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 527.34375, | |
| "epoch": 0.2751842751842752, | |
| "grad_norm": 0.7242965068485637, | |
| "kl": 0.0230255126953125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.001, | |
| "num_tokens": 2830716.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 140 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 565.8125, | |
| "epoch": 0.27714987714987716, | |
| "grad_norm": 0.6228089750555335, | |
| "kl": 0.02398681640625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0041, | |
| "num_tokens": 2853810.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 141 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 555.1875, | |
| "epoch": 0.2791154791154791, | |
| "grad_norm": 0.37035042072698693, | |
| "kl": 0.02410888671875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0047, | |
| "num_tokens": 2875056.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 142 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 589.65625, | |
| "epoch": 0.2810810810810811, | |
| "grad_norm": 0.658035150283911, | |
| "kl": 0.02394866943359375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.025, | |
| "num_tokens": 2897357.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.4375, | |
| "step": 143 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 543.8125, | |
| "epoch": 0.28304668304668307, | |
| "grad_norm": 0.7501331474001123, | |
| "kl": 0.025238037109375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0123, | |
| "num_tokens": 2917753.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 144 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 612.28125, | |
| "epoch": 0.28501228501228504, | |
| "grad_norm": 0.5446469883013957, | |
| "kl": 0.0233612060546875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0054, | |
| "num_tokens": 2940774.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 145 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 539.78125, | |
| "epoch": 0.28697788697788695, | |
| "grad_norm": 0.6092756185830317, | |
| "kl": 0.02280426025390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0263, | |
| "num_tokens": 2961247.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 146 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 466.28125, | |
| "epoch": 0.2889434889434889, | |
| "grad_norm": 0.6941889736363939, | |
| "kl": 0.0313873291015625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0046, | |
| "num_tokens": 2979622.0, | |
| "reward": 0.09375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.09375, | |
| "step": 147 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 567.53125, | |
| "epoch": 0.2909090909090909, | |
| "grad_norm": 0.3736888500181184, | |
| "kl": 0.023223876953125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0099, | |
| "num_tokens": 3002039.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 148 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 467.59375, | |
| "epoch": 0.29287469287469287, | |
| "grad_norm": 0.8319282300423898, | |
| "kl": 0.02935791015625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.043, | |
| "num_tokens": 3019816.0, | |
| "reward": 0.375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.375, | |
| "step": 149 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 492.25, | |
| "epoch": 0.29484029484029484, | |
| "grad_norm": 0.7325276211486558, | |
| "kl": 0.0263671875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0293, | |
| "num_tokens": 3039146.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.46875, | |
| "step": 150 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 592.1875, | |
| "epoch": 0.2968058968058968, | |
| "grad_norm": 0.43633918201867017, | |
| "kl": 0.0228424072265625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0122, | |
| "num_tokens": 3062102.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 151 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 563.1875, | |
| "epoch": 0.2987714987714988, | |
| "grad_norm": 0.38992921040393713, | |
| "kl": 0.025421142578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0327, | |
| "num_tokens": 3084284.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 152 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 634.9375, | |
| "epoch": 0.30073710073710075, | |
| "grad_norm": 0.41943188112426816, | |
| "kl": 0.0197906494140625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.001, | |
| "num_tokens": 3108216.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.1875, | |
| "step": 153 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 636.9375, | |
| "epoch": 0.3027027027027027, | |
| "grad_norm": 0.19968498604663434, | |
| "kl": 0.02194976806640625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0001, | |
| "num_tokens": 3133100.0, | |
| "reward": 0.09375, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.09375, | |
| "step": 154 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 568.21875, | |
| "epoch": 0.3046683046683047, | |
| "grad_norm": 0.7545074739064522, | |
| "kl": 0.02685546875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.04, | |
| "num_tokens": 3154579.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.46875, | |
| "step": 155 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 529.8125, | |
| "epoch": 0.30663390663390666, | |
| "grad_norm": 0.6884386708248261, | |
| "kl": 0.025360107421875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0075, | |
| "num_tokens": 3174845.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.40625, | |
| "step": 156 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 513.0625, | |
| "epoch": 0.3085995085995086, | |
| "grad_norm": 0.5265944726858988, | |
| "kl": 0.0264892578125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0245, | |
| "num_tokens": 3195631.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 157 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 520.0, | |
| "epoch": 0.31056511056511055, | |
| "grad_norm": 0.45924555811174794, | |
| "kl": 0.026092529296875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0189, | |
| "num_tokens": 3216143.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 158 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 568.15625, | |
| "epoch": 0.3125307125307125, | |
| "grad_norm": 0.5301940769841443, | |
| "kl": 0.025177001953125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0008, | |
| "num_tokens": 3238264.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 159 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 508.59375, | |
| "epoch": 0.3144963144963145, | |
| "grad_norm": 0.8755698562076372, | |
| "kl": 0.0290069580078125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0536, | |
| "num_tokens": 3257837.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 160 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 481.15625, | |
| "epoch": 0.31646191646191646, | |
| "grad_norm": 0.6081291396581243, | |
| "kl": 0.0274810791015625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0018, | |
| "num_tokens": 3276820.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 161 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 598.28125, | |
| "epoch": 0.31842751842751843, | |
| "grad_norm": 0.5215378127068343, | |
| "kl": 0.0220794677734375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0012, | |
| "num_tokens": 3299357.0, | |
| "reward": 0.375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.375, | |
| "step": 162 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 487.09375, | |
| "epoch": 0.3203931203931204, | |
| "grad_norm": 0.5432850798950747, | |
| "kl": 0.03083038330078125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0034, | |
| "num_tokens": 3318084.0, | |
| "reward": 0.5, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.5, | |
| "step": 163 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 531.125, | |
| "epoch": 0.32235872235872237, | |
| "grad_norm": 0.46027668679847555, | |
| "kl": 0.026092529296875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0154, | |
| "num_tokens": 3338328.0, | |
| "reward": 0.125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.125, | |
| "step": 164 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 624.8125, | |
| "epoch": 0.32432432432432434, | |
| "grad_norm": 0.18474075838750667, | |
| "kl": 0.02459716796875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0011, | |
| "num_tokens": 3361788.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.21875, | |
| "step": 165 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 549.78125, | |
| "epoch": 0.3262899262899263, | |
| "grad_norm": 0.7718886882448309, | |
| "kl": 0.027069091796875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.005, | |
| "num_tokens": 3382981.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.46875, | |
| "step": 166 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 575.78125, | |
| "epoch": 0.32825552825552823, | |
| "grad_norm": 0.5895346584970902, | |
| "kl": 0.02323150634765625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0178, | |
| "num_tokens": 3405274.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 167 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 576.875, | |
| "epoch": 0.3302211302211302, | |
| "grad_norm": 0.4959807619539926, | |
| "kl": 0.0241851806640625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0022, | |
| "num_tokens": 3427392.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 168 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 561.78125, | |
| "epoch": 0.33218673218673217, | |
| "grad_norm": 0.8681759456851222, | |
| "kl": 0.0249176025390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0274, | |
| "num_tokens": 3448681.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.34375, | |
| "step": 169 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 483.53125, | |
| "epoch": 0.33415233415233414, | |
| "grad_norm": 0.5297566985301704, | |
| "kl": 0.0336761474609375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0019, | |
| "num_tokens": 3467212.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 170 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 581.40625, | |
| "epoch": 0.3361179361179361, | |
| "grad_norm": 0.6192719333892511, | |
| "kl": 0.025787353515625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0075, | |
| "num_tokens": 3489355.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 171 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 709.09375, | |
| "epoch": 0.3380835380835381, | |
| "grad_norm": 0.47423370876978443, | |
| "kl": 0.0204620361328125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0251, | |
| "num_tokens": 3515834.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 172 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 614.21875, | |
| "epoch": 0.34004914004914005, | |
| "grad_norm": 0.6479887044288859, | |
| "kl": 0.0229034423828125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0212, | |
| "num_tokens": 3539067.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.3125, | |
| "step": 173 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 552.6875, | |
| "epoch": 0.342014742014742, | |
| "grad_norm": 0.4712074892004097, | |
| "kl": 0.0272216796875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0209, | |
| "num_tokens": 3560435.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 174 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 599.75, | |
| "epoch": 0.343980343980344, | |
| "grad_norm": 0.7325904251977917, | |
| "kl": 0.0259552001953125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0087, | |
| "num_tokens": 3582617.0, | |
| "reward": 0.25, | |
| "reward_std": 0.3535533845424652, | |
| "rewards/reward_func": 0.25, | |
| "step": 175 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 589.46875, | |
| "epoch": 0.34594594594594597, | |
| "grad_norm": 0.718404368627785, | |
| "kl": 0.024017333984375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0069, | |
| "num_tokens": 3606486.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 176 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 525.0625, | |
| "epoch": 0.34791154791154794, | |
| "grad_norm": 0.5807447821668908, | |
| "kl": 0.0262451171875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.006, | |
| "num_tokens": 3626528.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 177 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 539.875, | |
| "epoch": 0.34987714987714985, | |
| "grad_norm": 0.3836295461022733, | |
| "kl": 0.025665283203125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0233, | |
| "num_tokens": 3647890.0, | |
| "reward": 0.25, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.25, | |
| "step": 178 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 641.28125, | |
| "epoch": 0.3518427518427518, | |
| "grad_norm": 0.3325268331032711, | |
| "kl": 0.02361297607421875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0117, | |
| "num_tokens": 3672977.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.1875, | |
| "step": 179 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 611.1875, | |
| "epoch": 0.3538083538083538, | |
| "grad_norm": 0.5612204711943589, | |
| "kl": 0.025909423828125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.014, | |
| "num_tokens": 3696201.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 180 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 587.125, | |
| "epoch": 0.35577395577395576, | |
| "grad_norm": 0.48905557864261767, | |
| "kl": 0.0284881591796875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0135, | |
| "num_tokens": 3718645.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 181 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 532.8125, | |
| "epoch": 0.35773955773955773, | |
| "grad_norm": 0.6196793685491477, | |
| "kl": 0.0283050537109375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.021, | |
| "num_tokens": 3739419.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 182 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 491.25, | |
| "epoch": 0.3597051597051597, | |
| "grad_norm": 0.7943049071356907, | |
| "kl": 0.0270843505859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0343, | |
| "num_tokens": 3758175.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 183 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 629.3125, | |
| "epoch": 0.3616707616707617, | |
| "grad_norm": 0.4358875914582024, | |
| "kl": 0.02367401123046875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0133, | |
| "num_tokens": 3782789.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 184 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 540.59375, | |
| "epoch": 0.36363636363636365, | |
| "grad_norm": 0.7097269605040931, | |
| "kl": 0.028656005859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.023, | |
| "num_tokens": 3804202.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 185 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 517.40625, | |
| "epoch": 0.3656019656019656, | |
| "grad_norm": 0.8226505283368968, | |
| "kl": 0.0278778076171875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0374, | |
| "num_tokens": 3824813.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.3535533845424652, | |
| "rewards/reward_func": 0.3125, | |
| "step": 186 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 499.3125, | |
| "epoch": 0.3675675675675676, | |
| "grad_norm": 0.8593511870826509, | |
| "kl": 0.026092529296875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0443, | |
| "num_tokens": 3844419.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.3125, | |
| "step": 187 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 497.5, | |
| "epoch": 0.36953316953316956, | |
| "grad_norm": 0.6026343875240253, | |
| "kl": 0.0312042236328125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0226, | |
| "num_tokens": 3863979.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 188 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 547.0625, | |
| "epoch": 0.3714987714987715, | |
| "grad_norm": 0.6577878892956752, | |
| "kl": 0.030914306640625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.002, | |
| "num_tokens": 3885407.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 189 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 605.0625, | |
| "epoch": 0.37346437346437344, | |
| "grad_norm": 0.45127850046328344, | |
| "kl": 0.024169921875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.003, | |
| "num_tokens": 3908713.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.40625, | |
| "step": 190 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 529.46875, | |
| "epoch": 0.3754299754299754, | |
| "grad_norm": 0.5469774623716551, | |
| "kl": 0.025115966796875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0056, | |
| "num_tokens": 3928856.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 191 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 475.15625, | |
| "epoch": 0.3773955773955774, | |
| "grad_norm": 0.9374126682104605, | |
| "kl": 0.0326385498046875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0543, | |
| "num_tokens": 3947241.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.3125, | |
| "step": 192 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 547.1875, | |
| "epoch": 0.37936117936117936, | |
| "grad_norm": 0.5527175997380437, | |
| "kl": 0.0286407470703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0571, | |
| "num_tokens": 3968765.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 193 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 581.53125, | |
| "epoch": 0.38132678132678133, | |
| "grad_norm": 0.5091715028558698, | |
| "kl": 0.02716064453125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0032, | |
| "num_tokens": 3990744.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 194 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 545.3125, | |
| "epoch": 0.3832923832923833, | |
| "grad_norm": 0.6906699369220881, | |
| "kl": 0.026611328125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0317, | |
| "num_tokens": 4011242.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 195 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 553.0625, | |
| "epoch": 0.38525798525798527, | |
| "grad_norm": 0.803923186478945, | |
| "kl": 0.026153564453125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0114, | |
| "num_tokens": 4032454.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.34375, | |
| "step": 196 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 514.875, | |
| "epoch": 0.38722358722358724, | |
| "grad_norm": 0.5425045400963313, | |
| "kl": 0.029937744140625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0032, | |
| "num_tokens": 4053082.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 197 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 677.78125, | |
| "epoch": 0.3891891891891892, | |
| "grad_norm": 0.6672952970651821, | |
| "kl": 0.0213623046875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0023, | |
| "num_tokens": 4078691.0, | |
| "reward": 0.125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.125, | |
| "step": 198 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 512.03125, | |
| "epoch": 0.3911547911547912, | |
| "grad_norm": 0.476359304779961, | |
| "kl": 0.035552978515625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0066, | |
| "num_tokens": 4099720.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.21875, | |
| "step": 199 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 743.375, | |
| "epoch": 0.3931203931203931, | |
| "grad_norm": 0.4679126082574762, | |
| "kl": 0.0198974609375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0029, | |
| "num_tokens": 4127002.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 200 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 557.0, | |
| "epoch": 0.39508599508599507, | |
| "grad_norm": 0.6988574973859861, | |
| "kl": 0.028167724609375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0327, | |
| "num_tokens": 4148548.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 201 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 566.28125, | |
| "epoch": 0.39705159705159704, | |
| "grad_norm": 0.7768469504643993, | |
| "kl": 0.0265960693359375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0108, | |
| "num_tokens": 4170063.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 202 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 610.5, | |
| "epoch": 0.399017199017199, | |
| "grad_norm": 0.3681456638360366, | |
| "kl": 0.0287933349609375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0088, | |
| "num_tokens": 4194559.0, | |
| "reward": 0.25, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.25, | |
| "step": 203 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 639.34375, | |
| "epoch": 0.400982800982801, | |
| "grad_norm": 0.6149062975054262, | |
| "kl": 0.0237884521484375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0481, | |
| "num_tokens": 4218340.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.1875, | |
| "step": 204 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 524.15625, | |
| "epoch": 0.40294840294840295, | |
| "grad_norm": 0.7831991086164788, | |
| "kl": 0.0277557373046875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0474, | |
| "num_tokens": 4239031.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 205 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 553.5625, | |
| "epoch": 0.4049140049140049, | |
| "grad_norm": 0.7849973591038859, | |
| "kl": 0.0251617431640625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0378, | |
| "num_tokens": 4260199.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.40625, | |
| "step": 206 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 480.75, | |
| "epoch": 0.4068796068796069, | |
| "grad_norm": 0.6839800139372438, | |
| "kl": 0.03155517578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.016, | |
| "num_tokens": 4278919.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 207 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 606.34375, | |
| "epoch": 0.40884520884520886, | |
| "grad_norm": 0.3896782835395279, | |
| "kl": 0.026397705078125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0216, | |
| "num_tokens": 4302006.0, | |
| "reward": 0.25, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.25, | |
| "step": 208 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 629.21875, | |
| "epoch": 0.41081081081081083, | |
| "grad_norm": 0.5046764418165803, | |
| "kl": 0.0250244140625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0093, | |
| "num_tokens": 4325779.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 209 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 639.59375, | |
| "epoch": 0.41277641277641275, | |
| "grad_norm": 0.5479053649859522, | |
| "kl": 0.0309600830078125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.008, | |
| "num_tokens": 4349352.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 210 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 608.875, | |
| "epoch": 0.4147420147420147, | |
| "grad_norm": 0.27291739633494205, | |
| "kl": 0.025482177734375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.008, | |
| "num_tokens": 4372614.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.28125, | |
| "step": 211 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 700.3125, | |
| "epoch": 0.4167076167076167, | |
| "grad_norm": 0.6410316114692441, | |
| "kl": 0.0247802734375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0235, | |
| "num_tokens": 4398294.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 212 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 558.625, | |
| "epoch": 0.41867321867321866, | |
| "grad_norm": 0.30488486047826086, | |
| "kl": 0.028900146484375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0073, | |
| "num_tokens": 4419278.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.28125, | |
| "step": 213 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 454.71875, | |
| "epoch": 0.42063882063882063, | |
| "grad_norm": 0.6876906732910688, | |
| "kl": 0.0352935791015625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0201, | |
| "num_tokens": 4437525.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.4375, | |
| "step": 214 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 586.34375, | |
| "epoch": 0.4226044226044226, | |
| "grad_norm": 0.7070061651726116, | |
| "kl": 0.0324859619140625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0072, | |
| "num_tokens": 4459576.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.40625, | |
| "step": 215 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 598.25, | |
| "epoch": 0.4245700245700246, | |
| "grad_norm": 0.6961718312404855, | |
| "kl": 0.031036376953125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.025, | |
| "num_tokens": 4482428.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 216 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 550.5, | |
| "epoch": 0.42653562653562654, | |
| "grad_norm": 0.6664650861834673, | |
| "kl": 0.030609130859375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0165, | |
| "num_tokens": 4503774.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.40625, | |
| "step": 217 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 492.875, | |
| "epoch": 0.4285012285012285, | |
| "grad_norm": 0.6916569742061264, | |
| "kl": 0.0322113037109375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0245, | |
| "num_tokens": 4523210.0, | |
| "reward": 0.53125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.53125, | |
| "step": 218 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 556.125, | |
| "epoch": 0.4304668304668305, | |
| "grad_norm": 0.6079132124740041, | |
| "kl": 0.0360260009765625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0504, | |
| "num_tokens": 4544250.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 219 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 640.625, | |
| "epoch": 0.43243243243243246, | |
| "grad_norm": 0.49303925630539, | |
| "kl": 0.02655029296875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0042, | |
| "num_tokens": 4568206.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 220 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 591.15625, | |
| "epoch": 0.43439803439803437, | |
| "grad_norm": 0.3790789264426733, | |
| "kl": 0.026641845703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0029, | |
| "num_tokens": 4592127.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 221 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 559.5, | |
| "epoch": 0.43636363636363634, | |
| "grad_norm": 0.7205730486755546, | |
| "kl": 0.0282440185546875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0047, | |
| "num_tokens": 4613469.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.3125, | |
| "step": 222 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 616.78125, | |
| "epoch": 0.4383292383292383, | |
| "grad_norm": 0.2994169582864983, | |
| "kl": 0.0284271240234375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0124, | |
| "num_tokens": 4636780.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 223 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 546.96875, | |
| "epoch": 0.4402948402948403, | |
| "grad_norm": 0.6443329354884543, | |
| "kl": 0.03546142578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0477, | |
| "num_tokens": 4658003.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 224 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 691.59375, | |
| "epoch": 0.44226044226044225, | |
| "grad_norm": 0.5810139994156356, | |
| "kl": 0.0257415771484375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0137, | |
| "num_tokens": 4683758.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 225 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 654.46875, | |
| "epoch": 0.4442260442260442, | |
| "grad_norm": 0.4628887828349593, | |
| "kl": 0.0256500244140625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0144, | |
| "num_tokens": 4708083.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 226 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 480.71875, | |
| "epoch": 0.4461916461916462, | |
| "grad_norm": 0.5979661864229516, | |
| "kl": 0.0352325439453125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0019, | |
| "num_tokens": 4727130.0, | |
| "reward": 0.5625, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.5625, | |
| "step": 227 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 541.5, | |
| "epoch": 0.44815724815724817, | |
| "grad_norm": 0.5705252239874116, | |
| "kl": 0.0313720703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0057, | |
| "num_tokens": 4747992.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 228 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 612.46875, | |
| "epoch": 0.45012285012285014, | |
| "grad_norm": 0.5836545557559704, | |
| "kl": 0.027099609375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0091, | |
| "num_tokens": 4771251.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 229 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 637.8125, | |
| "epoch": 0.4520884520884521, | |
| "grad_norm": 0.4498303433751672, | |
| "kl": 0.0286102294921875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0034, | |
| "num_tokens": 4796355.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 230 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 610.625, | |
| "epoch": 0.4540540540540541, | |
| "grad_norm": 0.6013423970019107, | |
| "kl": 0.0283966064453125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0639, | |
| "num_tokens": 4819261.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.1875, | |
| "step": 231 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 584.0625, | |
| "epoch": 0.456019656019656, | |
| "grad_norm": 0.7699293584446985, | |
| "kl": 0.03173828125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0023, | |
| "num_tokens": 4841987.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.1875, | |
| "step": 232 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 617.21875, | |
| "epoch": 0.45798525798525797, | |
| "grad_norm": 0.5200942231005387, | |
| "kl": 0.078887939453125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0358, | |
| "num_tokens": 4865308.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 233 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 549.21875, | |
| "epoch": 0.45995085995085994, | |
| "grad_norm": 0.6584051899429707, | |
| "kl": 0.0289764404296875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0274, | |
| "num_tokens": 4886777.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 234 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 671.78125, | |
| "epoch": 0.4619164619164619, | |
| "grad_norm": 0.299051256935548, | |
| "kl": 0.0264892578125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0096, | |
| "num_tokens": 4911454.0, | |
| "reward": 0.25, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.25, | |
| "step": 235 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 571.28125, | |
| "epoch": 0.4638820638820639, | |
| "grad_norm": 0.6402318572716406, | |
| "kl": 0.03070068359375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0066, | |
| "num_tokens": 4933175.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 236 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 561.59375, | |
| "epoch": 0.46584766584766585, | |
| "grad_norm": 0.6512014891525182, | |
| "kl": 0.033782958984375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0072, | |
| "num_tokens": 4955004.0, | |
| "reward": 0.5, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.5, | |
| "step": 237 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 532.5625, | |
| "epoch": 0.4678132678132678, | |
| "grad_norm": 0.4328249392514584, | |
| "kl": 0.03594970703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0331, | |
| "num_tokens": 4976386.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.46875, | |
| "step": 238 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 584.03125, | |
| "epoch": 0.4697788697788698, | |
| "grad_norm": 0.4634554918945285, | |
| "kl": 0.032501220703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0107, | |
| "num_tokens": 4999461.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 239 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 618.0, | |
| "epoch": 0.47174447174447176, | |
| "grad_norm": 0.00408085066316832, | |
| "kl": 0.03265380859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 5023321.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.1875, | |
| "step": 240 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 562.40625, | |
| "epoch": 0.47371007371007373, | |
| "grad_norm": 0.7043144078754758, | |
| "kl": 0.03973388671875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0176, | |
| "num_tokens": 5044300.0, | |
| "reward": 0.53125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.53125, | |
| "step": 241 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 534.5, | |
| "epoch": 0.4756756756756757, | |
| "grad_norm": 1.0386591332218544, | |
| "kl": 0.0329742431640625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0069, | |
| "num_tokens": 5064534.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.39774755761027336, | |
| "rewards/reward_func": 0.34375, | |
| "step": 242 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 579.625, | |
| "epoch": 0.4776412776412776, | |
| "grad_norm": 0.5640869175323494, | |
| "kl": 0.030670166015625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0096, | |
| "num_tokens": 5086222.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 243 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 494.0625, | |
| "epoch": 0.4796068796068796, | |
| "grad_norm": 0.6343600501349801, | |
| "kl": 0.033843994140625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0089, | |
| "num_tokens": 5105218.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 244 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 631.71875, | |
| "epoch": 0.48157248157248156, | |
| "grad_norm": 0.5374583150390896, | |
| "kl": 0.02874755859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0166, | |
| "num_tokens": 5130717.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 245 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 533.84375, | |
| "epoch": 0.48353808353808353, | |
| "grad_norm": 0.833366654624395, | |
| "kl": 0.0404052734375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0297, | |
| "num_tokens": 5151728.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 246 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 718.875, | |
| "epoch": 0.4855036855036855, | |
| "grad_norm": 0.34812015193013685, | |
| "kl": 0.0258331298828125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.006, | |
| "num_tokens": 5178038.0, | |
| "reward": 0.09375, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.09375, | |
| "step": 247 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 537.5, | |
| "epoch": 0.48746928746928747, | |
| "grad_norm": 0.646907137960505, | |
| "kl": 0.031463623046875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.018, | |
| "num_tokens": 5198692.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 248 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 582.90625, | |
| "epoch": 0.48943488943488944, | |
| "grad_norm": 0.46174894142283385, | |
| "kl": 0.030975341796875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0172, | |
| "num_tokens": 5220241.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.40625, | |
| "step": 249 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 639.1875, | |
| "epoch": 0.4914004914004914, | |
| "grad_norm": 0.45612193916391053, | |
| "kl": 0.031524658203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0042, | |
| "num_tokens": 5246589.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 250 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 543.0625, | |
| "epoch": 0.4933660933660934, | |
| "grad_norm": 0.5938060228305441, | |
| "kl": 0.031219482421875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0037, | |
| "num_tokens": 5267427.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 251 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 541.0625, | |
| "epoch": 0.49533169533169535, | |
| "grad_norm": 0.6182124930877956, | |
| "kl": 0.032958984375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0064, | |
| "num_tokens": 5288851.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 252 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 603.5, | |
| "epoch": 0.4972972972972973, | |
| "grad_norm": 0.5011035982710935, | |
| "kl": 0.0300750732421875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0113, | |
| "num_tokens": 5312549.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 253 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 571.375, | |
| "epoch": 0.49926289926289924, | |
| "grad_norm": 0.7964580067571678, | |
| "kl": 0.0327606201171875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0314, | |
| "num_tokens": 5334683.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.1875, | |
| "step": 254 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 500.34375, | |
| "epoch": 0.5012285012285013, | |
| "grad_norm": 0.6228321163898966, | |
| "kl": 0.030609130859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0119, | |
| "num_tokens": 5353900.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 255 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 672.78125, | |
| "epoch": 0.5031941031941032, | |
| "grad_norm": 0.41977655143914255, | |
| "kl": 0.0276641845703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.005, | |
| "num_tokens": 5378597.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 256 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 603.875, | |
| "epoch": 0.5051597051597052, | |
| "grad_norm": 0.7823424230820542, | |
| "kl": 0.0303192138671875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0065, | |
| "num_tokens": 5402139.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.15625, | |
| "step": 257 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 521.46875, | |
| "epoch": 0.5071253071253071, | |
| "grad_norm": 0.5439456208837793, | |
| "kl": 0.0348052978515625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0027, | |
| "num_tokens": 5422456.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 258 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 481.3125, | |
| "epoch": 0.509090909090909, | |
| "grad_norm": 0.7976544572687917, | |
| "kl": 0.03790283203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0277, | |
| "num_tokens": 5440868.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 259 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 482.65625, | |
| "epoch": 0.5110565110565111, | |
| "grad_norm": 0.6137513319405157, | |
| "kl": 0.037567138671875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0212, | |
| "num_tokens": 5459865.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 260 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 582.53125, | |
| "epoch": 0.513022113022113, | |
| "grad_norm": 0.561088121226348, | |
| "kl": 0.03466796875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0325, | |
| "num_tokens": 5482278.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 261 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 610.9375, | |
| "epoch": 0.514987714987715, | |
| "grad_norm": 0.40737350155236235, | |
| "kl": 0.0304107666015625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0164, | |
| "num_tokens": 5505802.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 262 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 558.28125, | |
| "epoch": 0.5169533169533169, | |
| "grad_norm": 0.8967360376806364, | |
| "kl": 0.0367889404296875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0129, | |
| "num_tokens": 5526721.0, | |
| "reward": 0.375, | |
| "reward_std": 0.3535533845424652, | |
| "rewards/reward_func": 0.375, | |
| "step": 263 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 668.1875, | |
| "epoch": 0.518918918918919, | |
| "grad_norm": 0.5741886327516559, | |
| "kl": 0.0272674560546875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0009, | |
| "num_tokens": 5552191.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 264 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 592.75, | |
| "epoch": 0.5208845208845209, | |
| "grad_norm": 0.6032240213293185, | |
| "kl": 0.031951904296875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.005, | |
| "num_tokens": 5574565.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 265 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 637.90625, | |
| "epoch": 0.5228501228501229, | |
| "grad_norm": 0.41254200551494014, | |
| "kl": 0.03033447265625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0074, | |
| "num_tokens": 5598552.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 266 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 567.40625, | |
| "epoch": 0.5248157248157248, | |
| "grad_norm": 0.4342634053395224, | |
| "kl": 0.034881591796875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0119, | |
| "num_tokens": 5620875.0, | |
| "reward": 0.25, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.25, | |
| "step": 267 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 572.46875, | |
| "epoch": 0.5267813267813268, | |
| "grad_norm": 0.737529118820131, | |
| "kl": 0.028900146484375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0298, | |
| "num_tokens": 5642648.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.3125, | |
| "step": 268 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 685.625, | |
| "epoch": 0.5287469287469287, | |
| "grad_norm": 0.4500569807633216, | |
| "kl": 0.026824951171875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0106, | |
| "num_tokens": 5668694.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 269 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 679.28125, | |
| "epoch": 0.5307125307125307, | |
| "grad_norm": 0.4849462015076887, | |
| "kl": 0.024566650390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0078, | |
| "num_tokens": 5694389.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 270 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 528.0, | |
| "epoch": 0.5326781326781327, | |
| "grad_norm": 0.28404451611955495, | |
| "kl": 0.0351409912109375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0026, | |
| "num_tokens": 5715327.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.21875, | |
| "step": 271 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 568.78125, | |
| "epoch": 0.5346437346437346, | |
| "grad_norm": 0.6496110805941108, | |
| "kl": 0.0332794189453125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.007, | |
| "num_tokens": 5737064.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 272 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 619.03125, | |
| "epoch": 0.5366093366093366, | |
| "grad_norm": 0.5926254229816691, | |
| "kl": 0.034912109375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0138, | |
| "num_tokens": 5760589.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 273 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 536.1875, | |
| "epoch": 0.5385749385749385, | |
| "grad_norm": 0.8103580352644913, | |
| "kl": 0.0359039306640625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0085, | |
| "num_tokens": 5780855.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 274 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 639.84375, | |
| "epoch": 0.5405405405405406, | |
| "grad_norm": 0.4708407621573234, | |
| "kl": 0.0315093994140625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0003, | |
| "num_tokens": 5805138.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 275 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 499.40625, | |
| "epoch": 0.5425061425061425, | |
| "grad_norm": 0.7738022512202402, | |
| "kl": 0.0649566650390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0035, | |
| "num_tokens": 5824025.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 276 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 557.15625, | |
| "epoch": 0.5444717444717445, | |
| "grad_norm": 0.6506653997702296, | |
| "kl": 0.035491943359375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0106, | |
| "num_tokens": 5845784.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 277 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 480.4375, | |
| "epoch": 0.5464373464373464, | |
| "grad_norm": 0.7038827199007561, | |
| "kl": 0.0393524169921875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0202, | |
| "num_tokens": 5864738.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 278 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 573.53125, | |
| "epoch": 0.5484029484029485, | |
| "grad_norm": 0.9617362599461011, | |
| "kl": 0.0326080322265625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0444, | |
| "num_tokens": 5887175.0, | |
| "reward": 0.375, | |
| "reward_std": 0.3535533845424652, | |
| "rewards/reward_func": 0.375, | |
| "step": 279 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 522.5625, | |
| "epoch": 0.5503685503685504, | |
| "grad_norm": 0.615849444934388, | |
| "kl": 0.03692626953125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0023, | |
| "num_tokens": 5907477.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 280 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 524.5, | |
| "epoch": 0.5523341523341523, | |
| "grad_norm": 0.3464132826493981, | |
| "kl": 0.03509521484375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0217, | |
| "num_tokens": 5927731.0, | |
| "reward": 0.375, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.375, | |
| "step": 281 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 461.8125, | |
| "epoch": 0.5542997542997543, | |
| "grad_norm": 0.6909097974436648, | |
| "kl": 0.042205810546875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0181, | |
| "num_tokens": 5945637.0, | |
| "reward": 0.5625, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.5625, | |
| "step": 282 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 587.28125, | |
| "epoch": 0.5562653562653562, | |
| "grad_norm": 0.7289737680839917, | |
| "kl": 0.050811767578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0467, | |
| "num_tokens": 5969752.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.40625, | |
| "step": 283 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 495.5625, | |
| "epoch": 0.5582309582309583, | |
| "grad_norm": 0.5981873778704219, | |
| "kl": 0.0392303466796875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0157, | |
| "num_tokens": 5989426.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 284 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 516.21875, | |
| "epoch": 0.5601965601965602, | |
| "grad_norm": 0.9106360418769466, | |
| "kl": 0.033355712890625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0412, | |
| "num_tokens": 6009763.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.39774755761027336, | |
| "rewards/reward_func": 0.40625, | |
| "step": 285 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 557.3125, | |
| "epoch": 0.5621621621621622, | |
| "grad_norm": 0.602763648223161, | |
| "kl": 0.0350799560546875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0376, | |
| "num_tokens": 6031665.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 286 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 529.1875, | |
| "epoch": 0.5641277641277641, | |
| "grad_norm": 0.7103001938223117, | |
| "kl": 0.0367889404296875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0208, | |
| "num_tokens": 6052263.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.40625, | |
| "step": 287 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 579.3125, | |
| "epoch": 0.5660933660933661, | |
| "grad_norm": 0.6000018822425005, | |
| "kl": 0.0293121337890625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0064, | |
| "num_tokens": 6074509.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 288 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 561.0625, | |
| "epoch": 0.568058968058968, | |
| "grad_norm": 0.6084807148588314, | |
| "kl": 0.0366363525390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0132, | |
| "num_tokens": 6096751.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 289 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 445.21875, | |
| "epoch": 0.5700245700245701, | |
| "grad_norm": 0.897808318980553, | |
| "kl": 0.0430908203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0165, | |
| "num_tokens": 6114208.0, | |
| "reward": 0.25, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.25, | |
| "step": 290 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 507.65625, | |
| "epoch": 0.571990171990172, | |
| "grad_norm": 0.7237211163121627, | |
| "kl": 0.030975341796875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.005, | |
| "num_tokens": 6134967.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.40625, | |
| "step": 291 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 636.28125, | |
| "epoch": 0.5739557739557739, | |
| "grad_norm": 0.4667444055129069, | |
| "kl": 0.03082275390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0189, | |
| "num_tokens": 6158956.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 292 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 430.4375, | |
| "epoch": 0.5759213759213759, | |
| "grad_norm": 0.773152795620052, | |
| "kl": 0.04296875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0254, | |
| "num_tokens": 6176394.0, | |
| "reward": 0.59375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.59375, | |
| "step": 293 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 529.65625, | |
| "epoch": 0.5778869778869778, | |
| "grad_norm": 0.5498942657198764, | |
| "kl": 0.031890869140625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0031, | |
| "num_tokens": 6198079.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.1875, | |
| "step": 294 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 429.34375, | |
| "epoch": 0.5798525798525799, | |
| "grad_norm": 0.8165142043245753, | |
| "kl": 0.0404205322265625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0292, | |
| "num_tokens": 6214812.0, | |
| "reward": 0.375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.375, | |
| "step": 295 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 563.59375, | |
| "epoch": 0.5818181818181818, | |
| "grad_norm": 0.004534611785819621, | |
| "kl": 0.036376953125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 6236273.0, | |
| "reward": 0.5, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.5, | |
| "step": 296 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 611.09375, | |
| "epoch": 0.5837837837837838, | |
| "grad_norm": 0.3953649614977329, | |
| "kl": 0.0288238525390625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0081, | |
| "num_tokens": 6259916.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.21875, | |
| "step": 297 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 502.5625, | |
| "epoch": 0.5857493857493857, | |
| "grad_norm": 0.8629063325607663, | |
| "kl": 0.034912109375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0142, | |
| "num_tokens": 6279672.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.3125, | |
| "step": 298 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 512.8125, | |
| "epoch": 0.5877149877149878, | |
| "grad_norm": 0.7312949259699398, | |
| "kl": 0.0346527099609375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0231, | |
| "num_tokens": 6299108.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 299 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 669.3125, | |
| "epoch": 0.5896805896805897, | |
| "grad_norm": 0.3844211231718793, | |
| "kl": 0.0296630859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0046, | |
| "num_tokens": 6324478.0, | |
| "reward": 0.25, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.25, | |
| "step": 300 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 663.71875, | |
| "epoch": 0.5916461916461917, | |
| "grad_norm": 0.5222770323856585, | |
| "kl": 0.0269317626953125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0194, | |
| "num_tokens": 6349253.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 301 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 610.59375, | |
| "epoch": 0.5936117936117936, | |
| "grad_norm": 0.5352648071252188, | |
| "kl": 0.0316619873046875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.023, | |
| "num_tokens": 6372508.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 302 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 523.4375, | |
| "epoch": 0.5955773955773955, | |
| "grad_norm": 0.4400789874053765, | |
| "kl": 0.0358734130859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0053, | |
| "num_tokens": 6392690.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 303 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 617.3125, | |
| "epoch": 0.5975429975429976, | |
| "grad_norm": 0.5569387236547813, | |
| "kl": 0.038787841796875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0386, | |
| "num_tokens": 6416800.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 304 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 622.125, | |
| "epoch": 0.5995085995085995, | |
| "grad_norm": 0.3406849270712581, | |
| "kl": 0.031158447265625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0027, | |
| "num_tokens": 6442038.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.1875, | |
| "step": 305 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 492.75, | |
| "epoch": 0.6014742014742015, | |
| "grad_norm": 0.5946586449478722, | |
| "kl": 0.0381927490234375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0293, | |
| "num_tokens": 6461826.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 306 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 528.78125, | |
| "epoch": 0.6034398034398034, | |
| "grad_norm": 0.5089011679410208, | |
| "kl": 0.0418548583984375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0248, | |
| "num_tokens": 6482279.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 307 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 594.0, | |
| "epoch": 0.6054054054054054, | |
| "grad_norm": 0.6413836653957254, | |
| "kl": 0.03863525390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0207, | |
| "num_tokens": 6504697.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 308 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 634.28125, | |
| "epoch": 0.6073710073710074, | |
| "grad_norm": 0.3854408237264038, | |
| "kl": 0.0313720703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.038, | |
| "num_tokens": 6529342.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 309 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 525.65625, | |
| "epoch": 0.6093366093366094, | |
| "grad_norm": 0.4971690550435448, | |
| "kl": 0.0362091064453125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0044, | |
| "num_tokens": 6550177.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 310 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 533.0, | |
| "epoch": 0.6113022113022113, | |
| "grad_norm": 0.31552374439782604, | |
| "kl": 0.0349884033203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0034, | |
| "num_tokens": 6570467.0, | |
| "reward": 0.5, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.5, | |
| "step": 311 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 588.65625, | |
| "epoch": 0.6132678132678133, | |
| "grad_norm": 0.6494668114475654, | |
| "kl": 0.032928466796875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0078, | |
| "num_tokens": 6592632.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 312 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 490.375, | |
| "epoch": 0.6152334152334152, | |
| "grad_norm": 0.6225762736741876, | |
| "kl": 0.0364532470703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0188, | |
| "num_tokens": 6612088.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 313 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 504.84375, | |
| "epoch": 0.6171990171990172, | |
| "grad_norm": 0.7409698772731983, | |
| "kl": 0.0345001220703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0004, | |
| "num_tokens": 6631959.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 314 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 520.0625, | |
| "epoch": 0.6191646191646192, | |
| "grad_norm": 0.9716296912093264, | |
| "kl": 0.0350494384765625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0063, | |
| "num_tokens": 6652429.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.39774755761027336, | |
| "rewards/reward_func": 0.28125, | |
| "step": 315 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 463.5, | |
| "epoch": 0.6211302211302211, | |
| "grad_norm": 0.4764914574248498, | |
| "kl": 0.0382232666015625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0084, | |
| "num_tokens": 6671487.0, | |
| "reward": 0.25, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.25, | |
| "step": 316 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 578.15625, | |
| "epoch": 0.6230958230958231, | |
| "grad_norm": 0.5689759451374822, | |
| "kl": 0.0312042236328125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0122, | |
| "num_tokens": 6693304.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 317 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 450.125, | |
| "epoch": 0.625061425061425, | |
| "grad_norm": 0.5697978782946299, | |
| "kl": 0.04327392578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0086, | |
| "num_tokens": 6710562.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.46875, | |
| "step": 318 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 506.46875, | |
| "epoch": 0.6270270270270271, | |
| "grad_norm": 0.5626725420705256, | |
| "kl": 0.036956787109375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.007, | |
| "num_tokens": 6729951.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.4375, | |
| "step": 319 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 547.0625, | |
| "epoch": 0.628992628992629, | |
| "grad_norm": 0.5894973888067989, | |
| "kl": 0.033966064453125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0202, | |
| "num_tokens": 6750997.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 320 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 708.78125, | |
| "epoch": 0.630958230958231, | |
| "grad_norm": 0.37243224501970285, | |
| "kl": 0.0262451171875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0023, | |
| "num_tokens": 6777310.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 321 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 600.90625, | |
| "epoch": 0.6329238329238329, | |
| "grad_norm": 0.5442459669871826, | |
| "kl": 0.0332794189453125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.003, | |
| "num_tokens": 6800313.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 322 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 509.4375, | |
| "epoch": 0.634889434889435, | |
| "grad_norm": 0.5413789628110073, | |
| "kl": 0.0334320068359375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0002, | |
| "num_tokens": 6819943.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 323 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 499.71875, | |
| "epoch": 0.6368550368550369, | |
| "grad_norm": 0.3901208892455512, | |
| "kl": 0.0421600341796875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.004, | |
| "num_tokens": 6839360.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 324 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 455.90625, | |
| "epoch": 0.6388206388206388, | |
| "grad_norm": 0.5926575612429559, | |
| "kl": 0.038970947265625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0026, | |
| "num_tokens": 6858437.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 325 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 546.625, | |
| "epoch": 0.6407862407862408, | |
| "grad_norm": 0.7062156547375247, | |
| "kl": 0.0332183837890625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0253, | |
| "num_tokens": 6879563.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 326 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 419.21875, | |
| "epoch": 0.6427518427518427, | |
| "grad_norm": 0.8379864895035511, | |
| "kl": 0.046844482421875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.011, | |
| "num_tokens": 6895844.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.4375, | |
| "step": 327 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 489.03125, | |
| "epoch": 0.6447174447174447, | |
| "grad_norm": 0.6425469992752795, | |
| "kl": 0.03472900390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0442, | |
| "num_tokens": 6914839.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 328 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 490.21875, | |
| "epoch": 0.6466830466830467, | |
| "grad_norm": 0.5538877152424146, | |
| "kl": 0.0370941162109375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0017, | |
| "num_tokens": 6933922.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 329 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 605.84375, | |
| "epoch": 0.6486486486486487, | |
| "grad_norm": 0.2911469195673005, | |
| "kl": 0.02972412109375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0052, | |
| "num_tokens": 6956815.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.21875, | |
| "step": 330 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 524.96875, | |
| "epoch": 0.6506142506142506, | |
| "grad_norm": 0.6692638014108062, | |
| "kl": 0.0333404541015625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0261, | |
| "num_tokens": 6976844.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 331 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 504.09375, | |
| "epoch": 0.6525798525798526, | |
| "grad_norm": 0.7535594730898131, | |
| "kl": 0.0408782958984375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0078, | |
| "num_tokens": 6996319.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.3125, | |
| "step": 332 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 622.625, | |
| "epoch": 0.6545454545454545, | |
| "grad_norm": 0.37491504190078967, | |
| "kl": 0.030029296875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0127, | |
| "num_tokens": 7020507.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.34375, | |
| "step": 333 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 528.96875, | |
| "epoch": 0.6565110565110565, | |
| "grad_norm": 0.6920450435680606, | |
| "kl": 0.0441131591796875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0129, | |
| "num_tokens": 7041354.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 334 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 449.46875, | |
| "epoch": 0.6584766584766585, | |
| "grad_norm": 0.6734712281138222, | |
| "kl": 0.043548583984375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0095, | |
| "num_tokens": 7060813.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 335 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 566.625, | |
| "epoch": 0.6604422604422604, | |
| "grad_norm": 0.7077598401138161, | |
| "kl": 0.040802001953125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0183, | |
| "num_tokens": 7081947.0, | |
| "reward": 0.25, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.25, | |
| "step": 336 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 480.09375, | |
| "epoch": 0.6624078624078624, | |
| "grad_norm": 0.5853457266532924, | |
| "kl": 0.043914794921875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0052, | |
| "num_tokens": 7102076.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.4375, | |
| "step": 337 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 567.65625, | |
| "epoch": 0.6643734643734643, | |
| "grad_norm": 0.5842956050999084, | |
| "kl": 0.0388946533203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0072, | |
| "num_tokens": 7124387.0, | |
| "reward": 0.25, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.25, | |
| "step": 338 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 511.125, | |
| "epoch": 0.6663390663390664, | |
| "grad_norm": 0.4668262300707244, | |
| "kl": 0.03558349609375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0015, | |
| "num_tokens": 7144605.0, | |
| "reward": 0.375, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.375, | |
| "step": 339 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 508.59375, | |
| "epoch": 0.6683046683046683, | |
| "grad_norm": 0.8250382613151874, | |
| "kl": 0.0343170166015625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0932, | |
| "num_tokens": 7164278.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 340 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 489.96875, | |
| "epoch": 0.6702702702702703, | |
| "grad_norm": 0.7144131063171276, | |
| "kl": 0.0345001220703125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0074, | |
| "num_tokens": 7184667.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 341 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 544.21875, | |
| "epoch": 0.6722358722358722, | |
| "grad_norm": 0.8700749896413762, | |
| "kl": 0.0315093994140625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0151, | |
| "num_tokens": 7205714.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 342 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 610.375, | |
| "epoch": 0.6742014742014742, | |
| "grad_norm": 0.45694633600819196, | |
| "kl": 0.030517578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0112, | |
| "num_tokens": 7229098.0, | |
| "reward": 0.125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.125, | |
| "step": 343 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 521.46875, | |
| "epoch": 0.6761670761670762, | |
| "grad_norm": 0.23257958215792865, | |
| "kl": 0.0365447998046875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0096, | |
| "num_tokens": 7249091.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.34375, | |
| "step": 344 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 577.875, | |
| "epoch": 0.6781326781326781, | |
| "grad_norm": 0.4152502823473006, | |
| "kl": 0.0332489013671875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0014, | |
| "num_tokens": 7271555.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.4375, | |
| "step": 345 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 446.3125, | |
| "epoch": 0.6800982800982801, | |
| "grad_norm": 0.6310780819784905, | |
| "kl": 0.04241943359375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0235, | |
| "num_tokens": 7288979.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.46875, | |
| "step": 346 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 565.5625, | |
| "epoch": 0.682063882063882, | |
| "grad_norm": 0.6457400724427969, | |
| "kl": 0.041015625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0055, | |
| "num_tokens": 7311269.0, | |
| "reward": 0.375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.375, | |
| "step": 347 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 521.09375, | |
| "epoch": 0.684029484029484, | |
| "grad_norm": 0.5857246046657881, | |
| "kl": 0.037384033203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0203, | |
| "num_tokens": 7332160.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 348 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 558.28125, | |
| "epoch": 0.685995085995086, | |
| "grad_norm": 0.36864784062194783, | |
| "kl": 0.035400390625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0255, | |
| "num_tokens": 7354603.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 349 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 510.09375, | |
| "epoch": 0.687960687960688, | |
| "grad_norm": 0.37595426080670763, | |
| "kl": 0.0396881103515625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0045, | |
| "num_tokens": 7374876.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 350 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 606.25, | |
| "epoch": 0.6899262899262899, | |
| "grad_norm": 0.7050890671231397, | |
| "kl": 0.032562255859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0187, | |
| "num_tokens": 7397590.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.15625, | |
| "step": 351 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 639.0625, | |
| "epoch": 0.6918918918918919, | |
| "grad_norm": 0.623394039605989, | |
| "kl": 0.029327392578125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0084, | |
| "num_tokens": 7422452.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 352 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 514.75, | |
| "epoch": 0.6938574938574938, | |
| "grad_norm": 0.7054121163036189, | |
| "kl": 0.03265380859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0129, | |
| "num_tokens": 7442250.0, | |
| "reward": 0.375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.375, | |
| "step": 353 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 592.0625, | |
| "epoch": 0.6958230958230959, | |
| "grad_norm": 0.6548424627448317, | |
| "kl": 0.0325927734375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0205, | |
| "num_tokens": 7464522.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 354 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 526.125, | |
| "epoch": 0.6977886977886978, | |
| "grad_norm": 0.9159125424427512, | |
| "kl": 0.0395050048828125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0266, | |
| "num_tokens": 7486958.0, | |
| "reward": 0.375, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.375, | |
| "step": 355 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 524.8125, | |
| "epoch": 0.6997542997542997, | |
| "grad_norm": 0.747990779466027, | |
| "kl": 0.0347137451171875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.037, | |
| "num_tokens": 7508700.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 356 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 670.3125, | |
| "epoch": 0.7017199017199017, | |
| "grad_norm": 0.5139288731237621, | |
| "kl": 0.0316314697265625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0135, | |
| "num_tokens": 7533558.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 357 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 555.25, | |
| "epoch": 0.7036855036855036, | |
| "grad_norm": 0.7849105797708434, | |
| "kl": 0.034576416015625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0218, | |
| "num_tokens": 7554376.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.28125, | |
| "step": 358 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 517.125, | |
| "epoch": 0.7056511056511057, | |
| "grad_norm": 0.8194811532288894, | |
| "kl": 0.0421905517578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0362, | |
| "num_tokens": 7574152.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.46875, | |
| "step": 359 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 468.9375, | |
| "epoch": 0.7076167076167076, | |
| "grad_norm": 0.4941945178172363, | |
| "kl": 0.041046142578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0048, | |
| "num_tokens": 7592906.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 360 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 498.9375, | |
| "epoch": 0.7095823095823096, | |
| "grad_norm": 0.8121886110174811, | |
| "kl": 0.04071044921875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0818, | |
| "num_tokens": 7612886.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.40625, | |
| "step": 361 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 601.15625, | |
| "epoch": 0.7115479115479115, | |
| "grad_norm": 0.45512137045157774, | |
| "kl": 0.03326416015625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0033, | |
| "num_tokens": 7635799.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.4375, | |
| "step": 362 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 545.40625, | |
| "epoch": 0.7135135135135136, | |
| "grad_norm": 0.830092454914979, | |
| "kl": 0.0363311767578125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0188, | |
| "num_tokens": 7656276.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.40625, | |
| "step": 363 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 451.9375, | |
| "epoch": 0.7154791154791155, | |
| "grad_norm": 0.3436418620406507, | |
| "kl": 0.04986572265625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0065, | |
| "num_tokens": 7674878.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 364 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 473.40625, | |
| "epoch": 0.7174447174447175, | |
| "grad_norm": 0.6642678609768281, | |
| "kl": 0.0440521240234375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0141, | |
| "num_tokens": 7694745.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 365 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 480.71875, | |
| "epoch": 0.7194103194103194, | |
| "grad_norm": 0.7446714368531804, | |
| "kl": 0.046142578125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0123, | |
| "num_tokens": 7713650.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.4375, | |
| "step": 366 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 484.375, | |
| "epoch": 0.7213759213759213, | |
| "grad_norm": 0.8698002010484014, | |
| "kl": 0.042236328125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0213, | |
| "num_tokens": 7732482.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 367 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 461.875, | |
| "epoch": 0.7233415233415234, | |
| "grad_norm": 0.9898476778510431, | |
| "kl": 0.03765869140625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0047, | |
| "num_tokens": 7750484.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.39774755761027336, | |
| "rewards/reward_func": 0.46875, | |
| "step": 368 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 457.6875, | |
| "epoch": 0.7253071253071253, | |
| "grad_norm": 0.8069454227623065, | |
| "kl": 0.049163818359375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0093, | |
| "num_tokens": 7767968.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.4375, | |
| "step": 369 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 439.3125, | |
| "epoch": 0.7272727272727273, | |
| "grad_norm": 0.8746030989783482, | |
| "kl": 0.043609619140625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0183, | |
| "num_tokens": 7786330.0, | |
| "reward": 0.5625, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.5625, | |
| "step": 370 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 446.4375, | |
| "epoch": 0.7292383292383292, | |
| "grad_norm": 0.4845476792447974, | |
| "kl": 0.1286468505859375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0175, | |
| "num_tokens": 7804720.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.21875, | |
| "step": 371 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 434.53125, | |
| "epoch": 0.7312039312039312, | |
| "grad_norm": 0.7019254770499221, | |
| "kl": 0.0400238037109375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0082, | |
| "num_tokens": 7821607.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 372 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 624.34375, | |
| "epoch": 0.7331695331695331, | |
| "grad_norm": 0.6696488313790595, | |
| "kl": 0.03546142578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0218, | |
| "num_tokens": 7844870.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 373 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 561.6875, | |
| "epoch": 0.7351351351351352, | |
| "grad_norm": 0.8600206462556853, | |
| "kl": 0.0367431640625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0546, | |
| "num_tokens": 7866172.0, | |
| "reward": 0.53125, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.53125, | |
| "step": 374 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 497.125, | |
| "epoch": 0.7371007371007371, | |
| "grad_norm": 0.7638586068500758, | |
| "kl": 0.0342559814453125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0107, | |
| "num_tokens": 7885490.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.46875, | |
| "step": 375 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 566.53125, | |
| "epoch": 0.7390663390663391, | |
| "grad_norm": 0.4829267801913666, | |
| "kl": 0.0334930419921875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0167, | |
| "num_tokens": 7908661.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 376 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 480.40625, | |
| "epoch": 0.741031941031941, | |
| "grad_norm": 0.740899432881196, | |
| "kl": 0.0396575927734375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0194, | |
| "num_tokens": 7927568.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.46875, | |
| "step": 377 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 568.09375, | |
| "epoch": 0.742997542997543, | |
| "grad_norm": 0.5106656848937849, | |
| "kl": 0.036956787109375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0017, | |
| "num_tokens": 7950839.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 378 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 522.8125, | |
| "epoch": 0.744963144963145, | |
| "grad_norm": 0.6659830768357405, | |
| "kl": 0.03570556640625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0149, | |
| "num_tokens": 7971197.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 379 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 512.9375, | |
| "epoch": 0.7469287469287469, | |
| "grad_norm": 0.7524006374031159, | |
| "kl": 0.0421142578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0224, | |
| "num_tokens": 7991209.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 380 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 545.75, | |
| "epoch": 0.7488943488943489, | |
| "grad_norm": 0.4839548397489806, | |
| "kl": 0.0340118408203125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.014, | |
| "num_tokens": 8012103.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 381 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 532.15625, | |
| "epoch": 0.7508599508599508, | |
| "grad_norm": 0.7060415982161667, | |
| "kl": 0.0348968505859375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.003, | |
| "num_tokens": 8032922.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.40625, | |
| "step": 382 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 593.0625, | |
| "epoch": 0.7528255528255529, | |
| "grad_norm": 0.8703459584049694, | |
| "kl": 0.0374755859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0024, | |
| "num_tokens": 8055494.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.3125, | |
| "step": 383 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 605.28125, | |
| "epoch": 0.7547911547911548, | |
| "grad_norm": 0.531040583745666, | |
| "kl": 0.0342864990234375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.004, | |
| "num_tokens": 8078407.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 384 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 472.71875, | |
| "epoch": 0.7567567567567568, | |
| "grad_norm": 0.6413661425786513, | |
| "kl": 0.0486602783203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0229, | |
| "num_tokens": 8096908.0, | |
| "reward": 0.5, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.5, | |
| "step": 385 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 540.40625, | |
| "epoch": 0.7587223587223587, | |
| "grad_norm": 0.5887264630204668, | |
| "kl": 0.0734100341796875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.015, | |
| "num_tokens": 8118141.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 386 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 552.3125, | |
| "epoch": 0.7606879606879607, | |
| "grad_norm": 0.35879317767107755, | |
| "kl": 0.0392608642578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0237, | |
| "num_tokens": 8139923.0, | |
| "reward": 0.15625, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.15625, | |
| "step": 387 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 426.0, | |
| "epoch": 0.7626535626535627, | |
| "grad_norm": 0.6612874102908352, | |
| "kl": 0.0440216064453125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0187, | |
| "num_tokens": 8157183.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 388 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 431.96875, | |
| "epoch": 0.7646191646191646, | |
| "grad_norm": 0.7214155280951005, | |
| "kl": 0.0413360595703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0028, | |
| "num_tokens": 8174882.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 389 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 464.0625, | |
| "epoch": 0.7665847665847666, | |
| "grad_norm": 0.6542555963916522, | |
| "kl": 0.0447845458984375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0012, | |
| "num_tokens": 8192790.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 390 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 508.65625, | |
| "epoch": 0.7685503685503685, | |
| "grad_norm": 0.8263885848953906, | |
| "kl": 0.03887939453125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0084, | |
| "num_tokens": 8212111.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.46875, | |
| "step": 391 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 615.46875, | |
| "epoch": 0.7705159705159705, | |
| "grad_norm": 0.5049205742260084, | |
| "kl": 0.0379638671875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0336, | |
| "num_tokens": 8236940.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 392 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 600.15625, | |
| "epoch": 0.7724815724815725, | |
| "grad_norm": 0.6038728188859286, | |
| "kl": 0.047821044921875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0255, | |
| "num_tokens": 8260543.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.1875, | |
| "step": 393 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 490.40625, | |
| "epoch": 0.7744471744471745, | |
| "grad_norm": 0.48639402941702775, | |
| "kl": 0.0445709228515625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0125, | |
| "num_tokens": 8279292.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.40625, | |
| "step": 394 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 526.28125, | |
| "epoch": 0.7764127764127764, | |
| "grad_norm": 0.37376433880625853, | |
| "kl": 0.04119873046875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0067, | |
| "num_tokens": 8300045.0, | |
| "reward": 0.125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.125, | |
| "step": 395 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 587.78125, | |
| "epoch": 0.7783783783783784, | |
| "grad_norm": 0.5600748468776157, | |
| "kl": 0.034210205078125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0145, | |
| "num_tokens": 8323160.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 396 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 546.1875, | |
| "epoch": 0.7803439803439803, | |
| "grad_norm": 0.6575791496386095, | |
| "kl": 0.044769287109375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0228, | |
| "num_tokens": 8345354.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 397 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 550.65625, | |
| "epoch": 0.7823095823095824, | |
| "grad_norm": 0.7332146409060798, | |
| "kl": 0.0418243408203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0071, | |
| "num_tokens": 8367361.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 398 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 578.8125, | |
| "epoch": 0.7842751842751843, | |
| "grad_norm": 0.35869977857179974, | |
| "kl": 0.03509521484375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0202, | |
| "num_tokens": 8390055.0, | |
| "reward": 0.25, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.25, | |
| "step": 399 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 510.09375, | |
| "epoch": 0.7862407862407862, | |
| "grad_norm": 0.4437939683355927, | |
| "kl": 0.0397796630859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.002, | |
| "num_tokens": 8411168.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.28125, | |
| "step": 400 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 608.3125, | |
| "epoch": 0.7882063882063882, | |
| "grad_norm": 0.2883562751351335, | |
| "kl": 0.03594970703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0132, | |
| "num_tokens": 8433636.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.40625, | |
| "step": 401 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 561.09375, | |
| "epoch": 0.7901719901719901, | |
| "grad_norm": 0.5029309214073006, | |
| "kl": 0.0384063720703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0368, | |
| "num_tokens": 8454585.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.46875, | |
| "step": 402 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 500.96875, | |
| "epoch": 0.7921375921375922, | |
| "grad_norm": 0.5436913833491791, | |
| "kl": 0.0381011962890625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0333, | |
| "num_tokens": 8474234.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 403 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 555.71875, | |
| "epoch": 0.7941031941031941, | |
| "grad_norm": 0.7152534354108595, | |
| "kl": 0.0462188720703125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0401, | |
| "num_tokens": 8495313.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 404 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 538.4375, | |
| "epoch": 0.7960687960687961, | |
| "grad_norm": 0.7749306460061681, | |
| "kl": 0.04327392578125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0084, | |
| "num_tokens": 8515913.0, | |
| "reward": 0.25, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.25, | |
| "step": 405 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 548.25, | |
| "epoch": 0.798034398034398, | |
| "grad_norm": 0.7006524831087708, | |
| "kl": 0.0382537841796875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0414, | |
| "num_tokens": 8537121.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.40625, | |
| "step": 406 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 541.4375, | |
| "epoch": 0.8, | |
| "grad_norm": 0.654242660435957, | |
| "kl": 0.037353515625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 8559349.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 407 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 408.96875, | |
| "epoch": 0.801965601965602, | |
| "grad_norm": 0.6261245534491866, | |
| "kl": 0.049560546875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0053, | |
| "num_tokens": 8575850.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 408 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 507.8125, | |
| "epoch": 0.803931203931204, | |
| "grad_norm": 0.596943807393144, | |
| "kl": 0.045806884765625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0068, | |
| "num_tokens": 8596270.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.40625, | |
| "step": 409 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 467.25, | |
| "epoch": 0.8058968058968059, | |
| "grad_norm": 0.8480568951318361, | |
| "kl": 0.04742431640625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.015, | |
| "num_tokens": 8614880.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.3125, | |
| "step": 410 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 554.5625, | |
| "epoch": 0.8078624078624078, | |
| "grad_norm": 0.6934048214061209, | |
| "kl": 0.0448150634765625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0165, | |
| "num_tokens": 8636184.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.40625, | |
| "step": 411 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 488.28125, | |
| "epoch": 0.8098280098280098, | |
| "grad_norm": 0.48599362890989417, | |
| "kl": 0.042694091796875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0144, | |
| "num_tokens": 8655357.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 412 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 484.53125, | |
| "epoch": 0.8117936117936118, | |
| "grad_norm": 0.7147383773279342, | |
| "kl": 0.038909912109375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0102, | |
| "num_tokens": 8673852.0, | |
| "reward": 0.53125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.53125, | |
| "step": 413 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 575.5, | |
| "epoch": 0.8137592137592138, | |
| "grad_norm": 0.32192540490547855, | |
| "kl": 0.035552978515625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.007, | |
| "num_tokens": 8695206.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.34375, | |
| "step": 414 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 543.8125, | |
| "epoch": 0.8157248157248157, | |
| "grad_norm": 0.5700024867715143, | |
| "kl": 0.04132080078125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0493, | |
| "num_tokens": 8717744.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 415 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 523.1875, | |
| "epoch": 0.8176904176904177, | |
| "grad_norm": 0.4619033887646093, | |
| "kl": 0.0422515869140625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.017, | |
| "num_tokens": 8738210.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 416 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 542.0625, | |
| "epoch": 0.8196560196560196, | |
| "grad_norm": 0.4499473639990639, | |
| "kl": 0.0384521484375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.018, | |
| "num_tokens": 8759650.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 417 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 501.96875, | |
| "epoch": 0.8216216216216217, | |
| "grad_norm": 0.008040642029484348, | |
| "kl": 0.0464019775390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0, | |
| "num_tokens": 8780119.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.0, | |
| "rewards/reward_func": 0.4375, | |
| "step": 418 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 480.6875, | |
| "epoch": 0.8235872235872236, | |
| "grad_norm": 0.9431817721721498, | |
| "kl": 0.0443572998046875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0238, | |
| "num_tokens": 8798427.0, | |
| "reward": 0.25, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.25, | |
| "step": 419 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 485.125, | |
| "epoch": 0.8255528255528255, | |
| "grad_norm": 0.5837840088431165, | |
| "kl": 0.05145263671875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0048, | |
| "num_tokens": 8818277.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 420 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 474.53125, | |
| "epoch": 0.8275184275184275, | |
| "grad_norm": 0.7118362037590832, | |
| "kl": 0.04388427734375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0113, | |
| "num_tokens": 8837382.0, | |
| "reward": 0.59375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.59375, | |
| "step": 421 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 597.125, | |
| "epoch": 0.8294840294840294, | |
| "grad_norm": 0.3840894481957198, | |
| "kl": 0.0394287109375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0115, | |
| "num_tokens": 8861122.0, | |
| "reward": 0.375, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.375, | |
| "step": 422 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 572.375, | |
| "epoch": 0.8314496314496315, | |
| "grad_norm": 0.5331945067131889, | |
| "kl": 0.03997802734375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0258, | |
| "num_tokens": 8882982.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 423 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 459.25, | |
| "epoch": 0.8334152334152334, | |
| "grad_norm": 0.21463702529938705, | |
| "kl": 0.04638671875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0069, | |
| "num_tokens": 8901532.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.21875, | |
| "step": 424 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 499.25, | |
| "epoch": 0.8353808353808354, | |
| "grad_norm": 0.5417437150526841, | |
| "kl": 0.049285888671875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.041, | |
| "num_tokens": 8921268.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 425 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 470.71875, | |
| "epoch": 0.8373464373464373, | |
| "grad_norm": 0.8602038158937927, | |
| "kl": 0.047943115234375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0556, | |
| "num_tokens": 8940985.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.4375, | |
| "step": 426 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 581.78125, | |
| "epoch": 0.8393120393120393, | |
| "grad_norm": 0.8393394778722276, | |
| "kl": 0.0372772216796875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0111, | |
| "num_tokens": 8963234.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.40625, | |
| "step": 427 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 598.3125, | |
| "epoch": 0.8412776412776413, | |
| "grad_norm": 0.5416506654601897, | |
| "kl": 0.038848876953125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.008, | |
| "num_tokens": 8986074.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 428 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 509.96875, | |
| "epoch": 0.8432432432432433, | |
| "grad_norm": 0.7465918567434375, | |
| "kl": 0.04742431640625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0251, | |
| "num_tokens": 9006525.0, | |
| "reward": 0.375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.375, | |
| "step": 429 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 539.3125, | |
| "epoch": 0.8452088452088452, | |
| "grad_norm": 0.7890529810165291, | |
| "kl": 0.05438232421875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0181, | |
| "num_tokens": 9028675.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 430 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 571.59375, | |
| "epoch": 0.8471744471744471, | |
| "grad_norm": 0.505619144404018, | |
| "kl": 0.04241943359375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0142, | |
| "num_tokens": 9050618.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 431 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 536.71875, | |
| "epoch": 0.8491400491400491, | |
| "grad_norm": 0.7115120258334543, | |
| "kl": 0.037139892578125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0032, | |
| "num_tokens": 9071397.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 432 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 511.25, | |
| "epoch": 0.8511056511056511, | |
| "grad_norm": 0.5913109373359104, | |
| "kl": 0.0347747802734375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0226, | |
| "num_tokens": 9090691.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.46875, | |
| "step": 433 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 667.46875, | |
| "epoch": 0.8530712530712531, | |
| "grad_norm": 0.6879865599375653, | |
| "kl": 0.039642333984375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0025, | |
| "num_tokens": 9116912.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.1875, | |
| "step": 434 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 561.78125, | |
| "epoch": 0.855036855036855, | |
| "grad_norm": 0.7249650778428434, | |
| "kl": 0.046295166015625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0103, | |
| "num_tokens": 9140269.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 435 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 555.65625, | |
| "epoch": 0.857002457002457, | |
| "grad_norm": 0.7477379427688969, | |
| "kl": 0.0420379638671875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0045, | |
| "num_tokens": 9163224.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 436 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 612.1875, | |
| "epoch": 0.858968058968059, | |
| "grad_norm": 0.7596946251219937, | |
| "kl": 0.039398193359375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0197, | |
| "num_tokens": 9186472.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 437 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 559.4375, | |
| "epoch": 0.860933660933661, | |
| "grad_norm": 0.4603738577917549, | |
| "kl": 0.050079345703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0032, | |
| "num_tokens": 9208562.0, | |
| "reward": 0.25, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.25, | |
| "step": 438 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 599.3125, | |
| "epoch": 0.8628992628992629, | |
| "grad_norm": 0.7518131906274648, | |
| "kl": 0.0406341552734375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0063, | |
| "num_tokens": 9231966.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.28125, | |
| "step": 439 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 589.03125, | |
| "epoch": 0.8648648648648649, | |
| "grad_norm": 0.5163156103091184, | |
| "kl": 0.0401763916015625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0059, | |
| "num_tokens": 9254845.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.1875, | |
| "step": 440 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 516.5, | |
| "epoch": 0.8668304668304668, | |
| "grad_norm": 0.5186482651887447, | |
| "kl": 0.0468902587890625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0162, | |
| "num_tokens": 9274655.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 441 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 473.25, | |
| "epoch": 0.8687960687960687, | |
| "grad_norm": 0.8560947365720074, | |
| "kl": 0.04534912109375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0613, | |
| "num_tokens": 9293249.0, | |
| "reward": 0.5, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.5, | |
| "step": 442 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 535.5625, | |
| "epoch": 0.8707616707616708, | |
| "grad_norm": 0.6757770984662097, | |
| "kl": 0.037872314453125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0215, | |
| "num_tokens": 9313597.0, | |
| "reward": 0.375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.375, | |
| "step": 443 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 603.34375, | |
| "epoch": 0.8727272727272727, | |
| "grad_norm": 0.6313377990331057, | |
| "kl": 0.03936767578125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0334, | |
| "num_tokens": 9337692.0, | |
| "reward": 0.125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.125, | |
| "step": 444 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 424.34375, | |
| "epoch": 0.8746928746928747, | |
| "grad_norm": 0.8947636846783437, | |
| "kl": 0.0523681640625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0032, | |
| "num_tokens": 9354627.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.46875, | |
| "step": 445 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 654.25, | |
| "epoch": 0.8766584766584766, | |
| "grad_norm": 0.610019658699382, | |
| "kl": 0.048675537109375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0163, | |
| "num_tokens": 9379179.0, | |
| "reward": 0.375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.375, | |
| "step": 446 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 611.46875, | |
| "epoch": 0.8786240786240787, | |
| "grad_norm": 0.5312983210093951, | |
| "kl": 0.0400543212890625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0072, | |
| "num_tokens": 9402732.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 447 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 546.8125, | |
| "epoch": 0.8805896805896806, | |
| "grad_norm": 0.5318339403144223, | |
| "kl": 0.0393829345703125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0054, | |
| "num_tokens": 9423636.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 448 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 489.03125, | |
| "epoch": 0.8825552825552826, | |
| "grad_norm": 0.624630511202983, | |
| "kl": 0.0647735595703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0367, | |
| "num_tokens": 9442853.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 449 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 597.90625, | |
| "epoch": 0.8845208845208845, | |
| "grad_norm": 0.6474131690597978, | |
| "kl": 0.039093017578125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0014, | |
| "num_tokens": 9467312.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 450 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 577.75, | |
| "epoch": 0.8864864864864865, | |
| "grad_norm": 0.7064702158831349, | |
| "kl": 0.0397491455078125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0133, | |
| "num_tokens": 9489042.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 451 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 502.34375, | |
| "epoch": 0.8884520884520885, | |
| "grad_norm": 0.7709467427427181, | |
| "kl": 0.04876708984375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0313, | |
| "num_tokens": 9510571.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 452 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 541.9375, | |
| "epoch": 0.8904176904176904, | |
| "grad_norm": 0.6518397938731665, | |
| "kl": 0.0745697021484375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0193, | |
| "num_tokens": 9531091.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 453 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 570.75, | |
| "epoch": 0.8923832923832924, | |
| "grad_norm": 0.31940008744943943, | |
| "kl": 0.0417938232421875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0062, | |
| "num_tokens": 9553943.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.4375, | |
| "step": 454 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 547.15625, | |
| "epoch": 0.8943488943488943, | |
| "grad_norm": 0.8236104881717373, | |
| "kl": 0.050628662109375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0016, | |
| "num_tokens": 9574864.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.34375, | |
| "step": 455 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 563.25, | |
| "epoch": 0.8963144963144963, | |
| "grad_norm": 0.6320533026964233, | |
| "kl": 0.040283203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0232, | |
| "num_tokens": 9596684.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 456 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 587.0, | |
| "epoch": 0.8982800982800982, | |
| "grad_norm": 0.6014697704100915, | |
| "kl": 0.0406036376953125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0124, | |
| "num_tokens": 9618830.0, | |
| "reward": 0.375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.375, | |
| "step": 457 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 646.71875, | |
| "epoch": 0.9002457002457003, | |
| "grad_norm": 0.7833382380682595, | |
| "kl": 0.0423583984375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0277, | |
| "num_tokens": 9643945.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.40625, | |
| "step": 458 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 616.46875, | |
| "epoch": 0.9022113022113022, | |
| "grad_norm": 0.6286029969787078, | |
| "kl": 0.0343017578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0398, | |
| "num_tokens": 9667142.0, | |
| "reward": 0.125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.125, | |
| "step": 459 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 540.90625, | |
| "epoch": 0.9041769041769042, | |
| "grad_norm": 0.8483549546208754, | |
| "kl": 0.042694091796875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0306, | |
| "num_tokens": 9687767.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.46875, | |
| "step": 460 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 573.53125, | |
| "epoch": 0.9061425061425061, | |
| "grad_norm": 0.9783818715960411, | |
| "kl": 0.0431671142578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0273, | |
| "num_tokens": 9709752.0, | |
| "reward": 0.375, | |
| "reward_std": 0.3535533845424652, | |
| "rewards/reward_func": 0.375, | |
| "step": 461 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 543.09375, | |
| "epoch": 0.9081081081081082, | |
| "grad_norm": 0.9101620155547488, | |
| "kl": 0.047393798828125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0418, | |
| "num_tokens": 9731845.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.28125, | |
| "step": 462 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 649.75, | |
| "epoch": 0.9100737100737101, | |
| "grad_norm": 0.5177724982627493, | |
| "kl": 0.041015625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0269, | |
| "num_tokens": 9757043.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 463 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 566.3125, | |
| "epoch": 0.912039312039312, | |
| "grad_norm": 0.5097590808471448, | |
| "kl": 0.0486907958984375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0478, | |
| "num_tokens": 9779425.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 464 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 574.28125, | |
| "epoch": 0.914004914004914, | |
| "grad_norm": 0.46044888779207277, | |
| "kl": 0.039031982421875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0038, | |
| "num_tokens": 9801234.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 465 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 572.0, | |
| "epoch": 0.9159705159705159, | |
| "grad_norm": 0.4446137407494509, | |
| "kl": 0.0450439453125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0019, | |
| "num_tokens": 9823266.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 466 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 521.6875, | |
| "epoch": 0.917936117936118, | |
| "grad_norm": 0.5512563995618973, | |
| "kl": 0.045654296875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0158, | |
| "num_tokens": 9843372.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.3125, | |
| "step": 467 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 581.3125, | |
| "epoch": 0.9199017199017199, | |
| "grad_norm": 0.5062880324256598, | |
| "kl": 0.04962158203125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0024, | |
| "num_tokens": 9866160.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 468 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 570.625, | |
| "epoch": 0.9218673218673219, | |
| "grad_norm": 0.5198694178078531, | |
| "kl": 0.0443115234375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0046, | |
| "num_tokens": 9888922.0, | |
| "reward": 0.375, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.375, | |
| "step": 469 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 518.78125, | |
| "epoch": 0.9238329238329238, | |
| "grad_norm": 0.4443759237259831, | |
| "kl": 0.042236328125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.004, | |
| "num_tokens": 9909305.0, | |
| "reward": 0.5, | |
| "reward_std": 0.0883883461356163, | |
| "rewards/reward_func": 0.5, | |
| "step": 470 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 521.34375, | |
| "epoch": 0.9257985257985258, | |
| "grad_norm": 0.39487768918762367, | |
| "kl": 0.0467071533203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0012, | |
| "num_tokens": 9929824.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.28125, | |
| "step": 471 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 639.15625, | |
| "epoch": 0.9277641277641278, | |
| "grad_norm": 0.8482745853224053, | |
| "kl": 0.041168212890625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0007, | |
| "num_tokens": 9956533.0, | |
| "reward": 0.25, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.25, | |
| "step": 472 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 570.0, | |
| "epoch": 0.9297297297297298, | |
| "grad_norm": 0.7070985297382011, | |
| "kl": 0.0414276123046875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0434, | |
| "num_tokens": 9978539.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 473 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 599.0625, | |
| "epoch": 0.9316953316953317, | |
| "grad_norm": 0.635901530369511, | |
| "kl": 0.0579833984375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0389, | |
| "num_tokens": 10001113.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.3125, | |
| "step": 474 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 653.4375, | |
| "epoch": 0.9336609336609336, | |
| "grad_norm": 0.6523995575575142, | |
| "kl": 0.0434722900390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0513, | |
| "num_tokens": 10025803.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 475 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 643.1875, | |
| "epoch": 0.9356265356265356, | |
| "grad_norm": 0.23781963453628815, | |
| "kl": 0.0393218994140625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0014, | |
| "num_tokens": 10049597.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.34375, | |
| "step": 476 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 489.34375, | |
| "epoch": 0.9375921375921376, | |
| "grad_norm": 0.5471556352566244, | |
| "kl": 0.045623779296875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0121, | |
| "num_tokens": 10069512.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 477 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 574.125, | |
| "epoch": 0.9395577395577396, | |
| "grad_norm": 0.27015180079377904, | |
| "kl": 0.039276123046875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0048, | |
| "num_tokens": 10092064.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.28125, | |
| "step": 478 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 623.75, | |
| "epoch": 0.9415233415233415, | |
| "grad_norm": 0.6058280367141352, | |
| "kl": 0.0426177978515625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.044, | |
| "num_tokens": 10115488.0, | |
| "reward": 0.375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.375, | |
| "step": 479 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 600.4375, | |
| "epoch": 0.9434889434889435, | |
| "grad_norm": 0.39339140830706126, | |
| "kl": 0.0403594970703125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.011, | |
| "num_tokens": 10138394.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.04419417306780815, | |
| "rewards/reward_func": 0.21875, | |
| "step": 480 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 549.375, | |
| "epoch": 0.9454545454545454, | |
| "grad_norm": 0.6076812699572645, | |
| "kl": 0.04254150390625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0152, | |
| "num_tokens": 10159542.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.28125, | |
| "step": 481 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 604.03125, | |
| "epoch": 0.9474201474201475, | |
| "grad_norm": 0.7186549326531969, | |
| "kl": 0.04241943359375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0531, | |
| "num_tokens": 10182213.0, | |
| "reward": 0.25, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.25, | |
| "step": 482 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 533.375, | |
| "epoch": 0.9493857493857494, | |
| "grad_norm": 0.8531762575906214, | |
| "kl": 0.045928955078125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0259, | |
| "num_tokens": 10202439.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.4375, | |
| "step": 483 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 616.9375, | |
| "epoch": 0.9513513513513514, | |
| "grad_norm": 0.6000292269955634, | |
| "kl": 0.0395965576171875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0127, | |
| "num_tokens": 10226203.0, | |
| "reward": 0.1875, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.1875, | |
| "step": 484 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 643.3125, | |
| "epoch": 0.9533169533169533, | |
| "grad_norm": 0.5586717727855203, | |
| "kl": 0.0501861572265625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0167, | |
| "num_tokens": 10251063.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 485 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 593.65625, | |
| "epoch": 0.9552825552825552, | |
| "grad_norm": 0.4552158754845552, | |
| "kl": 0.0469207763671875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0343, | |
| "num_tokens": 10273626.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 486 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 474.25, | |
| "epoch": 0.9572481572481573, | |
| "grad_norm": 0.6308023679429978, | |
| "kl": 0.0457763671875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0072, | |
| "num_tokens": 10292442.0, | |
| "reward": 0.375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.375, | |
| "step": 487 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 583.5, | |
| "epoch": 0.9592137592137592, | |
| "grad_norm": 0.723580982968801, | |
| "kl": 0.03936767578125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0084, | |
| "num_tokens": 10315494.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.40625, | |
| "step": 488 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 582.0, | |
| "epoch": 0.9611793611793612, | |
| "grad_norm": 0.5067316106579862, | |
| "kl": 0.043212890625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0016, | |
| "num_tokens": 10337536.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 489 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 513.0, | |
| "epoch": 0.9631449631449631, | |
| "grad_norm": 0.4993171072444713, | |
| "kl": 0.043701171875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0167, | |
| "num_tokens": 10357220.0, | |
| "reward": 0.59375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.59375, | |
| "step": 490 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 615.65625, | |
| "epoch": 0.9651105651105651, | |
| "grad_norm": 0.5190916238994441, | |
| "kl": 0.043243408203125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0308, | |
| "num_tokens": 10380595.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.34375, | |
| "step": 491 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 562.59375, | |
| "epoch": 0.9670761670761671, | |
| "grad_norm": 0.5941246052934998, | |
| "kl": 0.043060302734375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0256, | |
| "num_tokens": 10404092.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.3125, | |
| "step": 492 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 601.28125, | |
| "epoch": 0.9690417690417691, | |
| "grad_norm": 0.41176728740618557, | |
| "kl": 0.04107666015625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0093, | |
| "num_tokens": 10429171.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.40625, | |
| "step": 493 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 523.96875, | |
| "epoch": 0.971007371007371, | |
| "grad_norm": 0.6874286928007323, | |
| "kl": 0.061553955078125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0308, | |
| "num_tokens": 10449826.0, | |
| "reward": 0.34375, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.34375, | |
| "step": 494 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 575.4375, | |
| "epoch": 0.972972972972973, | |
| "grad_norm": 0.9358489974466774, | |
| "kl": 0.0450592041015625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.009, | |
| "num_tokens": 10471846.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.30935921147465706, | |
| "rewards/reward_func": 0.28125, | |
| "step": 495 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 606.46875, | |
| "epoch": 0.9749385749385749, | |
| "grad_norm": 0.8563492393734905, | |
| "kl": 0.0441436767578125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0106, | |
| "num_tokens": 10495427.0, | |
| "reward": 0.375, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.375, | |
| "step": 496 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 482.9375, | |
| "epoch": 0.9769041769041769, | |
| "grad_norm": 0.6591307258770829, | |
| "kl": 0.049560546875, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0025, | |
| "num_tokens": 10514793.0, | |
| "reward": 0.5, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.5, | |
| "step": 497 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 592.9375, | |
| "epoch": 0.9788697788697789, | |
| "grad_norm": 0.7690953881646339, | |
| "kl": 0.041473388671875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0479, | |
| "num_tokens": 10537763.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.3125, | |
| "step": 498 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 631.8125, | |
| "epoch": 0.9808353808353808, | |
| "grad_norm": 0.6185240661307275, | |
| "kl": 0.0496673583984375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0006, | |
| "num_tokens": 10561673.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.21875, | |
| "step": 499 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 481.875, | |
| "epoch": 0.9828009828009828, | |
| "grad_norm": 0.7706877175063906, | |
| "kl": 0.043121337890625, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0013, | |
| "num_tokens": 10582179.0, | |
| "reward": 0.28125, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.28125, | |
| "step": 500 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 502.25, | |
| "epoch": 0.9847665847665847, | |
| "grad_norm": 0.6632447682844715, | |
| "kl": 0.05169677734375, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0109, | |
| "num_tokens": 10601247.0, | |
| "reward": 0.46875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.46875, | |
| "step": 501 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 593.4375, | |
| "epoch": 0.9867321867321868, | |
| "grad_norm": 0.6536626884644745, | |
| "kl": 0.0438232421875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0153, | |
| "num_tokens": 10624387.0, | |
| "reward": 0.3125, | |
| "reward_std": 0.2651650384068489, | |
| "rewards/reward_func": 0.3125, | |
| "step": 502 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 598.03125, | |
| "epoch": 0.9886977886977887, | |
| "grad_norm": 0.7565109176505256, | |
| "kl": 0.042327880859375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0069, | |
| "num_tokens": 10647090.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 503 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 486.8125, | |
| "epoch": 0.9906633906633907, | |
| "grad_norm": 0.5326768203635671, | |
| "kl": 0.051788330078125, | |
| "learning_rate": 1e-06, | |
| "loss": -0.0165, | |
| "num_tokens": 10665776.0, | |
| "reward": 0.40625, | |
| "reward_std": 0.13258251920342445, | |
| "rewards/reward_func": 0.40625, | |
| "step": 504 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 552.03125, | |
| "epoch": 0.9926289926289926, | |
| "grad_norm": 0.5981367015619785, | |
| "kl": 0.0459136962890625, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0203, | |
| "num_tokens": 10687387.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 505 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 595.75, | |
| "epoch": 0.9945945945945946, | |
| "grad_norm": 0.5606366476723692, | |
| "kl": 0.0438995361328125, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0304, | |
| "num_tokens": 10710265.0, | |
| "reward": 0.4375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.4375, | |
| "step": 506 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 558.375, | |
| "epoch": 0.9965601965601966, | |
| "grad_norm": 0.6917851268358196, | |
| "kl": 0.0489349365234375, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0192, | |
| "num_tokens": 10732003.0, | |
| "reward": 0.375, | |
| "reward_std": 0.1767766922712326, | |
| "rewards/reward_func": 0.375, | |
| "step": 507 | |
| }, | |
| { | |
| "clip_ratio": 0.0, | |
| "completion_length": 680.9166831970215, | |
| "epoch": 0.9985257985257985, | |
| "grad_norm": 0.6271385318204094, | |
| "kl": 0.03857421875, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0521, | |
| "num_tokens": 10757622.0, | |
| "reward": 0.21875, | |
| "reward_std": 0.22097086533904076, | |
| "rewards/reward_func": 0.21875, | |
| "step": 508 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 508, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 200, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |