From cc2f2abdf02f90e8cd3d5dd4cfbc37d0eb7b0ddc Mon Sep 17 00:00:00 2001 From: gio Date: Mon, 27 Apr 2026 15:03:10 -0500 Subject: [PATCH] speedrun.sh: use --num-iterations=6000 (88 min recipe, CORE 0.2646) --- runs/speedrun.sh | 11 ++++++----- 1 file changed, 6 insertions(+), 5 deletions(-) diff --git a/runs/speedrun.sh b/runs/speedrun.sh index ff7f285f..5ba55a74 100644 --- a/runs/speedrun.sh +++ b/runs/speedrun.sh @@ -69,11 +69,12 @@ python -m scripts.tok_eval echo "Waiting for dataset download to complete..." wait $DATASET_DOWNLOAD_PID -# d22 model (slightly overtrained to beat GPT-2 => increase data:params ratio from compute optimal 10.5 (default) to 12). -# Mirror of Run 6's d24+ratio=8 strategy from the other side of compute-optimal — d22 is below GPT-2 capability, -# so we overtrain rather than undertrain. Combined with --warmdown-ratio=0.85 (longer low-LR tail) and -# --muon-qk-clip-tau=100 (Kimi K2 §A QK-Clip) the recipe crosses GPT-2 CORE in 3.3% less wall-clock than Run 6. -torchrun --standalone --nproc_per_node=8 -m scripts.base_train -- --depth=22 --target-param-data-ratio=12 --total-batch-size=1048576 --device-batch-size=16 --warmdown-ratio=0.85 --muon-qk-clip-tau=100 --fp8 --run=$WANDB_RUN +# d22 model trained for 6000 iterations at 1M tokens/iter = 6B tokens (~ratio=11 against d22's +# scaling params, mirror of Run 6's d24+ratio=8 strategy from the other side of compute-optimal — +# d22 is below GPT-2 capability so we overtrain). Combined with --warmdown-ratio=0.85 (longer +# low-LR tail) and --muon-qk-clip-tau=100 (Kimi K2 §A QK-Clip) the recipe crosses GPT-2 CORE +# in 88 min — ~10.8% less wall-clock than Run 6 — at CORE 0.2646, val_bpb 0.7241. +torchrun --standalone --nproc_per_node=8 -m scripts.base_train -- --depth=22 --num-iterations=6000 --total-batch-size=1048576 --device-batch-size=16 --warmdown-ratio=0.85 --muon-qk-clip-tau=100 --fp8 --run=$WANDB_RUN # evaluate the model: CORE metric, BPB on train/val, and draw samples torchrun --standalone --nproc_per_node=8 -m scripts.base_eval -- --device-batch-size=16