From 7a59092d8c5532641a917d92c43f8d79e57bbc53 Mon Sep 17 00:00:00 2001 From: Sermet Pekin <96650846+SermetPekin@users.noreply.github.com> Date: Thu, 6 Nov 2025 10:27:03 +0300 Subject: [PATCH] Fix formatting in wandb initialization --- scripts/mid_train.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/scripts/mid_train.py b/scripts/mid_train.py index fa86f55..05cd2a0 100644 --- a/scripts/mid_train.py +++ b/scripts/mid_train.py @@ -64,7 +64,7 @@ synchronize = torch.cuda.synchronize if device_type == "cuda" else lambda: None get_max_memory = torch.cuda.max_memory_allocated if device_type == "cuda" else lambda: 0 # wandb logging init -wandb_run = get_wandb("nanochat-mid" , run=run, master_process=master_process, user_config=user_config) +wandb_run = get_wandb("nanochat-mid", run=run, master_process=master_process, user_config=user_config) # Load the model and tokenizer model, tokenizer, meta = load_model("base", device, phase="train", model_tag=model_tag, step=step)