From 5c5cff25a5c176b64f15321d765958d6d7fc5241 Mon Sep 17 00:00:00 2001 From: Dipesh Babu Date: Mon, 9 Feb 2026 20:51:15 -0500 Subject: [PATCH] remove unused import --- nanochat/gpt.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/nanochat/gpt.py b/nanochat/gpt.py index cb1fe7f..1409d8c 100644 --- a/nanochat/gpt.py +++ b/nanochat/gpt.py @@ -12,7 +12,6 @@ Notable features: - Flash Attention 3 integration """ -from functools import partial from dataclasses import dataclass import torch @@ -22,7 +21,6 @@ import torch.nn.functional as F from nanochat.common import get_dist_info, print0 from nanochat.optim import MuonAdamW, DistMuonAdamW -from typing import Optional # Our custom Flash Attention module that automatically uses FA3 on Hopper+ and SDPA fallback elsewhere from nanochat.flash_attention import flash_attn