mirror of
https://github.com/karpathy/nanochat.git
synced 2026-02-05 10:09:57 +00:00
Update nanochat/flash_attention.py
Code review suggestion Co-authored-by: Sofie Van Landeghem <svlandeg@users.noreply.github.com>
This commit is contained in:
parent
2e45b7800a
commit
4e70a2b678
|
|
@ -25,7 +25,7 @@ def _load_flash_attention_3():
|
|||
if not torch.cuda.is_available():
|
||||
return None
|
||||
try:
|
||||
major, minor = torch.cuda.get_device_capability()
|
||||
major, _ = torch.cuda.get_device_capability()
|
||||
# FA3 kernels are compiled for Hopper (sm90) only
|
||||
# Ada (sm89), Blackwell (sm100) need SDPA fallback until FA3 is recompiled
|
||||
if major != 9:
|
||||
|
|
|
|||
Loading…
Reference in New Issue
Block a user