mirror of
https://github.com/karpathy/nanochat.git
synced 2026-02-02 00:29:52 +00:00
Fix generate() crash when top_k=0
Prevent a crash in generate() by skipping top-k filtering when top_k is set to 0
This commit is contained in:
parent
c88bbf8133
commit
d4d014010b
|
|
@ -489,7 +489,7 @@ class GPT(nn.Module):
|
|||
for _ in range(max_tokens):
|
||||
logits = self.forward(ids) # (B, T, vocab_size)
|
||||
logits = logits[:, -1, :] # (B, vocab_size)
|
||||
if top_k is not None:
|
||||
if top_k is not None and top_k > 0:
|
||||
v, _ = torch.topk(logits, min(top_k, logits.size(-1)))
|
||||
logits[logits < v[:, [-1]]] = -float('Inf')
|
||||
if temperature > 0:
|
||||
|
|
|
|||
Loading…
Reference in New Issue
Block a user