mirror of
https://github.com/karpathy/nanochat.git
synced 2026-05-09 01:10:10 +00:00
Model R6 (97% pass rate on 33-probe eval, val_bpb 0.2635): - modal/serve.py + modal/_tools.py: tool-aware streaming with TavilySearchBackend auto-detect, python_start/end state machine, output_start/end forcing; mount tavily secret - modal/serve.py: MODEL_TAG=d24-sft-r6, model path points at new SFT r6 - services/chat-api/routes/messages.py: accept thinking_mode flag, inject samosaChaat system prompt (direct or <think> variant) into first user message before streaming to Modal - services/frontend/components/chat/ChatInput.tsx: Brain toggle 'Think' button next to send; when active, model uses think mode - services/frontend/components/chat/ChatWindow.tsx: track thinkingMode state, pass through to API body as thinking_mode - services/frontend/components/chat/MessageBubble.tsx: parse and render <think>...</think> as collapsible italic blocks; <|python_start|>...<|python_end|> as tool-call cards with icons per tool name; <|output_start|>...<|output_end|> as result cards with expandable JSON - nanochat/tools.py: TavilySearchBackend class + auto-detect - nanochat/ui.html: legacy UI reasoning toggle (kept for parity) Tool execution verified live: query -> web_search via Tavily -> Macron returned with grounded answer.
248 lines
7.9 KiB
TypeScript
248 lines
7.9 KiB
TypeScript
'use client';
|
|
|
|
import { useCallback, useEffect, useMemo, useRef, useState } from 'react';
|
|
import { PanelLeftOpen } from 'lucide-react';
|
|
import MessageBubble from './MessageBubble';
|
|
import EmptyState from './EmptyState';
|
|
import ChatInput from './ChatInput';
|
|
import { useChatStore } from '@/store/chatStore';
|
|
import { useAuth } from '@/hooks/useAuth';
|
|
import { authHeaders } from '@/lib/auth-client';
|
|
import { parseSlashCommand } from '@/lib/slashCommands';
|
|
import type { Message } from '@/types/chat';
|
|
|
|
export default function ChatWindow() {
|
|
const { user } = useAuth();
|
|
const {
|
|
conversations,
|
|
currentConversationId,
|
|
model,
|
|
temperature,
|
|
topK,
|
|
sidebarOpen,
|
|
toggleSidebar,
|
|
createConversation,
|
|
newConversation,
|
|
appendMessage,
|
|
updateMessage,
|
|
setTemperature,
|
|
setTopK,
|
|
} = useChatStore();
|
|
|
|
const [draft, setDraft] = useState('');
|
|
const [streamingMsgId, setStreamingMsgId] = useState<string | null>(null);
|
|
const [thinkingMode, setThinkingMode] = useState(false);
|
|
const streamingBufferRef = useRef('');
|
|
const scrollRef = useRef<HTMLDivElement>(null);
|
|
|
|
const active = useMemo(
|
|
() => conversations.find((c) => c.id === currentConversationId) ?? null,
|
|
[conversations, currentConversationId],
|
|
);
|
|
|
|
const messages: Message[] = active?.messages ?? [];
|
|
const isEmpty = messages.length === 0;
|
|
|
|
const scrollToBottom = useCallback(() => {
|
|
const el = scrollRef.current;
|
|
if (el) el.scrollTop = el.scrollHeight;
|
|
}, []);
|
|
|
|
useEffect(() => {
|
|
scrollToBottom();
|
|
}, [messages.length, streamingMsgId, scrollToBottom]);
|
|
|
|
const [isStreaming, setIsStreaming] = useState(false);
|
|
const abortRef = useRef<AbortController | null>(null);
|
|
|
|
const stop = useCallback(() => {
|
|
abortRef.current?.abort();
|
|
abortRef.current = null;
|
|
setIsStreaming(false);
|
|
}, []);
|
|
|
|
const streamFromApi = useCallback(async (convId: string, assistantMsgId: string, content: string, temp?: number, topk?: number, thinking?: boolean) => {
|
|
stop();
|
|
const ac = new AbortController();
|
|
abortRef.current = ac;
|
|
setIsStreaming(true);
|
|
|
|
try {
|
|
const headers: Record<string, string> = {
|
|
'Content-Type': 'application/json',
|
|
...authHeaders(),
|
|
};
|
|
|
|
// Call chat-api directly via nginx — no Next.js proxy
|
|
const res = await fetch(`/api/conversations/${convId}/messages`, {
|
|
method: 'POST',
|
|
headers,
|
|
body: JSON.stringify({ content, temperature: temp, max_tokens: 512, top_k: topk, thinking_mode: !!thinking }),
|
|
signal: ac.signal,
|
|
});
|
|
|
|
if (!res.ok || !res.body) {
|
|
throw new Error(`HTTP ${res.status}`);
|
|
}
|
|
|
|
const reader = res.body.getReader();
|
|
const decoder = new TextDecoder();
|
|
let buffer = '';
|
|
|
|
while (true) {
|
|
const { done, value } = await reader.read();
|
|
if (done) break;
|
|
|
|
buffer += decoder.decode(value, { stream: true });
|
|
let nl: number;
|
|
while ((nl = buffer.indexOf('\n')) !== -1) {
|
|
const line = buffer.slice(0, nl).trim();
|
|
buffer = buffer.slice(nl + 1);
|
|
if (!line.startsWith('data:')) continue;
|
|
const payload = line.slice(5).trim();
|
|
if (!payload) continue;
|
|
try {
|
|
const data = JSON.parse(payload);
|
|
if (data.done) {
|
|
setStreamingMsgId(null);
|
|
streamingBufferRef.current = '';
|
|
setIsStreaming(false);
|
|
return;
|
|
}
|
|
if (typeof data.token === 'string') {
|
|
streamingBufferRef.current += data.token;
|
|
updateMessage(convId, assistantMsgId, streamingBufferRef.current);
|
|
}
|
|
} catch { /* skip malformed */ }
|
|
}
|
|
}
|
|
|
|
setStreamingMsgId(null);
|
|
streamingBufferRef.current = '';
|
|
} catch (err) {
|
|
if ((err as Error).name !== 'AbortError') {
|
|
console.error('[chat] stream error:', err);
|
|
if (assistantMsgId && convId) {
|
|
updateMessage(convId, assistantMsgId, `Error: ${(err as Error).message}`);
|
|
}
|
|
}
|
|
setStreamingMsgId(null);
|
|
streamingBufferRef.current = '';
|
|
} finally {
|
|
setIsStreaming(false);
|
|
abortRef.current = null;
|
|
}
|
|
}, [stop, updateMessage]);
|
|
|
|
const ensureConversation = useCallback(async () => {
|
|
if (currentConversationId) return currentConversationId;
|
|
// Try creating via the API first
|
|
const apiId = await createConversation();
|
|
if (apiId) return apiId;
|
|
// Fallback to local-only conversation (mock mode)
|
|
return newConversation();
|
|
}, [currentConversationId, createConversation, newConversation]);
|
|
|
|
const handleSend = useCallback(
|
|
async (rawInput?: string) => {
|
|
const text = (rawInput ?? draft).trim();
|
|
if (!text || isStreaming) return;
|
|
|
|
const convId = await ensureConversation();
|
|
|
|
const slash = parseSlashCommand(text, { temperature, topK });
|
|
if (slash.handled) {
|
|
setDraft('');
|
|
if (slash.setTemperature !== undefined) setTemperature(slash.setTemperature);
|
|
if (slash.setTopK !== undefined) setTopK(slash.setTopK);
|
|
if (slash.clear) {
|
|
const apiId = await createConversation();
|
|
if (!apiId) newConversation();
|
|
return;
|
|
}
|
|
if (slash.consoleMessage) {
|
|
appendMessage(convId, { role: 'console', content: slash.consoleMessage });
|
|
}
|
|
return;
|
|
}
|
|
|
|
setDraft('');
|
|
appendMessage(convId, { role: 'user', content: text });
|
|
|
|
const assistantId = appendMessage(convId, { role: 'assistant', content: '' });
|
|
setStreamingMsgId(assistantId);
|
|
streamingBufferRef.current = '';
|
|
|
|
await streamFromApi(convId, assistantId, text, temperature, topK, thinkingMode);
|
|
},
|
|
[
|
|
draft,
|
|
isStreaming,
|
|
ensureConversation,
|
|
temperature,
|
|
topK,
|
|
thinkingMode,
|
|
appendMessage,
|
|
streamFromApi,
|
|
setTemperature,
|
|
setTopK,
|
|
createConversation,
|
|
newConversation,
|
|
],
|
|
);
|
|
|
|
return (
|
|
<section className="flex-1 flex flex-col min-w-0 bg-white dark:bg-ink">
|
|
<header className="flex items-center justify-between px-4 md:px-6 py-3 border-b border-cream-border dark:border-ink-border">
|
|
<div className="flex items-center gap-3">
|
|
{!sidebarOpen && (
|
|
<button
|
|
type="button"
|
|
onClick={toggleSidebar}
|
|
aria-label="Open sidebar"
|
|
className="p-1.5 rounded-md hover:bg-cream dark:hover:bg-ink-elev text-brown-light dark:text-ink-text-soft"
|
|
>
|
|
<PanelLeftOpen size={18} />
|
|
</button>
|
|
)}
|
|
<span className="inline-flex items-center gap-1.5 text-xs px-2.5 py-1 rounded-full border border-warm-grey dark:border-ink-border bg-cream-light dark:bg-ink-soft text-brown dark:text-ink-text-soft">
|
|
<span className="w-1.5 h-1.5 rounded-full bg-saffron" />
|
|
{model}
|
|
</span>
|
|
</div>
|
|
<div className="text-sm text-gray-600 dark:text-ink-text-soft font-medium">
|
|
{user?.name ? `Hi, ${user.name.split(' ')[0]}` : ''}
|
|
</div>
|
|
</header>
|
|
|
|
<div ref={scrollRef} className="flex-1 overflow-y-auto nice-scrollbar">
|
|
<div className="max-w-3xl mx-auto px-4 md:px-6 py-6 flex flex-col min-h-full">
|
|
{isEmpty ? (
|
|
<EmptyState onPick={(p) => handleSend(p)} />
|
|
) : (
|
|
<div className="flex flex-col">
|
|
{messages.map((m) => (
|
|
<MessageBubble
|
|
key={m.id}
|
|
message={m}
|
|
isStreaming={streamingMsgId === m.id && isStreaming}
|
|
/>
|
|
))}
|
|
</div>
|
|
)}
|
|
</div>
|
|
</div>
|
|
|
|
<ChatInput
|
|
value={draft}
|
|
onChange={setDraft}
|
|
onSubmit={() => handleSend()}
|
|
onStop={stop}
|
|
isStreaming={isStreaming}
|
|
thinkingMode={thinkingMode}
|
|
onToggleThinking={() => setThinkingMode((v) => !v)}
|
|
/>
|
|
</section>
|
|
);
|
|
}
|