mirror of
https://github.com/karpathy/nanochat.git
synced 2026-05-12 10:50:16 +00:00
Build services/frontend/ replacing the legacy nanochat/ui.html single-file UI. Landing, login, and chat pages ported with full design system: Devanagari + Great Vibes hero, samosa/chai/toran SVG animations, gold/cream palette. - App Router pages: / (hero + floating illustrations), /login (split-screen OAuth with mandala motif), /chat (260px collapsible sidebar, suggestion chips, markdown + code-copy, auto-expanding input, slash commands) - SSE streaming via useSSE hook and /api/chat/stream BFF route (proxies to CHAT_API_URL when set, falls back to mock echo for local dev) - NextAuth.js v5 with Google + GitHub providers; middleware gates /chat/* - Zustand store with localStorage persistence for conversations/settings - Tailwind theme carries all ui.html tokens + keyframes (pendulum, float, wobble, steamFloat, steamType); SVG assets componentized under components/svg - Multi-stage node:20-alpine Dockerfile with Next standalone output Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
96 lines
2.6 KiB
TypeScript
96 lines
2.6 KiB
TypeScript
import { NextRequest } from 'next/server';
|
|
|
|
export const runtime = 'nodejs';
|
|
export const dynamic = 'force-dynamic';
|
|
|
|
interface StreamBody {
|
|
messages: Array<{ role: string; content: string }>;
|
|
model?: string;
|
|
temperature?: number;
|
|
topK?: number;
|
|
}
|
|
|
|
const encoder = new TextEncoder();
|
|
|
|
function sseEvent(data: Record<string, unknown>) {
|
|
return encoder.encode(`data: ${JSON.stringify(data)}\n\n`);
|
|
}
|
|
|
|
async function proxyUpstream(body: StreamBody, upstreamUrl: string) {
|
|
const upstream = await fetch(upstreamUrl, {
|
|
method: 'POST',
|
|
headers: { 'Content-Type': 'application/json' },
|
|
body: JSON.stringify({
|
|
messages: body.messages,
|
|
temperature: body.temperature ?? 0.8,
|
|
top_k: body.topK ?? 50,
|
|
max_tokens: 512,
|
|
model: body.model,
|
|
}),
|
|
});
|
|
|
|
if (!upstream.ok || !upstream.body) {
|
|
throw new Error(`upstream HTTP ${upstream.status}`);
|
|
}
|
|
|
|
return new Response(upstream.body, {
|
|
headers: {
|
|
'Content-Type': 'text/event-stream',
|
|
'Cache-Control': 'no-cache, no-transform',
|
|
Connection: 'keep-alive',
|
|
},
|
|
});
|
|
}
|
|
|
|
function mockEcho(body: StreamBody): Response {
|
|
const last = body.messages[body.messages.length - 1]?.content ?? '';
|
|
const greetings = [
|
|
'Namaste! ',
|
|
"Here's what I can offer for that question: ",
|
|
"Let's think about it together. ",
|
|
];
|
|
const greeting = greetings[Math.floor(Math.random() * greetings.length)];
|
|
const echo = last.trim() ? `You asked: "${last.trim()}".` : 'I am listening.';
|
|
const full = `${greeting}${echo}\n\nThis is a mock response from the samosaChaat frontend — once the chat service is wired, real streaming tokens will land here.`;
|
|
|
|
const stream = new ReadableStream({
|
|
async start(controller) {
|
|
const words = full.split(/(\s+)/);
|
|
for (const w of words) {
|
|
controller.enqueue(sseEvent({ token: w, gpu: 0 }));
|
|
await new Promise((r) => setTimeout(r, 25));
|
|
}
|
|
controller.enqueue(sseEvent({ done: true }));
|
|
controller.close();
|
|
},
|
|
});
|
|
|
|
return new Response(stream, {
|
|
headers: {
|
|
'Content-Type': 'text/event-stream',
|
|
'Cache-Control': 'no-cache, no-transform',
|
|
Connection: 'keep-alive',
|
|
},
|
|
});
|
|
}
|
|
|
|
export async function POST(req: NextRequest) {
|
|
let body: StreamBody;
|
|
try {
|
|
body = (await req.json()) as StreamBody;
|
|
} catch {
|
|
return new Response('Invalid JSON', { status: 400 });
|
|
}
|
|
|
|
const upstream = process.env.CHAT_API_URL;
|
|
if (upstream) {
|
|
try {
|
|
return await proxyUpstream(body, `${upstream.replace(/\/$/, '')}/chat/completions`);
|
|
} catch (err) {
|
|
console.warn('[chat/stream] upstream failed, falling back to mock:', err);
|
|
}
|
|
}
|
|
|
|
return mockEcho(body);
|
|
}
|