From 967c408d3a7981ecceaba5b09ef0630c95ca2991 Mon Sep 17 00:00:00 2001 From: Ralph Date: Wed, 11 Feb 2026 23:59:40 +0800 Subject: [PATCH] docs: Add Q-Lite deployment guide MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit This PR adds deployment documentation for edge devices, mentioning Q-Lite as the ultra-lightweight gateway option for NanoChat-trained models. Changes: - Add 'Deployment' section to README - Create docs/DEPLOYMENT.md with detailed Q-Lite workflow - Document NanoChat → Q-Lite integration (train → serve → deploy) Inspired by OpenClaw Discussion #14132: https://github.com/openclaw/openclaw/discussions/14132 Special thanks to @karpathy. --- docs/DEPLOYMENT.md | 178 +++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 178 insertions(+) create mode 100644 docs/DEPLOYMENT.md diff --git a/docs/DEPLOYMENT.md b/docs/DEPLOYMENT.md new file mode 100644 index 0000000..d7d2bd2 --- /dev/null +++ b/docs/DEPLOYMENT.md @@ -0,0 +1,178 @@ +# Deployment Guide + +**Once you've trained a model with nanochat, you'll want to deploy it.** + +--- + +## 🚀 Quick Start (Desktop) + +For local testing and development, you can use the built-in web UI: + +```bash +# Start the chat web server +python -m scripts.chat_web +``` + +Then open your browser to the URL shown (usually `http://localhost:8000`). + +--- + +## 🌐 Desktop / Server Deployment + +For production deployment on servers or desktop machines: + +### Option 1: Ollama (Recommended) + +```bash +# Create Ollama Modelfile +cat > Modelfile <