diff --git a/README.md b/README.md index 7126429..bc01055 100644 --- a/README.md +++ b/README.md @@ -6,16 +6,6 @@ This repo is a full-stack implementation of an LLM like ChatGPT in a single, clean, minimal, hackable, dependency-lite codebase. nanochat is designed to run on a single 8XH100 node via scripts like [speedrun.sh](speedrun.sh), that run the entire pipeline start to end. This includes tokenization, pretraining, finetuning, evaluation, inference, and web serving over a simple UI so that you can talk to your own LLM just like ChatGPT. nanochat will become the capstone project of the course LLM101n being developed by Eureka Labs. ---- - -### Architecture Overview - -Here’s an overview of the nanochat architecture: - -![nanochat architecture](dev/nanochat_architecture.jpg) - ---- - ## Quick start The fastest way to feel the magic is to run the speedrun script [speedrun.sh](speedrun.sh), which trains and inferences the $100 tier of nanochat. On an 8XH100 node at $24/hr, this gives a total run time of about 4 hours. Boot up a new 8XH100 GPU box from your favorite provider (e.g. I use and like [Lambda](https://lambda.ai/service/gpu-cloud)), and kick off the training script: diff --git a/dev/nanochat_architecture.jpg b/dev/nanochat_architecture.jpg deleted file mode 100644 index 066a0c7..0000000 Binary files a/dev/nanochat_architecture.jpg and /dev/null differ