This repository contains the source content for LLM Inference Handbook, a practical guide for understanding, optimizing, scaling, and operating LLM inference.
To preview the site locally:
pnpm install
pnpm start
It will be running at http://localhost:3000/llm/.
Contributions are welcome! Feel free to open issues, suggest improvements, or submit pull requests.
This project is licensed under the MIT License.