This repository contains the source content for LLM Inference Handbook, a practical guide for understanding, optimizing, scaling, and operating LLM inference.
To preview the site locally:
pnpm install
pnpm start
It will be running at http://localhost:3000/llm/.
Contributions are welcome! Feel free to open issues, suggest improvements, or submit pull requests.
This repository is dual-licensed:
- All files in the
docs/
folder are licensed under the Creative Commons Attribution 4.0 International (CC BY 4.0) License. - All other files are licensed under the Apache License 2.0.