Model preparation
Model preparation is the process of preparing the model for inference.
📄️ LLM fine-tuning
Understand LLM fine-tuning and different fine-tuning frameworks.
📄️ LLM quantization
Understand LLM quantization and different quantization formats and methods.
📄️ LLM distillation
Learn how LLM distillation works, how it compares to quantization, and how to use it to build smaller, faster, and more efficient models for inference.
Stay updated with the handbook
Get the latest insights and updates on LLM inference and optimization techniques.
- Monthly insights
- Latest techniques
- Handbook updates