Deploy and fine-tune small language models entirely on your own hardware. Over five weeks, go from ML baselines through the SLM ecosystem, local quantized deployment, LoRA fine-tuning, and containerized serving — ending with an air-gapped support bot.