This summary provides an overview of the text, highlighting key points about building a novel routing framework for Large Language Models (LLMs) using human preference data. The framework directs simple queries to more cost-effective models while maintaining high response quality. The tutorial covers every step from data labeling and fine-tuning LLMs to offline evaluation and conducting offline evaluations on standard benchmarks. It also discusses the importance of balancing the dataset and optimizing inference speed. The final section evaluates the performance of the router against a random router on GSM8K, demonstrating its effectiveness in out-of-domain generalization.