AI Codex
Foundation Models & LLMsDevelopersCTOs

Model Distillation

Also: knowledge distillation

Training a smaller 'student' model to replicate the behavior of a larger 'teacher' model — producing smaller faster models that retain most of the original's capability.