3/26/2026 at 5:49:58 PM
>Knowledge distillation works like this: you take a large model, have it perform tasks with detailed reasoning, then feed those reasoning traces to a smaller model until the student learns to mimic the teacher. The smaller model ends up far more capable than if you’d trained it from scratch on the same data. Apple can now do this with the full Gemini, not just their own in-house models, and the distilled output runs locally. No internet required.No freaking way. AI companies see this as tantamount to pirating their models. There is no way that Google is not explicitly banning this in their agreement to allow Apple use their models.
by NewsaHackO
3/26/2026 at 6:40:18 PM
Here is a technical comparison to show how Knowledge Distillation works for on-device AI. It has visual maps for the Teacher vs Student models and explains things like 'Dark Knowledge' - https://vectree.io/compare/apple-intelligence-vs-knowledge-d...by functional_dev