distillation – a technique by which larger, more capable models can be used to train smaller models to similar performance within certain domains – Replit has a great writeup on how they created their first release codegen model in just a few weeks this way!

Nathan Labenz on AI pricing
from Tyler Cowen favicon