pruning for sparsity – turns out some LLMs work just as well if you set 60% of the weights to zero (though this likely isn’t true if you’re using Chinchilla-optimal training)

Nathan Labenz on AI pricing
from Tyler Cowen favicon