pruning for sparsity – turns out some LLMs work just as well if you set 60% of the weights to zero (though this likely isn’t true if you’re using Chinchilla-optimal training)
Nathan Labenz on AI pricing
from Tyler Cowen
Filed under:
Same Source
Related Notes
- Original layout ![](https://www.joelsimon.net/imgs/evo_plans/resul...from joelsimon.net
- The upshot for the industry at large, is: the **LLM-as-Moat model h...from Steve Yegge
- More things than you would think are dynamic strategic problems. If...from marcelo.rinesi
- We don't quite know what to do with language models yet. But we...from Maggie Appleton
- These are wonderful non-chat interfaces for LLMs that I would total...from Maggie Appleton
- combining search and AI chat is actually the wrong way to go and I ...from Garbage Day
- the tech am I digging recently is a software framework called **Lan...from Interconnected
- Part of what makes LoRA so effective is that - like other forms of ...from Dylan Patel