The way double descent is normally presented, increasing the number of model parameters can make performance worse before it gets better. But there is another even more shocking phenomenon called data double descent, where increasing the number of training samples can cause performance to get worse before it gets better. These two phenomena are essentially mirror images of each other. That’s because the explosion in test error depends on the ratio of parameters to training samples.
Double descent in human learning
from chris-said.io
Filed under:
Same Source
Related Notes
- Often, people who don’t have access to the raw data expect one narr...from Josh Beckman
- The upshot for the industry at large, is: the **LLM-as-Moat model h...from Steve Yegge
- The first image ever transmitted to Earth from another planet was r...from Instagram
- My experience is companies do not anticipate that the cost of monit...from Mathew Duggan
- Ad-hoc validation leads to a phenomenon that the [language-theoreti...from Alexis King
- the difference between validation and parsing lies almost entirely ...from Alexis King
- few days ago I did something that I never thought I’d do again, and...from brandur.org
- But you can already see the idea of a “prompt” evolving into someth...from Ryan Broderick