Tech

‘Catastrophic overtraining’ could harm large language AI models that are trained on more data for the sake of training

Share
Share


  • Researchers from top US universities warn extending pre-training can be detrimental to performance
  • Too much pre-training can deliver worse performance due to something akin to the butterfly effect
  • The more they are pre-trained, the more they become sensitive to small changes that could disrupt the end result

Researchers from Carnegie Mellon, Stanford, Harvard, and Princeton are challenging one of AI development’s accepted core beliefs – that the more pre-training data the better the performance.

As reported by HPCwire, a new paper discuses the concept of “catastrophic overtraining,” whereby extended pre-training can harm a model’s performance after fine-tuning.

Share

Leave a comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Related Articles
We just got another big hint that the Samsung Galaxy S25 FE is on the way
Tech

We just got another big hint that the Samsung Galaxy S25 FE is on the way

References to Galaxy S25 FE firmware have appeared The phone could launch...

You won’t believe what 700+ projectors and AI can do in Abu Dhabi’s new immersive art world
Tech

You won’t believe what 700+ projectors and AI can do in Abu Dhabi’s new immersive art world

Over 700 Epson projectors transform walls into moving, responsive works of living...

When the school bell rings, the bandwidth drops: How post-15:40 internet surges affect UK broadband quality
Tech

When the school bell rings, the bandwidth drops: How post-15:40 internet surges affect UK broadband quality

Half of parents work after school, causing a broadband battle with streaming-addicted...