Tech

‘Catastrophic overtraining’ could harm large language AI models that are trained on more data for the sake of training

Share
Share


  • Researchers from top US universities warn extending pre-training can be detrimental to performance
  • Too much pre-training can deliver worse performance due to something akin to the butterfly effect
  • The more they are pre-trained, the more they become sensitive to small changes that could disrupt the end result

Researchers from Carnegie Mellon, Stanford, Harvard, and Princeton are challenging one of AI development’s accepted core beliefs – that the more pre-training data the better the performance.

As reported by HPCwire, a new paper discuses the concept of “catastrophic overtraining,” whereby extended pre-training can harm a model’s performance after fine-tuning.

Share

Leave a comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Related Articles
IPVanish’s malware protection confirmed among the best on the market
Tech

IPVanish’s malware protection confirmed among the best on the market

IPVanish’s malware and tracker blocking technology has been confirmed as one of...

Removable battery is one very useful feature that this Samsung rugged smartphone brings to the table
Tech

Removable battery is one very useful feature that this Samsung rugged smartphone brings to the table

Samsung Galaxy XCover7 Pro revives the removable battery for power in rugged...

You freak out when battery life hits 38%, but here’s how to extend it and calm the heck down
Tech

You freak out when battery life hits 38%, but here’s how to extend it and calm the heck down

The moment when you think it’s time to start charging your smartphone...

Opera Mini stuffs a whole AI assistant into a tiny Android browser
Tech

Opera Mini stuffs a whole AI assistant into a tiny Android browser

Opera has added its AI assistant Aria to the Opera Mini browser...