The “catastrophic oversight” could damage the models of the large language that are trained in more data for the sake of training


  • Researchers from the best American universities warn that extending pre-training can be harmful to performance
  • Too pre-training can offer worse performance due to something similar to the butterfly effect
  • The more they train, the more they become sensitive to small changes that could interrupt the final result

Carnegie Mellon, Stanford, Harvard and Princeton researchers are challenging one of the accepted central beliefs of Ai Development, which the more training data, the better the performance will be.

As reported by HPCWIREA new document describes the concept of “catastrophic over -entry”, whereby the extended pretending can damage the performance of a model after adjustment.

Leave a Comment

Your email address will not be published. Required fields are marked *