• BananaTrifleViolin@lemmy.world
    link
    fedilink
    English
    arrow-up
    19
    ·
    9 months ago

    The “solutions” to model collapse - essentially retraining on the original data set - suggests LLMs plateau or deteriorate. Especially without a way to separate out good and bad quality data (or ad they euohemistically try and say human vs AI data).

    Were increasingly seeing the limitations and flaws with LLMs. “Hallucinations” or better described as serious errors, model collapse and complete collapse suggest the current approach to LLMs is probably not going to lead to some gone of general AI. We have models we don’t really understand that have fundamental flaws and limitations.

    Unsurprising that they probably can’t live up to the hype.

    • zwaetschgeraeuber@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      8
      ·
      9 months ago

      Even if it will plateau, same was said with moorrs law, which held up way longer than expected. There are so many ways to improve this. Open source community is getting to the point where you can actually run decent models on normal private hardware (talking about 70-120b model)