• catloaf@lemm.ee
    link
    fedilink
    English
    arrow-up
    21
    arrow-down
    5
    ·
    3 months ago

    AI doesn’t grok anything. It doesn’t have any capability of understanding at all. It’s a Markov chain on steroids.

    • Thorny_Insight@lemm.ee
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      11
      ·
      edit-2
      3 months ago

      …is how generative-AI haters redefine terms and move the goalposts to fight their cognitive dissonance.

      • yesman@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        4
        ·
        3 months ago

        I read the abstract, and the connection to your title is a mystery. Are you using “grock” as in “transcendental understanding” or as Musk’s branded AI?

        • Hackworth@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          17
          ·
          3 months ago

          No c, just grok, originally from Stranger in a Strange Land. But a more technical definition is provided and expanded upon in the paper. Mystery easily dispelled!

          • yesman@lemmy.world
            link
            fedilink
            English
            arrow-up
            9
            arrow-down
            4
            ·
            3 months ago

            In that case I refer you to u/catloaf 's post. A machine cannot grock, not at any speed.

          • Blueberrydreamer@lemmynsfw.com
            link
            fedilink
            English
            arrow-up
            7
            arrow-down
            4
            ·
            3 months ago

            Thanks for clarifying, now please refer to the poster’s original statement:

            AI doesn’t grok anything. It doesn’t have any capability of understanding at all. It’s a Markov chain on steroids.

            • Hackworth@lemmy.worldOP
              link
              fedilink
              English
              arrow-up
              5
              arrow-down
              10
              ·
              3 months ago

              We follow the classic experimental paradigm reported in Power et al. (2022) for analyzing “grokking”, a poorly understood phenomenon in which validation accuracy dramatically improves long after the train loss saturates. Unlike the previous templates, this one is more amenable to open-ended empirical analysis (e.g. what conditions grokking occurs) rather than just trying to improve performance metrics

              • catloaf@lemm.ee
                link
                fedilink
                English
                arrow-up
                5
                arrow-down
                3
                ·
                3 months ago

                Oh okay so they’re just redefining words that are already well-defined so they can make fancy claims.

                • Hackworth@lemmy.worldOP
                  link
                  fedilink
                  English
                  arrow-up
                  4
                  arrow-down
                  4
                  ·
                  3 months ago

                  Well-defined for casual use is very different than well-defined for scholarly research. It’s standard practice to take colloquial vocab and more narrowly define it for use within a scientific discipline. Sometimes different disciplines will narrowly define the same word two different ways, which makes interdisciplinary communication pretty funny.

    • Hackworth@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      11
      ·
      3 months ago

      I appreciate it. I’ve had little luck engaging people in conversation about AI research in general. Since abandoning reddit, I’ve been litmus testing other platforms. I’m afraid this puts lemmy at about a 12. /shrug. Still better than reddit.

      • Thorny_Insight@lemm.ee
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        5
        ·
        3 months ago

        I couldn’t think of a worse platform to try and discuss this topic than Lemmy. The consensus here is essentially that big companies = bad, AI companies = big, and thus AI = bad.

            • Hackworth@lemmy.worldOP
              link
              fedilink
              English
              arrow-up
              2
              ·
              3 months ago

              You’d think, but not really. /r/chatgpt and a couple AI art subs are the only active ones. And “active” means bots and parrots in this case.

              • Thorny_Insight@lemm.ee
                link
                fedilink
                English
                arrow-up
                2
                ·
                edit-2
                3 months ago

                Well, I’m on Lemmy myself, so perhaps that’s some sort of an indication of where I prefer to discuss thing with people in general, not just about AI. My list of blocked users is rather vast though, so a big part of the loudest haters are being filtered out from my feed. That surely contributes to the better experience here - or atleast less bad.

                Definitely prefered it over there at reddit, but I’m a man of principle so I’m not going back either.