• 𝕊𝕚𝕤𝕪𝕡𝕙𝕖𝕒𝕟@programming.devOPM
      link
      fedilink
      English
      arrow-up
      4
      ·
      edit-2
      1 year ago

      Or when it tells you that it can do something it actually can’t, and it hallucinates like crazy. In the early days of ChatGPT I asked it to summarize an article at a link, and it gave me a very believable but completely false summary based on the words in the URL.

      This was the first time I saw wild hallucination. It was astounding.

      • Phoenix@programming.dev
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        It’s even better when you ask it to write code for you, it generates a decent looking block, but upon closer inspection it imports a nonexistent library that just happens to do exactly what you were looking for.

        That’s the best sort of hallucination, because it gets your hopes up.

        • 𝕊𝕚𝕤𝕪𝕡𝕙𝕖𝕒𝕟@programming.devOPM
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          Yes, for a moment you think “oh, there’s such a convenient API for this” and then you realize…

          But we programmers can at least compile/run the code and find out if it’s wrong (most of the time). It is much harder in other fields.