• Bloonface@kbin.social
        link
        fedilink
        arrow-up
        6
        ·
        2 years ago

        Yeah but you can tell from the context that search results are just a list of random web pages that maybe what Google says is bollocks.

        Google gives you a bunch of results and says “here, look at these”. LLMs confidently tell you things that they may have simply made up and present them as if they’re real.

        • Fizz@lemmy.nz
          link
          fedilink
          arrow-up
          1
          ·
          2 years ago

          Its an infant tech so there’s obviously issues but if you go ask chat gpt how do I play poker you will get a short concise and accurate answer. In my opinion that’s the strength of ai as a search engine. Sometimes I just want an answer to my question and don’t want to dig through multiple webpages all filled with ads and hostile design.

      • noodlejetski@kbin.social
        link
        fedilink
        arrow-up
        4
        ·
        edit-2
        2 years ago

        Google actually pulls results from web pages.

        you know how some smartphone keyboards predict the next word that you’re going to use, and you can form a comprehensible sentence that sometimes even makes sense by simply tapping the next word on the prediction bar over and over? that’s what those language models do. they don’t actually search for anything, they just create sequences of words that sound probable.

        • Fizz@lemmy.nz
          link
          fedilink
          arrow-up
          1
          ·
          2 years ago

          It seems that Bing chat bot searches then reads the results and gives you the answer.

          I know it’s basically predictive text but if the prompt contains a relevant info then the predictive text is likely to be the answer you’re looking for so it works well.