• Jesus_666@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    3 days ago

    Mind you, LLMs can be quite inconsistent. If you repeat the same question in new chats, you can easily get a mix of good answers, bad answers, bafflingly insane answers, and “I’m sorry but I cannot support terrorism”.

    • glimse@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      3 days ago

      Yeah, that is definitely true. I just wanted to take back my knee jerk reaction of “OF COURSE CHATGPT WOULD SPIT OUT SOMETHING SO STUPID” after being proven wrong.

      I tried it both on my phone and work PC and got different but equally decent replies. Nothing you couldn’t find elsewhere without boiling the planet, though

      • Promethiel@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        2 days ago

        That’s the true danger of the tools, imo. They aren’t the digital All Seers their makers want to market them as, but they also aren’t the utterly useless slop machines the consensus on the Fediverse appears to vehemently be.

        Of course they’re somewhere in-between. Spicy auto complete sounds like a put down but there is spice. I prefer to think of LLMs as word calculators, and I do mean that I have found them analogous in the sense that if you approach them with a similar plan of action as you would an actual calculator, you’ll get similar results.

        Difference is of course that math is not fuzzy and open to nuance, and 1+1 will always equal 2.

        But language isn’t math, and it’s trickier to get a sense for what the “equations” that will yield results are, so it’s easy to disparage the technology as a concept given (as you rightly point out) the boiling of the planet.

        Work has been insisting we tool with an LLM and they’re checking but thankfully my role doesn’t require relying on any facts the machine spits out.

        Which is another part of why the technology is so reviled/misunderstood; the part of the LLM that determines the next word isn’t and can’t judge the veracity of it’s output. Any landing on factual info is either a coincidence or the fact that you as the user knew to “coach” things in such a way as to arrive as the most likely output which the user already knew is correct.

        Because of the uncertainty, it is simply unwise to take any LLM’s output as factual, as any fact checking capacity isn’t innate but other operations being done on the output, if any.

        Then there’s all the other reasons to hate the things like who makes them, how they’re made, how they’re wielded, etc, and I frankly can’t blame anyone for vehemently hating LLM’s as a concept.

        But it’s disingenous to think the tech is wholly incapable of anything of merit to anyone and only idiots out there are using it (even if that may still be often the case).

        Or put another way: A sailor hidden within the ship’s hold will still drown and die alongside the ones that don’t resist the siren call above and pilot straight for the rocks.

        Cool stuff, deployed in maximally foolish ways. I think I rambled a bit there, but hopefully a bit of my point made it across.