• TommySoda@lemmy.world
    link
    fedilink
    arrow-up
    30
    arrow-down
    2
    ·
    edit-2
    11 months ago

    I have nothing against the people that are working on AI and appreciate the work they do. However every time I see an article about a company using AI like this I just get the vibe that it’s a bunch of middle aged men trying desperately to make things like the “future” they saw when they were a kid. I’ve seen amazing implementations of AI in a lot of different ways but I’m so sick of dumb ideas like this because some guy that used to watch Star Trek as a kid wants to feel like they live in the future while piggybacking on someone else’s work. It’s like the painted tunnel in cartoons where it looks like a real tunnel but in reality it’s just a very convincing lie. And that’s all that it is. Complexity does not mean sophistication when it comes to AI and never has and to treat it as such is just a forceful way to make your ideas come true without putting in the real effort.

    Sorry, I had to get that out. Also I have nothing against Star Trek and I used to watch it as a kid because my parents watched it all the time.

    • Thorned_Rose@kbin.social
      link
      fedilink
      arrow-up
      3
      ·
      11 months ago

      some guy that used to watch Star Trek as a kid wants to feel like they live in the future while piggybacking on someone else’s work.

      I don’t think they care about their own nostalgia. I think they ant to use other people’s dreams to make a lot of money. I’m also sure some of them genuinely just ant to push the technological envelope just cause they can, ethics be damned. But ultimately, it’s just money.

      I would love nothing more than the utopian future Trek promised but greed is killing it.

    • Lvxferre@lemmy.ml
      link
      fedilink
      arrow-up
      4
      arrow-down
      2
      ·
      11 months ago

      Complexity does not mean sophistication when it comes to AI and never has and to treat it as such is just a forceful way to make your ideas come true without putting in the real effort.

      It’s a bit off-topic, but what I really want is a language model that assigns semantic values to the tokens, and handles those values instead of directly working with the tokens themselves. That would be probably far less complex than current state-of-art LLMs, but way more sophisticated, and require far less data for “training”.

      • njordomir@lemmy.world
        link
        fedilink
        arrow-up
        3
        ·
        11 months ago

        I’m not sure I understand. Do you mean hearing codewords triggering actions as opposed to trying to understand the users intent through language? Or is are there a few more layers to this whole thing than my moderate nerd cred will allow me to understand?

        • Lvxferre@lemmy.ml
          link
          fedilink
          arrow-up
          6
          ·
          11 months ago

          Not quite. I’m focusing on chatbots like Bard, ChatGPT and the likes, and their technology (LLM, or large language model).

          At the core those LLMs work like this: they pick words, split them into “tokens”, and then perform a few operations on those tokens, across multiple layers. But at the end of the day they still work with the words themselves, not with the meaning being encoded by those words.

          What I want is an LLM that assigns multiple meanings for those words, and performs the operations above on the meaning itself. In other words the LLM would actually understand you, not just chain words.