I’m usually the one saying “AI is already as good as it’s gonna get, for a long while.”

This article, in contrast, is quotes from folks making the next AI generation - saying the same.

  • ChicoSuave@lemmy.world
    link
    fedilink
    English
    arrow-up
    44
    arrow-down
    2
    ·
    10 months ago

    I understand folks don’t like AI but this “article” is like a reddit post with lots of links to subjects which are vague and need the link text to tell us what is important, instead of relying on the actual article.

    • 11111one11111@lemmy.world
      link
      fedilink
      English
      arrow-up
      25
      arrow-down
      2
      ·
      edit-2
      10 months ago

      What the fuck you aren’t kidding. I have comment replies to trolls that are longer than that article. The over the top citations also makes me think this was entirely written by an actual AI bot that was lrompted to supply x amoint of sources in their article. Lol

    • Korne127@lemmy.world
      link
      fedilink
      English
      arrow-up
      15
      arrow-down
      2
      ·
      10 months ago

      LLMs are one version of AI. It’s just one tiny part of AIs that are used every day, from chess bots to voice transcription, but they also are AI.

      • buddascrayon@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        1
        ·
        edit-2
        10 months ago

        I would replace the word version with aspect. LLMs are merely one part of the puzzle that would be AI. Essentially what’s been constructed is the mouth and the part of the brain that can form words but without any of the reasoning or intelligence behind what the mouth says.

        The same goes for the art AIs. They can paint pictures based on input but they can’t reason how those pictures should look. Which is why it requires so much tweaking to get them to output something that doesn’t look like it came out of a Lovecraft novel.

  • Greg Clarke@lemmy.ca
    link
    fedilink
    English
    arrow-up
    36
    arrow-down
    11
    ·
    10 months ago

    OpenAI, Google, Anthropic admit they can’t scale up their chatbots any further

    Lol, no they didn’t. The quotes this articles are using are talking about LLMs not chatbots. This is yet another stupid article from someone who doesn’t understand the technology. There is a lot of legitimate criticism for the way this technology is being implemented but FFS get the basics right at least.

    • MajorHavoc@programming.devOP
      link
      fedilink
      English
      arrow-up
      18
      arrow-down
      5
      ·
      10 months ago

      Are you asserting that chatbots are so fundamentally different from LLMs that “oh shit we can’t just throw more CPU and data at this anymore” doesn’t apply to roughly the same degree?

        • Greg Clarke@lemmy.ca
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          5
          ·
          10 months ago

          People that don’t understand those terms are using them interchangeably

          • Buffalox@lemmy.world
            link
            fedilink
            English
            arrow-up
            6
            ·
            10 months ago

            LLM is the technology, Chatbot is an implementation of it. So yes a Chatbot as it’s talked about here is an LLM. Although obviously chatbots don’t have to be LLM, those that are not are irrelevant.

            • Greg Clarke@lemmy.ca
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              7
              ·
              10 months ago

              No, a chat bot as it’s talked about here is not an LLM. This article is discussing limitations of LLM training data and inferring that chat bots can not scale as a result. There are many techniques that can be used to continue to improve chat bots.

              • Buffalox@lemmy.world
                link
                fedilink
                English
                arrow-up
                4
                arrow-down
                1
                ·
                10 months ago

                The chatbot is a front end to an LLM, you are being needlessly pedantic. What the chatbot serves you, is the result of LLM queries.

                • Greg Clarke@lemmy.ca
                  link
                  fedilink
                  English
                  arrow-up
                  3
                  arrow-down
                  5
                  ·
                  10 months ago

                  That may have been true for the early LLM chatbots but not anymore. ChatGPT for instance, now writes code to answer logical questions. The o1 models have background token usage because each response is actually the result of multiple background LLM responses.

      • Greg Clarke@lemmy.ca
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        8
        ·
        10 months ago

        Yes of course I’m asserting that. While the performance of LLMs may be plateauing, the cost, context window, and efficiency is still getting much better. When you chat with a modern chat bot it’s not just sending your input to an LLM like the first public version of ChatGPT. Nowadays a single chat bot response may require many LLM requests along with other techniques to mitigate the deficiencies of LLMs. Just ask the free version of ChatGPT a question that requires some calculation and you’ll have a better understanding of what’s going on and the direction of the industry.

        • MajorHavoc@programming.devOP
          link
          fedilink
          English
          arrow-up
          4
          ·
          10 months ago

          I think you’re agreeing, just in a rude and condescending way.

          There’s a lot of ways left to improve, but they’re not as simple as just throwing more data and CPU at the problem, anymore.

          • Greg Clarke@lemmy.ca
            link
            fedilink
            English
            arrow-up
            3
            arrow-down
            3
            ·
            edit-2
            10 months ago

            I’m sorry if I’m coming across as condescending, that’s not my intent. It’s never been “as simple as just throwing more data and CPU at the problem”. There were algorithmic challenges for every LLM evolution. There are still lots of potential improvements using the existing training data. But even if there wasn’t, we’ll still see loads of improvements in chat bots because of other techniques.

            Edit: typo

  • Ragdoll X@lemmy.worldBanned
    link
    fedilink
    English
    arrow-up
    5
    ·
    edit-2
    10 months ago

    It’s a known problem - though of course, because these companies are trying to push AI into everything and oversell it to build hype and please investors, they usually try to avoid recognizing its limitations.

    Frankly I think that now they should focus on making these models smaller and more efficient instead of just throwing more compute at the wall, and actually train them to completion so they’ll generalize properly and be more useful.

  • Tux@lemmy.world
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    3
    ·
    10 months ago

    Looks, like AI buble is slowly coming to end just like what happned to crypto and NFT buble.

      • MajorHavoc@programming.devOP
        link
        fedilink
        English
        arrow-up
        1
        ·
        10 months ago

        The bubble was when we were being sold block chain as the solution to every problem. I feel like that bubble ended in 2019 or 2020.

        Things that actually benefitted from block chain are still around, of course.

        Unrelated side rant: I’m pissed about pogs going away, though. Pogs were fun. I should still be able to buy pogs.

  • nialv7@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    2
    ·
    10 months ago

    They might be right but I read some of the linked articles on this blog (?), the authors just come off as not really knowing much about current AI technologies, and at the same time very very arrogant.

    • raspberriesareyummy@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      10 months ago

      The article talks about LLM developers / operators. Not sure how you got from that to “current AI technologies” - a completely unrelated topic.