• NeilBrü
    link
    fedilink
    English
    44
    edit-2
    4 days ago

    An LLM is a poor computational/predictive paradigm for playing chess.

      • NeilBrü
        link
        fedilink
        English
        24 days ago

        I’m impressed, if that’s true! In general, an LLM’s training cost vs. an LSTM, RNN, or some other more appropriate DNN algorithm suitable for the ruleset is laughably high.

        • @Takapapatapaka@lemmy.world
          link
          fedilink
          English
          24 days ago

          Oh yes, cost of training are ofc a great loss here, it’s not optimized at all, and it’s stuck at an average level.

          Interestingly, i believe some people did research on it and found some parameters in the model that seemed to represent the state of the chess board (as in, they seem to reflect the current state of the board, and when artificially modified, the model takes modification into account in its playing). It was used by a french youtuber to show how LLMs can somehow have a kinda representation of the world. I can try to get the sources back if you’re interested.

          • NeilBrü
            link
            fedilink
            English
            1
            edit-2
            4 days ago

            Absolutely interested. Thank you for your time to share that.

            My career path in neural networks began as a researcher for cancerous tissue object detection in medical diagnostic imaging. Now it is switched to generative models for CAD (architecture, product design, game assets, etc.). I don’t really mess about with fine-tuning LLMs.

            However, I do self-host my own LLMs as code assistants. Thus, I’m only tangentially involved with the current LLM craze.

            But it does interest me, nonetheless!

            • @Takapapatapaka@lemmy.world
              link
              fedilink
              English
              13 days ago

              Here is the main blog post that i remembered : it has a follow up, a more scientific version, and uses two other articles as a basis, so you might want to dig around what they mention in the introduction.

              It is indeed a quite technical discovery, and it still lacks complete and wider analysis, but it is very interesting for the fact that it kinda invalidates the common gut feeling that llms are pure lucky random.

    • @Bleys@lemmy.world
      link
      fedilink
      English
      44 days ago

      The underlying neural network tech is the same as what the best chess AIs (AlphaZero, Leela) use. The problem is, as you said, that ChatGPT is designed specifically as an LLM so it’s been optimized strictly to write semi-coherent text first, and then any problem solving beyond that is ancillary. Which should say a lot about how inconsistent ChatGPT is at solving problems, given that it’s not actually optimized for any specific use cases.

      • NeilBrü
        link
        fedilink
        English
        3
        edit-2
        4 days ago

        Yes, I agree wholeheartedly with your clarification.

        My career path, as I stated in a different comment in regards to neural networks, is focused on generative DNNs for CAD applications and parametric 3D modeling. Before that, I began as a researcher in cancerous tissue classification and object detection in medical diagnostic imaging.

        Thus, large language models are well out of my area of expertise in terms of the architecture of their models.

        However, fundamentally it boils down to the fact that the specific large language model used was designed to predict text and not necessarily solve problems/play games to “win”/“survive”.

        (I admit that I’m just parroting what you stated and maybe rehashing what I stated even before that, but I like repeating and refining in simple terms to practice explaining to laymen and, dare I say, clients. It helps me feel as if I don’t come off too pompously when talking about this subject to others; forgive my tedium.)

  • @FMT99@lemmy.world
    link
    fedilink
    English
    1355 days ago

    Did the author thinks ChatGPT is in fact an AGI? It’s a chatbot. Why would it be good at chess? It’s like saying an Atari 2600 running a dedicated chess program can beat Google Maps at chess.

    • snooggums
      link
      fedilink
      English
      1265 days ago

      AI including ChatGPT is being marketed as super awesome at everything, which is why that and similar AI is being forced into absolutely everything and being sold as a replacement for people.

      Something marketed as AGI should be treated as AGI when proving it isn’t AGI.

      • NoiseColor
        cake
        link
        fedilink
        English
        -175 days ago

        I don’t think ai is being marketed as awesome at everything. It’s got obvious flaws. Right now its not good for stuff like chess, probably not even tic tac toe. It’s a language model, its hard for it to calculate the playing field. But ai is in development, it might not need much to start playing chess.

        • @vinnymac@lemmy.world
          link
          fedilink
          English
          175 days ago

          What the tech is being marketed as and what it’s capable of are not the same, and likely never will be. In fact all things are very rarely marketed how they truly behave, intentionally.

          Everyone is still trying to figure out what these Large Reasoning Models and Large Language Models are even capable of; Apple, one of the largest companies in the world just released a white paper this past week describing the “illusion of reasoning”. If it takes a scientific paper to understand what these models are and are not capable of, I assure you they’ll be selling snake oil for years after we fully understand every nuance of their capabilities.

          TL;DR Rich folks want them to be everything, so they’ll be sold as capable of everything until we repeatedly refute they are able to do so.

          • NoiseColor
            cake
            link
            fedilink
            English
            -55 days ago

            I think in many cases people intentionally or unintentionally disregard the time component here. Ai is in development. I think what is being marketed here, just like in the stock market, is a piece of the future. I don’t expect the models I use to be perfect and not make mistakes, so I use them accordingly. They are useful for what I use them for and I wouldn’t use them for chess. I don’t expect that laundry detergent to be just as perfect in the commercial either.

        • @BassTurd@lemmy.world
          link
          fedilink
          English
          115 days ago

          Marketing does not mean functionality. AI is absolutely being sold to the public and enterprises as something that can solve everything. Obviously it can’t, but it’s being sold that way. I would bet the average person would be surprised by this headline solely on what they’ve heard about the capabilities of AI.

          • NoiseColor
            cake
            link
            fedilink
            English
            -105 days ago

            I don’t think anyone is so stupid to believe current ai can solve everything.

            And honestly, I didn’t see any marketing material that would claim that.

            • @BassTurd@lemmy.world
              link
              fedilink
              English
              95 days ago

              You are both completely over estimating the intelligence level of “anyone” and not living in the same AI marketed universe as the rest of us. People are stupid. Really stupid.

              • NoiseColor
                cake
                link
                fedilink
                English
                15 days ago

                I don’t understand why this is so important, marketing is all about exaggerating, why expect something different here.

                • @BassTurd@lemmy.world
                  link
                  fedilink
                  English
                  25 days ago

                  It’s not important. You said AI isn’t being marketed to be able to do everything. I said yes it is. That’s it.

    • @suburban_hillbilly@lemmy.ml
      link
      fedilink
      English
      205 days ago

      Most people do. It’s just called AI in the media everywhere and marketing works. I think online folks forget that something as simple as getting a Lemmy account by yourself puts you into the top quintile of tech literacy.

    • @Broken@lemmy.ml
      link
      fedilink
      English
      45 days ago

      I agree with your general statement, but in theory since all ChatGPT does is regurgitate information back and a lot of chess is memorization of historical games and types, it might actually perform well. No, it can’t think, but it can remember everything so at some point that might tip the results in it’s favor.

      • @Eagle0110@lemmy.world
        link
        fedilink
        English
        3
        edit-2
        5 days ago

        Regurgitating an impression of, not regurgitating verbatim, that’s the problem here.

        Chess is 100% deterministic, so it falls flat.

        • Raltoid
          link
          fedilink
          English
          4
          edit-2
          5 days ago

          I’m guessing it’s not even hard to get it to “confidently” violate the rules.

      • @FMT99@lemmy.world
        link
        fedilink
        English
        14 days ago

        I mean it may be possible but the complexity would be so many orders of magnitude greater. It’d be like learning chess by just memorizing all the moves great players made but without any context or understanding of the underlying strategy.

    • @iAvicenna@lemmy.world
      link
      fedilink
      English
      65 days ago

      well so much hype has been generated around chatgpt being close to AGI that now it makes sense to ask questions like “can chatgpt prove the Riemann hypothesis”

    • @x00z@lemmy.world
      link
      fedilink
      English
      35 days ago

      In all fairness. Machine learning in chess engines is actually pretty strong.

      AlphaZero was developed by the artificial intelligence and research company DeepMind, which was acquired by Google. It is a computer program that reached a virtually unthinkable level of play using only reinforcement learning and self-play in order to train its neural networks. In other words, it was only given the rules of the game and then played against itself many millions of times (44 million games in the first nine hours, according to DeepMind).

      https://www.chess.com/terms/alphazero-chess-engine

      • @jeeva@lemmy.world
        link
        fedilink
        English
        14 days ago

        Sure, but machine learning like that is very different to how LLMs are trained and their output.

      • @FMT99@lemmy.world
        link
        fedilink
        English
        14 days ago

        Oh absolutely you can apply machine learning to game strategy. But you can’t expect a generalized chatbot to do well at strategic decision making for a specific game.

    • @saltesc@lemmy.world
      link
      fedilink
      English
      15 days ago

      I like referring to LLMs as VI (Virtual Intelligence from Mass Effect) since they merely give the impression of intelligence but are little more than search engines. In the end all one is doing is displaying expected results based on a popularity algorithm. However they do this inconsistently due to bad data in and limited caching.

  • @finitebanjo@lemmy.world
    link
    fedilink
    English
    64 days ago

    All these comments asking “why don’t they just have chatgpt go and look up the correct answer”.

    That’s not how it works, you buffoons, it trains off of datasets long before it releases. It doesn’t think. It doesn’t learn after release, it won’t remember things you try to teach it.

    Really lowering my faith in humanity when even the AI skeptics don’t understand that it generates statistical representations of an answer based on answers given in the past.

  • @floofloof@lemmy.ca
    link
    fedilink
    English
    19
    edit-2
    5 days ago

    I suppose it’s an interesting experiment, but it’s not that surprising that a word prediction machine can’t play chess.

  • @Furbag@lemmy.world
    cake
    link
    fedilink
    English
    135 days ago

    Can ChatGPT actually play chess now? Last I checked, it couldn’t remember more than 5 moves of history so it wouldn’t be able to see the true board state and would make illegal moves, take it’s own pieces, materialize pieces out of thin air, etc.

    • @skisnow@lemmy.ca
      link
      fedilink
      English
      3
      edit-2
      4 days ago

      It can’t, but that didn’t stop a bunch of gushing articles a while back about how it had an ELO of 2400 and other such nonsense. Turns out you could get it to have an ELO of 2400 under a very very specific set of circumstances that, include correcting it every time it hallucinated pieces or attempted to make illegal moves.

  • OBJECTION!
    link
    fedilink
    English
    315 days ago

    Tbf, the article should probably mention the fact that machine learning programs designed to play chess blow everything else out of the water.

    • @andallthat@lemmy.world
      link
      fedilink
      English
      7
      edit-2
      4 days ago

      Machine learning has existed for many years, now. The issue is with these funding-hungry new companies taking their LLMs, repackaging them as “AI” and attributing every ML win ever to “AI”.

      ML programs designed and trained specifically to identify tumors in medical imaging have become good diagnostic tools. But if you read in news that “AI helps cure cancer”, it makes it sound like it was a lone researcher who spent a few minutes engineering the right prompt for Copilot.

      Yes a specifically-designed and finely tuned ML program can now beat the best human chess player, but calling it “AI” and bundling it together with the latest Gemini or Claude iteration’s “reasoning capabilities” is intentionally misleading. That’s why articles like this one are needed. ML is a useful tool but far from the “super-human general intelligence” that is meant to replace half of human workers by the power of wishful prompting

  • @arc99@lemmy.world
    link
    fedilink
    English
    134 days ago

    Hardly surprising. Llms aren’t -thinking- they’re just shitting out the next token for any given input of tokens.

  • @jsomae@lemmy.ml
    link
    fedilink
    English
    94 days ago

    Using an LLM as a chess engine is like using a power tool as a table leg. Pretty funny honestly, but it’s obviously not going to be good at it, at least not without scaffolding.

    • @kent_eh@lemmy.ca
      link
      fedilink
      English
      23 days ago

      is like using a power tool as a table leg.

      Then again, our corporate lords and masters are trying to replace all manner of skilled workers with those same LLM “AI” tools.

      And clearly that will backfire on them and they’ll eventually scramble to find people with the needed skills, but in the meantime tons of people will have lost their source of income.

      • @jsomae@lemmy.ml
        link
        fedilink
        English
        1
        edit-2
        3 days ago

        If you believe LLMs are not good at anything then there should be relatively little to worry about in the long-term, but I am more concerned.

        It’s not obvious to me that it will backfire for them, because I believe LLMs are good at some things (that is, when they are used correctly, for the correct tasks). Currently they’re being applied to far more use cases than they are likely to be good at – either because they’re overhyped or our corporate lords and masters are just experimenting to find out what they’re good at and what not. Some of these cases will be like chess, but others will be like code*.

        (* not saying LLMs are good at code in general, but for some coding applications I believe they are vastly more efficient than humans, even if a human expert can currently write higher-quality less-buggy code.)

        • @kent_eh@lemmy.ca
          link
          fedilink
          English
          12 days ago

          I believe LLMs are good at some things

          The problem is that they’re being used for all the things, including a large number of tasks that thwy are not well suited to.

          • @jsomae@lemmy.ml
            link
            fedilink
            English
            22 days ago

            yeah, we agree on this point. In the short term it’s a disaster. In the long-term, assuming AI’s capabilities don’t continue to improve at the rate they have been, our corporate overlords will only replace people for whom it’s actually worth it to them to replace with AI.

  • I Cast Fist
    link
    fedilink
    English
    24 days ago

    So, it fares as well as the average schmuck, proving it is human

    /s

  • @vane@lemmy.world
    link
    fedilink
    English
    25 days ago

    It’s not that hard to beat dumb 6 year old who’s only purpose is mine your privacy to sell you ads or product place some shit for you in future.