• Steve Dice@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    0
    ·
    4 days ago

    2025 Mazda MX-5 Miata ‘got absolutely wrecked’ by Inflatable Boat in beginner’s boat racing match — Mazda’s newest model bamboozled by 1930s technology.

  • FourWaveforms@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    4 days ago

    If you don’t play chess, the Atari is probably going to beat you as well.

    LLMs are only good at things to the extent that they have been well-trained in the relevant areas. Not just learning to predict text string sequences, but reinforcement learning after that, where a human or some other agent says “this answer is better than that one” enough times in enough of the right contexts. It mimics the way humans learn, which is through repeated and diverse exposure.

    If they set up a system to train it against some chess program, or (much simpler) simply gave it a tool call, it would do much better. Tool calling already exists and would be by far the easiest way.

    It could also be instructed to write a chess solver program and then run it, at which point it would be on par with the Atari, but it wouldn’t compete well with a serious chess solver.

  • jsomae@lemmy.ml
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 days ago

    Using an LLM as a chess engine is like using a power tool as a table leg. Pretty funny honestly, but it’s obviously not going to be good at it, at least not without scaffolding.

    • kent_eh@lemmy.ca
      link
      fedilink
      English
      arrow-up
      0
      ·
      4 days ago

      is like using a power tool as a table leg.

      Then again, our corporate lords and masters are trying to replace all manner of skilled workers with those same LLM “AI” tools.

      And clearly that will backfire on them and they’ll eventually scramble to find people with the needed skills, but in the meantime tons of people will have lost their source of income.

      • jsomae@lemmy.ml
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        4 days ago

        If you believe LLMs are not good at anything then there should be relatively little to worry about in the long-term, but I am more concerned.

        It’s not obvious to me that it will backfire for them, because I believe LLMs are good at some things (that is, when they are used correctly, for the correct tasks). Currently they’re being applied to far more use cases than they are likely to be good at – either because they’re overhyped or our corporate lords and masters are just experimenting to find out what they’re good at and what not. Some of these cases will be like chess, but others will be like code*.

        (* not saying LLMs are good at code in general, but for some coding applications I believe they are vastly more efficient than humans, even if a human expert can currently write higher-quality less-buggy code.)

        • kent_eh@lemmy.ca
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 days ago

          I believe LLMs are good at some things

          The problem is that they’re being used for all the things, including a large number of tasks that thwy are not well suited to.

          • jsomae@lemmy.ml
            link
            fedilink
            English
            arrow-up
            0
            ·
            3 days ago

            yeah, we agree on this point. In the short term it’s a disaster. In the long-term, assuming AI’s capabilities don’t continue to improve at the rate they have been, our corporate overlords will only replace people for whom it’s actually worth it to them to replace with AI.

    • Bleys@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 days ago

      The underlying neural network tech is the same as what the best chess AIs (AlphaZero, Leela) use. The problem is, as you said, that ChatGPT is designed specifically as an LLM so it’s been optimized strictly to write semi-coherent text first, and then any problem solving beyond that is ancillary. Which should say a lot about how inconsistent ChatGPT is at solving problems, given that it’s not actually optimized for any specific use cases.

      • NeilBrü@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        5 days ago

        Yes, I agree wholeheartedly with your clarification.

        My career path, as I stated in a different comment, In regards to neural networks is focused on generative DNNs for CAD applications and parametric 3D modeling. Before that, I began as a researcher in cancerous tissue classification and object detection in medical diagnostic imaging.

        Thus, large language models are well out of my area of expertise in terms of the architecture of their models.

        However, fundamentally it boils down to the fact that the specific large language model used was designed to predict text and not necessarily solve problems/play games to “win”/“survive”.

        (I admit that I’m just parroting what you stated and maybe rehashing what I stated even before that, but I like repeating and refining in simple terms to practice explaining to laymen and, dare I say, clients. It helps me feel as if I don’t come off too pompously when talking about this subject to others; forgive my tedium.)

      • NeilBrü@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 days ago

        I’m impressed, if that’s true! In general, an LLM’s training cost vs. an LSTM, RNN, or some other more appropriate DNN algorithm suitable for the ruleset is laughably high.

        • Takapapatapaka@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          5 days ago

          Oh yes, cost of training are ofc a great loss here, it’s not optimized at all, and it’s stuck at an average level.

          Interestingly, i believe some people did research on it and found some parameters in the model that seemed to represent the state of the chess board (as in, they seem to reflect the current state of the board, and when artificially modified, the model takes modification into account in its playing). It was used by a french youtuber to show how LLMs can somehow have a kinda representation of the world. I can try to get the sources back if you’re interested.

          • NeilBrü@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            5 days ago

            Absolutely interested. Thank you for your time to share that.

            My career path in neural networks began as a researcher for cancerous tissue object detection in medical diagnostic imaging. Now it is switched to generative models for CAD (architecture, product design, game assets, etc.). I don’t really mess about with fine-tuning LLMs.

            However, I do self-host my own LLMs as code assistants. Thus, I’m only tangentially involved with the current LLM craze.

            But it does interest me, nonetheless!

  • finitebanjo@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 days ago

    All these comments asking “why don’t they just have chatgpt go and look up the correct answer”.

    That’s not how it works, you buffoons, it trains off of datasets long before it releases. It doesn’t think. It doesn’t learn after release, it won’t remember things you try to teach it.

    Really lowering my faith in humanity when even the AI skeptics don’t understand that it generates statistical representations of an answer based on answers given in the past.

  • nednobbins@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 days ago

    Sometimes it seems like most of these AI articles are written by AIs with bad prompts.

    Human journalists would hopefully do a little research. A quick search would reveal that researches have been publishing about this for over a year so there’s no need to sensationalize it. Perhaps the human journalist could have spent a little time talking about why LLMs are bad at chess and how researchers are approaching the problem.

    LLMs on the other hand, are very good at producing clickbait articles with low information content.

    • Lovable Sidekick@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      5 days ago

      In this case it’s not even bad prompts, it’s a problem domain ChatGPT wasn’t designed to be good at. It’s like saying modern medicine is clearly bullshit because a doctor loses a basketball game.

    • nova_ad_vitum@lemmy.ca
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 days ago

      Gotham chess has a video of making chatgpt play chess against stockfish. Spoiler: chatgpt does not do well. It plays okay for a few moves but then the moment it gets in trouble it straight up cheats. Telling it to follow the rules of chess doesn’t help.

      This sort of gets to the heart of LLM-based “AI”. That one example to me really shows that there’s no actual reasoning happening inside. It’s producing answers that statistically look like answers that might be given based on that input.

      For some things it even works. But calling this intelligence is dubious at best.

      • JacksonLamb@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 days ago

        ChatGPT versus Deepseek is hilarious. They both cheat like crazy and then one side jedi mind tricks the winner into losing.

      • interdimensionalmeme@lemmy.ml
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 days ago

        I think the biggest problem is it’s very low ability to “test time adaptability”. Even when combined with a reasonning model outputting into its context, the weights do not learn out of the immediate context.

        I think the solution might be to train a LoRa overlay on the fly against the weights and run inference with that AND the unmodified weights and then have an overseer model self evaluate and recompose the raw outputs.

        Like humans are way better at answering stuff when it’s a collaboration of more than one person. I suspect the same is true of LLMs.

      • Ultraviolet@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        5 days ago

        Because it doesn’t have any understanding of the rules of chess or even an internal model of the game state, it just has the text of chess games in its training data and can reproduce the notation, but nothing to prevent it from making illegal moves, trying to move or capture pieces that don’t exist, incorrectly declaring check/checkmate, or any number of nonsensical things.

  • Pamasich@kbin.earth
    link
    fedilink
    arrow-up
    0
    ·
    5 days ago

    Isn’t the Atari just a game console, not a chess engine?

    Like, Wikipedia doesn’t mention anything about the Atari 2600 having a built-in chess engine.

    If they were willing to run a chess game on the Atari 2600, why did they not apply the same to ChatGPT? There are custom GPTs which claim to use a stockfish API or play at a similar level.

    Like this, it’s just unfair. Both platforms are not designed to deal with the task by themselves, but one of them is given the necessary tooling, the other one isn’t. No matter what you think of ChatGPT, that’s not a fair comparison.

    • NutWrench@lemmy.ml
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 days ago

      The Atari 2600 is just hardware. The software came on plug-in cartridges. Video Chess was released for it in 1979.

    • jj4211@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 days ago

      GPTs which claim to use a stockfish API

      Then the actual chess isn’t LLM. If you are going stockfish, then the LLM doesn’t add anything, stockfish is doing everything.

      The whole point is the marketing rage is that LLMs can do all kinds of stuff, doubling down on this with the branding of some approaches as “reasoning” models, which are roughly “similar to ‘pre-reasoning’, but forcing use of more tokens on disposable intermediate generation steps”. With this facet of LLM marketing, the promise would be that the LLM can “reason” itself through a chess game without particular enablement. In practice, people trying to feed in gobs of chess data to an LLM end up with an LLM that doesn’t even comply to the rules of the game, let alone provide reasonable competitive responses to an oppone.

      • Pamasich@kbin.earth
        link
        fedilink
        arrow-up
        0
        ·
        5 days ago

        Then the actual chess isn’t LLM.

        And neither did the Atari 2600 win against ChatGPT. Whatever game they ran on it did.

        That’s my point here. The fact that neither Atari 2600 nor ChatGPT are capable of playing chess on their own. They can only do so if you provide them with the necessary tools. Which applies to both of them. Yet only one of them was given those tools here.

        • jj4211@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          5 days ago

          Fine, a chess engine that is capable of running with affordable even for the time 1970s electronics will best what marketing folks would have you think is an arbitrarily capable “reasoning” model running on top of the line 2025 hardware.

          You can split hairs about “well actually, the 2600 is hardware and a chess engine is the software” but everyone gets the point.

          As to assertions that no one should expect an LLM to be a chess engine, well tell that to the industry that is asserting the LLMs are now “reasoning” and provides a basis to replace most of the labor pool. We need stories like this to calibrate expectations in a way common people can understand…

  • Halosheep@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 days ago

    I swear every single article critical of current LLMs is like, “The square got BLASTED by the triangle shape when it completely FAILED to go through the triangle shaped hole.”

  • arc99@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 days ago

    Hardly surprising. Llms aren’t -thinking- they’re just shitting out the next token for any given input of tokens.