• jj4211@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 day ago

    I saw one article actually going all in on how incredible GPT-5 was.

    https://www.oneusefulthing.org/p/gpt-5-it-just-does-stuff

    The thing is, the biggest piece that really made the author excited was the “startup idea” and it proceeded to generate a mountain of business-speak that says nothing. He proceeds to proclaim a whole team of MBAs would take hours to produce something so magnificent. This pretty much made him just lose it, and I guess that is exactly the sort of content idiot executives slurp up.

  • ExLisper@lemmy.curiana.net
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 days ago

    One thing I didn’t expect in AI evolution was all the “safety” features. Basically since people are to stupid to use LLMs and not fall in love or poison themselves all models have more and more guardrails making it way less useful than normal search. I think it was fairly clear from the beginning that LLMs will always be bullshit machines but I didn’t think they will be less and less useful bullshit machines.

  • some_guy@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 days ago

    If you believe that Sam Altman is full of shit and the whole AI hype machine is a bubble (regardless of any real-world uses that do exist) built on lies about where this specific form of the technology can actually go, congratulations. You might enjoy listening to or reading the work of Ed Zitron. He has a podcast and a newsletter and he’s been pointing this out for over a year, among other topics.

  • miridius@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 days ago

    Nothing to see here, just another “token based LLMs can’t count letters in words” post

    • shane@feddit.nl
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 days ago

      That’s what I thought but there’s slightly more than that.

      The writer tried to trick ChatGPT 5, saying Vermont has no R in it. ChatGPT did say “wait, it does”. But then when pushed it said, “oh right there is no R in Vermont”.

      I mean… the inability to know what it knows or not is a real problem for most use cases…

      • jj4211@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 days ago

        Yeah, the fact you can “gaslight” a chat is just as much of a symptom of a difficulty as the usual mistakes. It shows that it doesn’t deal with facts, but structurally sound content, which is correlated with facts, especially when the prompt has context/rag stuffing the prompt using more traditional approaches that actually will tend to get more factual stuff crammed in.

        To all the people white knighting for the LLM, for the thousandth time, we know that it is useful, but it’s usefulness is only tenuously connected to the marketing reality. Making the mistake in counting letters is less important than the fact that it “acts” like it can when it can’t.

  • NotASharkInAManSuit@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 days ago

    That’s because they’re not trying to make AI, they are just programming LLMs to be bullshit confirmation bias machines. They don’t want to create intelligence, they want to create unregulated revenue streams.

  • DominusOfMegadeus@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 days ago

    I tried ChatGPT-5 for the first time this morning. I asked it to help me create an RSS list for some world news articles. For context I have never used RSS before. That was 90 minutes of my life I’ll never get back. Also, you can no longer choose any other models except for 5 and the 5-thinking model. Access to 4o is gone.

      • DominusOfMegadeus@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 days ago

        It was consistently wrong about how to go about setting up a group feed on every single website it suggested. And I started out trying to work on my iPad, and it kept telling me where to find things, and it was wrong, and how to do something, and it was wrong, and it kept telling me supposed desktop instructions, even after I told it I was on an iPad, (an iPad is somehow different it claimed). So I went to the desktop and it was exactly the same as the iPad, as o knew it would be, meaning its instructions were just wrong. When I called it out and I asked it on its final recommendation to be absolutely sure that the process it was going to tell me was correct and up-to-date, and to read the website first before it gave me the information; it lied to me and said it was and that it would check first, and then didn’t do what it said it was going to do. Plus about 70% of RSS links it gave me were bad. It was just a really frustrating experience. Essentially going from version 4o to version 5 was a step backwards in AI evolution from what I can tell things,

  • Mereo@lemmy.ca
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 days ago

    I ran the tests with the thinking model. It got them right. For these kind of tasks, choosing the thinking model is key,

    • very_well_lost@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      the thinking model

      Ugh… can we all just stop for a moment to acknowledge how obnoxious this branding is? They’ve already corrupted the term “AI” to the point of being completely meaningless, are they going to remove all meaning from the word “thinking” now too?

      • Lemminary@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 days ago

        They’ve already corrupted the term “AI” to the point of being completely meaningless

        Did they? Afaik, LLMs are an application of AI that falls under natural language processing. It’s like calling a rhombus a geometric shape because that’s what it is. And this usage goes back decades to, for example, A* pathfinding algorithms and hard-coded decision trees for NPCs.

        • DefederateLemmyMl@feddit.nl
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 days ago

          I think the problem stems from how LLMs are marketed to, and perceived by the public. They are not marketed as: this is a specific application of this-or-that AI or ML technology. They are marketed as “WE HAVE AI NOW!”, and the general public who is not familiar with AI/ML technologies equates this to AGI, because that’s what they know from the movies. The promotional imagery that some of these companies put out, with humanoid robots that look like they came straight out of Ex Machina doesn’t help either.

          And sure enough, upon first contact, an LLM looks like a duck and quacks like a duck … so people assume it is a duck, but they don’t realize that it’s a cardboard model of a duck with a taperecorder inside that plays back quacking sounds.

  • A_norny_mousse@feddit.org
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    3 days ago

    Sam Altman Is Still a Bullshit Machine

    Just like all billionaire tech bros: totally removed from reality, living in their own bubbles of unimaginable wealth and loneliness. They convinced themselves that the constant grift is real and call it a utopia… please, someone take their dangerous toys (and most of their wealth) away. And since the government certainly isn’t going to step in

  • brucethemoose@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    3 days ago

    ChatGPT Is Still a Bullshit Machine

    Just like Sam Altman.


    I will add that they aren’t even tackling basic issues like the randomness of sampling; all OpenAI models still need a highish temperature. It’s like the poster child of corporate scaling vs actually innovating.