• DeathToUS@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    Just use DeepAI never bothered with ChatGPT because its so primitive same goes for Gemini.

  • brsrklf@jlai.lu
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    Somehow the chatGPT US map generated a clean, correctly spelled French word for “Mondays” instead of Illinois.

    Sorry, Illinois, Garfield hates your guts.

  • ExLisper@lemmy.curiana.net
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    One thing I didn’t expect in AI evolution was all the “safety” features. Basically since people are to stupid to use LLMs and not fall in love or poison themselves all models have more and more guardrails making it way less useful than normal search. I think it was fairly clear from the beginning that LLMs will always be bullshit machines but I didn’t think they will be less and less useful bullshit machines.

  • some_guy@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    If you believe that Sam Altman is full of shit and the whole AI hype machine is a bubble (regardless of any real-world uses that do exist) built on lies about where this specific form of the technology can actually go, congratulations. You might enjoy listening to or reading the work of Ed Zitron. He has a podcast and a newsletter and he’s been pointing this out for over a year, among other topics.

  • A_norny_mousse@feddit.org
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    2 months ago

    Sam Altman Is Still a Bullshit Machine

    Just like all billionaire tech bros: totally removed from reality, living in their own bubbles of unimaginable wealth and loneliness. They convinced themselves that the constant grift is real and call it a utopia… please, someone take their dangerous toys (and most of their wealth) away. And since the government certainly isn’t going to step in

  • miridius@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    Nothing to see here, just another “token based LLMs can’t count letters in words” post

    • shane@feddit.nl
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      That’s what I thought but there’s slightly more than that.

      The writer tried to trick ChatGPT 5, saying Vermont has no R in it. ChatGPT did say “wait, it does”. But then when pushed it said, “oh right there is no R in Vermont”.

      I mean… the inability to know what it knows or not is a real problem for most use cases…

      • jj4211@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        Yeah, the fact you can “gaslight” a chat is just as much of a symptom of a difficulty as the usual mistakes. It shows that it doesn’t deal with facts, but structurally sound content, which is correlated with facts, especially when the prompt has context/rag stuffing the prompt using more traditional approaches that actually will tend to get more factual stuff crammed in.

        To all the people white knighting for the LLM, for the thousandth time, we know that it is useful, but it’s usefulness is only tenuously connected to the marketing reality. Making the mistake in counting letters is less important than the fact that it “acts” like it can when it can’t.

  • Mereo@lemmy.ca
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    I ran the tests with the thinking model. It got them right. For these kind of tasks, choosing the thinking model is key,

    • very_well_lost@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      the thinking model

      Ugh… can we all just stop for a moment to acknowledge how obnoxious this branding is? They’ve already corrupted the term “AI” to the point of being completely meaningless, are they going to remove all meaning from the word “thinking” now too?

      • Lemminary@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        They’ve already corrupted the term “AI” to the point of being completely meaningless

        Did they? Afaik, LLMs are an application of AI that falls under natural language processing. It’s like calling a rhombus a geometric shape because that’s what it is. And this usage goes back decades to, for example, A* pathfinding algorithms and hard-coded decision trees for NPCs.

        • DefederateLemmyMl@feddit.nl
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          I think the problem stems from how LLMs are marketed to, and perceived by the public. They are not marketed as: this is a specific application of this-or-that AI or ML technology. They are marketed as “WE HAVE AI NOW!”, and the general public who is not familiar with AI/ML technologies equates this to AGI, because that’s what they know from the movies. The promotional imagery that some of these companies put out, with humanoid robots that look like they came straight out of Ex Machina doesn’t help either.

          And sure enough, upon first contact, an LLM looks like a duck and quacks like a duck … so people assume it is a duck, but they don’t realize that it’s a cardboard model of a duck with a taperecorder inside that plays back quacking sounds.

  • BallShapedMan@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    A lady I can’t stand who is condescending to everyone like the worst version of a grade school teacher resigned recently so she can market her GPT “models” full time.

    Last week she used “Salem” her GPT prompt BS to read through a detailed document my department put together to see what was missing. She shared her screen with obvious gpt slop and started pointing out all the things we didn’t answer. The lady on my team who put the very specific and detailed document together just started reading the answers right from the document that showed it was clearly and precisely answered. The lady I can’t stand stopped sharing her screen and quit talking.

    The moral of the story, gpt did me a solid by convincing this person who’s clearly never heard of the Dunning Kruger effect that she needs to quit her well paying job and stop being a pain in my ass.

    Thank you GPT!! Best thing is ever done for me.

  • NotASharkInAManSuit@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    That’s because they’re not trying to make AI, they are just programming LLMs to be bullshit confirmation bias machines. They don’t want to create intelligence, they want to create unregulated revenue streams.

  • DominusOfMegadeus@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    I tried ChatGPT-5 for the first time this morning. I asked it to help me create an RSS list for some world news articles. For context I have never used RSS before. That was 90 minutes of my life I’ll never get back. Also, you can no longer choose any other models except for 5 and the 5-thinking model. Access to 4o is gone.

      • DominusOfMegadeus@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        It was consistently wrong about how to go about setting up a group feed on every single website it suggested. And I started out trying to work on my iPad, and it kept telling me where to find things, and it was wrong, and how to do something, and it was wrong, and it kept telling me supposed desktop instructions, even after I told it I was on an iPad, (an iPad is somehow different it claimed). So I went to the desktop and it was exactly the same as the iPad, as o knew it would be, meaning its instructions were just wrong. When I called it out and I asked it on its final recommendation to be absolutely sure that the process it was going to tell me was correct and up-to-date, and to read the website first before it gave me the information; it lied to me and said it was and that it would check first, and then didn’t do what it said it was going to do. Plus about 70% of RSS links it gave me were bad. It was just a really frustrating experience. Essentially going from version 4o to version 5 was a step backwards in AI evolution from what I can tell things,