By 2030, AI will greatly outperform humans in some complex intellectual tasks. Discover how LLMs are doubling their capabilities every seven months.

    • Eager Eagle@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      That sounds like a coin flip, but 50% reliability can be really useful.

      If a model has 50% chance of completing a task that would cost me an hour - and I can easily check it was completed correctly - on average, I’m saving half of the time it would take to complete this.

      That said, exponentials don’t exist in the real world, we’re just seeing the middle of a sigmoid curve, which will soon yield diminishing returns.

      • Womble@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 months ago

        That said, exponentials don’t exist in the real world, we’re just seeing the middle of a sigmoid curve, which will soon yield diminishing returns.

        Yes, but the tricky thing is we have no idea when the seemingly exponential growth will flip over into the plateuing phase. We could be there already or it could be another 30 years.

        For comparison Moores law is almost certainly a sigmoid too, but weve been seeing exponential growth for 50 years now.

  • snooggums@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    This is like measuring the increasing speeds of cars in the early years and extrapolating that they would be supersonic by now by ignoring the exponential impact that air resistance has.

    • abruptly8951@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      Or like looking at the early days of semiconductors and extrapolating that CPU speed will double every 18 months …smh these people

      • snooggums@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        3 months ago

        Since CPU speeds are still doubling every 18 months you have a solid point!

        Or maybe not since you are probably referring to the doubling of transistors that was an observation which was accurate over a lengthy period of time in the context of when the observation was made. Nobody said that would continue indefinitely either.

    • Lovable Sidekick@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      3 months ago

      Very good analogy. They’re also ignoring that getting faster and faster at reaching a 50% success rate (a totally unacceptable success rate for meaningful tasks) doesn’t imply ever achieving consistently acceptable success.

    • Voroxpete@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      My son has doubled in size every month for the last few months. At this rate he’ll be fifty foot tall by the time he’s seven years old.

      Yeah, it’s a stupid claim to make on the face of it. It also ignores practical realities. The first is those is training data, and the second is context windows. The idea that AI will successfully write a novel or code a large scale piece of software like a video game would require them to be able to hold that entire thing in their context window at once. Context windows are strongly tied to hardware usage, so scaling them to the point where they’re big enough for an entire novel may not ever be feasible (at least from a cost/benefit perspective).

      I think there’s also the issue of how you define “success” for the purpose of a study like this. The article claims that AI may one day write a novel, but how do you define “successfully” writing a novel? Is the goal here that one day we’ll have a machine that can produce algorithmically mediocre works of art? What’s the value in that?

  • SatanClaws@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    Is the performance increase related to computing power? I suspect the undelying massive datacenters running the cloud based LLMs are expanding at a similar rate…

  • db0@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    This is such bullshit. Models have already consumed all available data and have nothing left to consume, whole needing exponentially more data for progressive advancements

    • squaresinger@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      This. It’s the old “to the moon” mentality.

      If my 2yo continues to grow at the current rate, we’ll have to buy new doors soon becouse at age 10 the kid will be the tallest person on Earth.

    • PushButton@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      Apparently, throwing more data at it will not help much from now on… But anyway what they’re saying, I can’t trust the snake oil seller, he is suspicious…

  • ReallyActuallyFrankenstein@lemmynsfw.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    3 months ago

    Is it just me, or is this graph (first graph in the article) completely unintelligible?

    The X-axis being time is self-explanatory, but the Y-axis is somehow exponential time but then also mapping random milestones of performance, meaning those milestones are hard-linked to that time-based Y-axis? What?

  • alehel@lemmy.zip
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    Then why do I feel like it’s programming abilites are getting worse? I’ve stopped paying for it now because it causes more frustration than anything else. Works for simple “how can I simplyfi this code” queries when my head hurts, but that’s about it.

    • squaresinger@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      Only if people give up on the whole concept by then. Each new generation of AI model takes more energy than the last.

  • confuser@lemmy.zip
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    I saw something once that explained how you can have an ai trained on a set of soccer games and have it generate soccer games as a use for it.

    The idea is that the model has compressed all the soccer games into a smaller data size form than the total of having let’s say 100+ games on video or whatever.

    That’s the real utility I see in generative ai that I know can keep going basically as long as we want to.

  • vane@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    I very much like those huge generalizations in AI articles that makes you small and stupid. Those generalizations proves nothing but they sound like something big is coming. It’s parody. How long we see them before people wake up ? Just wait 2 more years and AI will be better bro. You’re not using AI properly, you need to learn how to use AI bro. You need to use different model for this task bro. Just pay for corporate products bro. Amount of junk of top of this pile of shit is amusing.

    • catty@lemmy.worldBanned
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      Because so much money has been thrown at it, for startups, for power generation, for investors, that this is little more than marketing for retail investors to buy into.

      • vane@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 months ago

        That’s no doubt that they pour money to machines instead of people. We all see that in statistics that machines get more support these days than people.

  • altphoto@lemmy.today
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    They need to invent an inquiring-gpt or Q-GPT. Otherwise they’ll need humans to do the digging.