• osugi_sakae@midwest.social
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 days ago

    Education is one area where GenAI is having a huge impact. Teachers work with text and language all day long. They have too much to do and not enough time to do it. Ideally, for example, they should “differentiate” for EACH and EVERY student. Of course that almost never happens, but second best is to differentiate for specific groups - students with IEPs (special ed), English Learners, maybe advanced / gifted.

    More tech aware teachers are now using ChatGPT and friends to help them do this. They are (usually) subject area experts, so they can quickly read through a generated or modified text and fix or remove errors - hallucinations are less (ime) of an issue in this situation. Now, instead of one reading that only a few students can actually understand, they have three at different levels, each with their own DOK questions.

    People have started saying “AI won’t replace teachers. Teachers who use AI will replace teachers who don’t.”

    Of course, it will be interesting to see what happens when VC funding dries up, and the AI companies can’t afford to lose money on every single interaction. Like with everything else in USA education, better off districts may be able to afford AI, and less-well-off (aka black / brown / poor) districts may not be able to.

  • eran_morad@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 days ago

    I’m buying semis. I don’t see AI, construed broadly, as ever shrinking from its current position.

  • NigelFrobisher@aussie.zone
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 days ago

    At a beach restaurant the other night I kept hearing a loud American voice cut across all conversation, going on and on about “AI” and how it would get into all human “workflows” (new buzzword?). His confidence and loudness was only matched by his obvious lack of understanding of how LLMs actually work.

      • ameancow@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        2 days ago

        I would also add “hopeful delusionals” and “unhinged cultist” to that list of labels.

        Seriously, we have people right now making their plans for what they’re going to do with their lives once Artificial Super Intelligence emerges and changes the entire world to some kind of post-scarcity, Star-Trek world where literally everyone is wealthy and nobody has to work. They think this is only several years away. Not a tiny number either, and they exist on a broad spectrum.

        Our species is so desperate for help from beyond, a savior that will change the current status-quo. We’ve been making fantasies and stories to indulge this desire for millenia and this is just the latest incarnation.

        No company on Earth is going to develop any kind of machine or tool that will destabilize the economic markets of our capitalist world. A LOT has to change before anyone will even dream of upending centuries of wealth-building.

      • AItoothbrush@lemmy.zip
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 days ago

        AI itself too i guess. Also i have to point this out every time but my username was chosen way before all this shit blew up into our faces. Ive used this one on every platform for years.

    • Chaotic Entropy@feddit.uk
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 days ago

      Some people can only hear “AI means I can pay people less/get rid of them entirely” and stop listening.

      • anon_8675309@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 days ago

        AI means C level jobs should be on the block as well. The board can make decisions based on their output.

        • Knock_Knock_Lemmy_In@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 days ago

          The whole ex-Mckinsey management layer is at risk. Whole teams of people who were dedicated to producing pretty slides with “action titles” for managers higher up the chain to consume and regurgitate are now having their lunch eaten by AI.

    • Blackmist@feddit.uk
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 days ago

      I’ve noticed that the people most vocal about wanting to use AI get very coy when you ask them what it should actually do.

      • ameancow@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        2 days ago

        I also notice the ONLY people who can offer firsthand reports how it’s actually useful in any way are in a very, very narrow niche.

        Basically, if you’re not a programmer, and even then a very select set of programmers, then your life is completely unimpacted by generative AI broadly. (Not counting the millions of students who used it to write papers for them.)

        AI is currently one of those solutions in search of a problem. In its current state, it can’t really do anything useful broadly. It can make your written work sound more professional and at the same time, more mediocre. It can generate very convincing pictures if you invest enough time into trying to decode the best sequence of prompts and literally just get lucky, but it’s far too inacurate and inconsistent to generate say, a fully illustrated comic book or cartoon, unless you already have a lot of talent in that field. I have tried many times to use AI in my current job to analyze PDF documents and spreadsheets and it’s still completely unable to do work that requires mathematics as well as contextual understanding of what that math represents.

        You can have really fun or cool conversations with it, but it’s not exactly captivating. It is also wildly inaccurate for daily use. I ask it for help finding songs by describing the lyrics and other clues, and it confidentially points me to non-existing albums by hallucinated artists.

        I have no doubt in time it’s going to radically change our world, but that time frame is going to require a LOT more time and baking before it’s done. Despite how excited a few select people are, nothing is changing overnight. We’re going to have a century-long “singularity” and won’t realize we’ve been through it until it’s done. As history tends to go.

        • icecreamtaco@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          2 days ago

          “AI, how do I do <obscure thing> in <complex programming framework>”

          “Here is some <language> code. Please fix any errors: <paste code here>”

          These save me hours of work on a regular basis and I don’t even use the paid tier of ChatGPT for it. Especially the first one because I used to read half the documentation to answer that question. Results are accurate 80% of the time, and the other 20% is close enough that I can fix it in a few minutes. I’m not in an obscure AI related field, any programmer can benefit from stuff like this.

    • Zement@feddit.nl
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      2 days ago

      I really like the idea of an LLM being narrowly configured to filter, summarize data which comes in at a irregular/organic form.

      You would have to do it multiples in parallel with different models and slightly different configurations to reduce hallucinations (Similar to sensor redundancies in Industrial Safety Levels) but still, … that alone is a game changer in “parsing the real world” … that energy amount needed to do this “right >= 3x” is cut short by removing the safety and redundancy because the hallucinations only become apparent down the line somewhere and only sometimes.

      They poison their own well because they jump directly to the enshittyfication stage.

      So people talking about embedding it into workflow… hi… here I am! =D

      • AA5B@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 days ago

        A buddy of mine has been doing this for months. As a manager, his first use case was summarizing the statuses of his team into a team status. Arguably hallucinations aren’t critical

        • sudneo@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 days ago

          I would argue that this makes the process microscopically more efficient and macroscopically way less efficient. That whole process probably is useless, and imagine wasting so much energy, water and computing power just to speed this useless process up and saving a handful of minutes (I am a lead and it takes me 2/3 minutes to put together a status of my team, and I don’t usually even request a status from each member).

          I keep saying this to everyone in my company who pushes for LLMs for administrative tasks: if you feel like LLMs can do this task, we should stop doing it at all because it means we are just going through the motions and pleasing a process without purpose. You will have people producing reports via LLM from a one-line prompt, the manager assembling it together with LLM and at vest someone reading it distilling it once again with LLMs. It is all a great waste of money, energy, time, cognitive effort that doesn’t benefit anybody.

          As soon as someone proposes to introduce LLMs in a process, raise with cutting that process altogether. Let’s produce less bullshit, instead of more while polluting even more in the process.

  • iAvicenna@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 days ago

    oh wow who would have guessed that business consultancy companies are generally built on top of bullshitting about things which you dont really have a grasp on

  • computerscientistII@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 days ago

    I saved a lot of time due to ChatGPT. Need to sign up some of my pupils for a competition by uploading their data in a csv-File to some plattform? Just copy and paste their data into chsatgpt and prompt it to create the file. The boss (headmaster) wants some reasoning why I need some paid time for certain projects? Let ChatGPT do the reasoning. Need some exercises for one of my classes that doesn’t really come to grips with while-loops? let ChatGPT create those exercises (some smartasses will of course have ChatGPT then solve those exercises). The list goes on…

    • FiskFisk33@startrek.website
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      Just copy and paste [student personal data] into [3rd parties database]

      Yeah, thats a problem, especially in Europe. Im unsure about US, but its definitely a breach of GDPR.

    • solstice@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      ChatGPT is basically like a really good intern, and I use it heavily that way. I run literally every email through it and say “respond to so and so, say xyz” and then maybe a little refining, copy paste, done.

      The other day, my boss sent me an excel file with a shitload of data in it that he wanted me to analyze some such way. I just copy pasted it into gpt and asked it, and it spit out the correct response. Then my boss asked me to do something else that required a bit of excel finagling that I didn’t really know how to do, so i asked gpt, and it told me the formula, which worked immediately first try.

      So basically it helps me accomplish tasks in seconds that previously would’ve taken hours. If anything, I think markets are currently undervalued, because remarkably, fucking NONE of my colleagues or friends are using it at all yet. Once there’s widespread adoption, which will pretty much have to happen if anyone wants to stay competitive once it gains more traction, look out…

      • AA5B@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 days ago

        The thing about tech bubbles is everyone rushes in full bore, on the hope that they can be the ones whose moonshot goes the distance. However even in the case where the technology achieves all its promise, most of those early attempts will not. Soon enough, we’ll be down to the top few, and only their datacenters will need to exist. Many of these failures will go away

        • Flying Squid@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 days ago

          Many of these failures will go away

          In a rational, non-capitalist world, yes. In our world, all of those data centres will last until they can’t find a way to squeeze some sort of profit out of them.

    • wrekone@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      One time, I needed to convince my boss’s boss that we needed to do something, and he wanted it in writing. Guess who wrote the proposal? And far more eloquently than I could have alone, in the time allowed. It required some good prompts, attentive proofreading, and a few drafts. But in the end, it was quite effective.

    • ssfckdt@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      Yeah, and Wikipedia is one of the most useful sites on the net, but it didn’t exactly result in the entire web becoming crowdsourced.

    • funkless_eck@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      The poem about AI that often gets posted says “What are you trying to avoid? The living [of a life]?”

      And yeah, that’s what it’s for, dodging shit you don’t want to do. I gotta produce some useless bullshit that no one’s going to read or care about: AI.

      I don’t even mind AI art for things like LinkedIn posts, blogs like “What is warehouse management?” or “Top 10 finance trends in 2025” - SEO spam that no human will read. No one wants to write it, read it, or care about it- its just a x kb file to tell Google to look here.

  • nroth@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 days ago

    “Built to do my art and writing so I can do my laundry and dishes” – Embodied agents is where the real value is. The chatbots are just fancy tech demos that folks started selling because people were buying.

    • bradd@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      Eh, my best coworker is an LLM. Full of shit, like the rest of them, but always available and willing to help out.

        • AA5B@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 days ago

          Just like every other coworker, it’s important to know what tasks they do well and where they typically need help

          • finitebanjo@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            2 days ago

            Lmao your stance is really “every coworker makes all product lower quality by nature of existence”? Thats some hardcore Cope you’re smoking.

            • AA5B@lemmy.world
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 days ago

              Every coworker has a specific type of task they do well and known limits you should pay attention to.

              • finitebanjo@lemmy.world
                link
                fedilink
                English
                arrow-up
                0
                ·
                2 days ago

                Yes and therefor any two employees must never be allowed to speak to each other. You know, because it makes all of their work worse quality. /s

                • sugar_in_your_tea@sh.itjust.works
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  1 day ago

                  That’s quite the extreme interpretation.

                  I’m a lead software dev, and when deadlines are close, I absolutely divvy up tasks based on ability. We’re a webapp shop with 2D and 3D components, and I have the following on my team:

                  • 2 BE devs with solid math experience
                  • 1 senior BE without formal education, but lots of knowledge on frameworks
                  • 1 junior fullstack that we hired as primarily backend (about 75/25 split)
                  • 2 senior FE devs, one with a QA background
                  • 2 mid level FEs who crank out code (but miss some edge cases)
                  • 1 junior FE

                  That’s across two teams, and one of the senior FEs is starting to take over the other team.

                  If we’re at the start of development, I’ll pair tasks between juniors and seniors so the juniors get more experience. When deadlines are close, I’ll pair tasks with the most competent dev in that area and have the juniors provide support (write tests, fix tech debt, etc).

                  The same goes for AI. It’s useful at the start of a project to understand the code and gen some boilerplate, but I’m going to leave it to the side when tricky bugs need to get fixed or we can’t tolerate as many new bugs. AI is like a really motivated junior, it’s quick to give answers but slow to check their accuracy.

    • nroth@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      Though the image generators are actually good. The visual arts will never be the same after this

      • TheBrideWoreCrimson@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        2 days ago

        I’ve been thinking about this a lot recently. No, we’re not there yet, may never be. Compare what Jesar, one of my favorite artists, can do - and that was in the oh-so-long-ago 2000s - and what an AI can do. It’s simply not up to the task. I do use AI a lot to create what is basically utility art. But it depends on pre-defined textual or visual inputs whereas only an artist can have divine inspiration. AI is more of a sterile tool, like interactive clipart, if you will.

      • LifeInMultipleChoice@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        3 days ago

        Compare it to the microwave. Is it good at something, yes. But if you shoot your fucking turkey in it at Thanksgiving and expect good results, you’re ignorant of how it works. Most people are expecting language models to do shit that aren’t meant to. Most of it isn’t new technology but old tech that people slapped a label on as well. I wasn’t playing Soul Caliber on the Dreamcast against AI openents… Yet now they are called AI opponents with no requirements to be different. GoldenEye on N64 was man VS AI. Madden 1995… AI. “Where did this AI boom come from!”

        Marketing and mislabeling. Online classes, call it AI. Photo editors, call it AI.

        • sugar_in_your_tea@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 day ago

          I wasn’t playing Soul Caliber on the Dreamcast against AI openents…

          Maybe terminology differs by region, but I absolutely played against AI as a kid. When I set up a game of Command and Conquer or something, I’d pick the number of AI opponents. Sometimes we’d call them bots (more common in FPS) or “the computer” or “CPU” (esp in Civ and other TBS), but I distinctly remember calling RTS SP opponents “AI” and I think many games used that terminology during the 90s.

          What frustrates me is the opposite of what you’re saying, people have changed the meaning of “AI” from a human programmed opponent to a statistical model. When I played against “AI” 20-30 years ago, I was playing against something a human crafted and tuned. These days, I don’t play against “AI” because “AI” generates text, images, and video from a statistical model and can’t really play games. AI is something that runs in the cloud, with maybe a small portion on phones and Windows computers to do simple tasks where the network would add too much latency.

  • 2pt_perversion@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    3 days ago

    There is this seeming need to discredit AI from some people that goes overboard. Some friends and family who have never really used LLMs outside of Google search feel compelled to tell me how bad it is.

    But generative AIs are really good at tasks I wouldn’t have imagined a computer doing just a few year ago. Even if they plateaued in place where they are right now it would lead to major shakeups in humanity’s current workflow. It’s not just hype.

    The part that is over hyped is companies trying to jump the gun and wholesale replace workers with unproven AI substitutes. And of course the companies who try to shove AI where it doesn’t really fit, like AI enabled fridges and toasters.

    • andallthat@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      2 days ago

      Goldman Sachs, quote from the article:

      “AI technology is exceptionally expensive, and to justify those costs, the technology must be able to solve complex problems, which it isn’t designed to do.”

      Generative AI can indeed do impressive things from a technical standpoint, but not enough revenue has been generated so far to offset the enormous costs. Like for other technologies, It might just take time (remember how many billions Amazon burned before turning into a cash-generating machine? And Uber has also just started turning some profit) + a great deal of enshittification once more people and companies are dependent. Or it might just be a bubble.

      As humans we’re not great at predicting these things including of course me. My personal prediction? A few companies will make money, especially the ones that start selling AI as a service at increasingly high costs, many others will fail and both AI enthusiasts and detractors will claim they were right all along.

    • Valmond@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 days ago

      Like what outcome?

      I have seen gains on cell detection, but it’s “just” a bit better.

    • ssfckdt@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      This is easy to say about the output of AIs… if you don’t check their work.

      Alas, checking for accuracy these days seems to be considered old fogey stuff.

    • Modern_medicine_isnt@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      See now, I would prefer AI in my toaster. It should be able to learn to adjust the cook time to what I want no matter what type of bread I put in it. Though is that realky AI? It could be. Same with my fridge. Learn what gets used and what doesn’t. Then give my wife the numbers on that damn clear box of salad she buys at costco everytime, which take up a ton of space and always goes bad before she eats even 5% of it. These would be practical benefits to the crap that is day to day life. And far more impactful then search results I can’t trust.

      • AA5B@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        2 days ago

        I agree with your wife: there’s always an aspirational salad in the fridge. For most foods, I’m pretty good at not buying stuff we won’t eat, but we always should eat more veggies. I don’t know how to persuade us to eat more veggies, but step 1 is availability. Like that Reddit meme

        1. Availability
        2. ???
        3. Profit by improved health
      • Someonelol@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 days ago

        You better believe that AI-powered toaster would only accept authorized bread from a bakery that paid top dollar to the company that makes them. To ensure the best quality possible and save you from inferior toast, of course.

      • ssfckdt@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        3 days ago

        There’s a good point here that like about 80% of what we’re calling AI right now… isn’t even AI or even LLM. It’s just… algorithm, code, plain old math. I’m pretty sure someone is going to refer to a calculator as AI soon. “Wow, it knows math! Just like a person! Amazing technology!”

        (That’s putting aside the very question of whether LLMs should even qualify as AIs at all.)

        • Modern_medicine_isnt@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 days ago

          In my professional experience, AI seems to be just a faster way to generate an algorithm that is really hard to debug. Though I am dev-ops/sre so I am not as deep in it as the devs.

          • AA5B@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 days ago

            As a devops person, I’m constantly jumping back and forth to whatever programming language and tools each team uses. Sometimes it takes a bit to find the context, and I’m hoping ai can help. Unfortunately, allowing the ai to see code is currently off limits by corporate policy, so it only helps in those situations where I need to generate boilerplate

            • Modern_medicine_isnt@lemmy.world
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 days ago

              In my jobs there have slways been certain stule requirements to the code. AI doesn’t take those into account. So I would have to rework the code anyway. And of course there are the local libraries it know nothing about.

              • AA5B@lemmy.world
                link
                fedilink
                English
                arrow-up
                0
                ·
                2 days ago

                Fight technology with technology. I’m sure you can specify a style for it to generate, but we already run everything through a prettifier configured for what we look for …. Unless you mean a higher order like naming or architecture

                • Modern_medicine_isnt@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  1 day ago

                  Lol, the lead can’t spec the style, he just reviews the code and asks for changes. Sometimes it’s just that we already have a method that does a similar thing, so we should use it. Of course an AI wouldn’t know about that unless you gave it access to your code. And given how speed first AI companies are, I would never trust that data with them. But other times it’s just the leads personal preference.

    • Eldritch@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      Computers have always been good at pattern recognition. This isn’t new. LLM are not a type of actual AI. They are programs capable of recognizing patterns and Loosely reproducing them in semi randomized ways. The reason these so-called generative AI Solutions have trouble generating the right number of fingers. Is not only because they have no idea how many fingers a person is supposed to have. They have no idea what a finger is.

      The same goes for code completion. They will just generate something that fills the pattern they’re told to look for. It doesn’t matter if it’s right or wrong. Because they have no concept of what is right or wrong Beyond fitting the pattern. Not to mention that we’ve had code completion software for over a decade at this point. Llms do it less efficiently and less reliably. The only upside of them is that sometimes they can recognize and suggest a pattern that those programming the other coding helpers might have missed. Outside of that. Such as generating act like whole blocks of code or even entire programs. You can’t even get an llm to reliably spit out a hello world program.

      • brie@programming.dev
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 days ago

        Large context window LLMs are able to do quite a bit more than filling the gaps and completion. They can edit multiple files.

        Yet, they’re unreliable, as they hallucinate all the time. Debugging LLM-generated code is a new skill, and it’s up to you to decide to learn it or not. I see quite an even split among devs. I think it’s worth it, though once it took me two hours to find a very obscure bug in LLM-generated code.

        • sudneo@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 days ago

          Humans are notoriously worse at tasks that have to do with reviewing than they are at tasks that have to do with creating. Editing an article is more boring and painful than writing it. Understanding and debugging code is much harder than writing it etc., observing someone cooking to spot mistakes is more boring than cooking etc.

          This also fights with the attention required to perform those tasks, which means a higher ratio of reviewing vs creating tasks leads to lower quality output because attention is depleted at some point and mistakes slip in. All this with the additional “bonus” to have to pay for the tool AND the human reviewing while also wasting tons of water and energy. I think it’s wise to ask ourselves whether this makes sense at all.

          • brie@programming.dev
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 days ago

            To make sense of that, figure out what pays more observing/editing or cooking/writing. Big shekels will make boring parts exciting

            • sudneo@lemm.ee
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 days ago

              Think also the amount of people doing both. Also writers earn way more than editors, and stellar chefs earn way more than cooking critics.

              If you think devs will be paid more to review GPT code, well, I would love to have your optimism.

              • brie@programming.dev
                link
                fedilink
                English
                arrow-up
                0
                ·
                2 days ago

                I’m too unfamiliar with the cooking and writing/publishing biz. I’d rather not use this analogy.

                I can see many business guys paying for something like Devin, making a mess, then hiring someone to fix it. I can see companies not hiring junior devs, and requiring old devs to learn to generate and debug. Just like they required devs to be “full stack”. You can easily prevent that if you have your own company. If … Do you have your own company?

                • sudneo@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  2 days ago

                  I don’t, like 99% of people don’t or won’t. My job is safe, I am arguing from a collective perspective.

                  I simply don’t think companies will act like that. Also the mere reduction of total number of positions will compress salaries.

        • NigelFrobisher@aussie.zone
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          2 days ago

          I have one of those at work now, but my experience with it is still quite limited. With Copilot it was quite useful for knocking up quick boutique solutions for particular problems (stitch together a load of PDFs sorted on a name heading), with the proviso that you might end up having to repair bleed between dependency versions and repair syntax. I couldn’t trust it with big refactors of existing systems.

        • cley_faye@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 days ago

          If you consider debugging broken LLM-generated code to be a skill… sure, go for it. But, since generated code is able to use tons of unknown side effects and other seemingly (for humans) random stuff to achieve its goal, I’d rather take the other approach, where it takes a human half an hour to write the code that some LLM could generate in seconds, and not have to learn how to parse random mumbo jumbo from a machine, while getting a working result.

          Writing code is far from being the longest part of the job; and you gingerly decided that making the tedious part even more tedious is a great idea to shorten the already short part of it…

          • brie@programming.dev
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 days ago

            It’s similar to fixing code written by interns. Why hire interns at all, eh?

            Is it faster to generate then debug or write everything? Needs to be properly tested. At the very least many devs have the perception of being faster, and perception sells.

            It actually makes writing web apps less tedious. The longest part of a dev job is pretending to work actually, but that’s no different from any office jerb.

      • JohnEdwa@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 days ago

        “It’s part of the history of the field of artificial intelligence that every time somebody figured out how to make a computer do something—play good checkers, solve simple but relatively informal problems—there was a chorus of critics to say, ‘that’s not thinking’”
        -Pamela McCorduck

        “AI is whatever hasn’t been done yet.”
        - Larry Tesler

        That’s the curse of the AI Effect.
        Nothing will ever be “an actual AI” until we cross the barrier to an actual human-like general artificial intelligence like Cortana from Halo, and even then people will claim it isn’t actually intelligent.

        • ssfckdt@lemmy.blahaj.zone
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 days ago

          I mean, I think intelligence requires the ability to integrate new information into one’s knowledge base. LLMs can’t do that, they have to be trained on a fixed corpus.

          Also, LLMs have a pretty shit-tastic track record of being able to differentiate correct data from bullshit, which is a pretty essential facet of intelligence IMO

          • JohnEdwa@sopuli.xyz
            link
            fedilink
            English
            arrow-up
            0
            ·
            3 days ago

            LLMs have a perfect track record of doing exactly what they were designed to, take an input and create a plausible output that looks like it was written by a human. They just completely lack the part in the middle that properly understands what it gets as the input and makes sure the output is factually correct, because if it did have that then it wouldn’t be an LLM any more, it would be an AGI.
            The “artificial” in AI does also stand for the meaning of “fake” - something that looks and feels like it is intelligent, but actually isn’t.

        • Eldritch@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 days ago

          Well at least until those who study intelligence and self-awareness actually come up with a comprehensive definition for it. Something we don’t even have currently. Which makes the situation even more silly. The people selling LLMs and AGNs as artificial intelligence are the PT Barnum of the modern era. This way to the egress folks come see the magnificent egress!

          • JohnEdwa@sopuli.xyz
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            3 days ago

            They already did. AGI - artificial general intelligence.

            The thing is, AGI and AI are different things. Like your “LLMs aren’t real AI” thing , large language models are a type of machine learning model, and machine learning is a field of study in artificial intelligence.
            LLMs are AI. Search engines are AI. Recommendation algorithms are AI. Siri, Alexa, self driving cars, Midjourney, Elevenlabs, every single video game with computer players, they are all AI. Because the term “Artificial Intelligence” by itself is extremely loose, and includes the types of narrow AI all of those are.
            Which then get hit by the AI Effect, and become “just another thing computers can do now”, and therefore, “not AI”.

            • ssfckdt@lemmy.blahaj.zone
              link
              fedilink
              English
              arrow-up
              0
              ·
              3 days ago

              That’s a disturbing handwave. “We don’t really know what intelligence is, so therefore, anything we call intelligence is fair game”

              A thermometer tells me what temperature it is. It senses the ambient heat energy and responds with a numeric indicator. Is that intelligence?

              My microwave stops when it notices steam from my popcorn bag. Is that intelligence?

              If I open an encyclopedia book to a page about computers, it tells me a bunch of information about computers. Is that intelligence?

            • Eldritch@lemmy.world
              link
              fedilink
              English
              arrow-up
              0
              ·
              3 days ago

              That just Compares it to human level intelligence. Something which we cannot currently even quantify. Let alone understand. It’s ultimately a comparison, a simile not a scientific definition.

              Search engines have always been databases. With interfaces programmed by humans. Not ai. They’ve never suddenly gained new functionality inexplicably. If there’s a new feature someone programmed it.

              Search engines are however becoming llms and are getting worse for it. Unless you think eating rocks and glue is particularly intelligent. Because there is no comprehension there. It’s simply trying to make its output match patterns it recognizes. Which is a precursor step. But is not “intelligence”. Unless a program doing what it’s programed to do is artificial intelligence. Which is such a meaningless measure because that would mean notepad is artificial intelligence. Windows is artificial intelligence. Linux is artificial intelligence.

                • Saledovil@sh.itjust.works
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  2 days ago

                  You can’t just throw out random Wikipedia links. For example, the Article on AGI explicitly says we don’t have a definition of what human level cognition actually is. Which is what the person you were replying to was saying. You’re doing a fallacious appeal to authority, except that the authority doesn’t agree with you.

      • Mak'@pawb.social
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 days ago

        I never know what to think when I come across a comment like this one—which does describe, even if only at a surface level, how an LLM works—with 50% downvotes. Like, are people angry at reality, is that it?

        • Naz@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 days ago

          Downvoting someone on the Internet is easier than tangentially modifying reality in a measurable way

        • Eldritch@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 days ago

          With as much misinformation that’s being spread about regarding LLMs. It would only lose more people’s comprehension to go into anything more than a generalization.

          The problem is people are being sold AGI. But chat GPT and all these other tools don’t even remotely qualify for that. They’re really nothing more than a glorified Alice chatbot system on steroids. The one neat new trick to all this is that they’ve automated the training a bit. But these llms have no more comprehension of their output or the input they were given than something like the old Alice chatbot.

          These tools have been described as artificial intelligence to layman for decades at this point. It makes it really hard to change that calcified opinion. People would rather believe that it’s some magical thing not just probability and maths.

          • snooggums@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            3 days ago

            They are bullshit machines, trained to output something that users think is the right output.

    • buddascrayon@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      The part that is over hyped is companies trying to jump the gun and wholesale replace workers with unproven AI substitutes. And of course the companies who try to shove AI where it doesn’t really fit, like AI enabled fridges and toasters.

      This is literally the hype. This is the hype that is dying and needs to die. Because generative AI is a tool with fairly specific uses. But it is being marketed by literally everyone who has it as General AI that can “DO ALL THE THINGS!” which it’s not and never will be.

      • five82@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 days ago

        The obsession with replacing workers with AI isn’t going to die. It’s too late. The large financial company that I work for has been obsessively tracking hours saved in developer time with GitHub Copilot. I’m an older developer and I was warned this week that my job will be eliminated soon.

        • buddascrayon@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          3 days ago

          The large financial company that I work for

          So the company that is obsessed with money that you work for has discovered a way to (they think) make more money by getting rid of you and you’re surprised by this?

          At least you’ve been forewarned. Take the opportunity to abandon ship. Don’t be the last one standing when the music stops.

    • sudneo@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      Even if they plateaued in place where they are right now it would lead to major shakeups in humanity’s current workflow

      Like which one? Because it’s now 2 years we have chatGPT and already quite a lot of (good?) models. Which shakeup do you think is happening or going to happen?

      • AA5B@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 days ago

        I don’t know anything about the online news business but it certainly appears to have changed. Most of it is dreck, either way, and those organizations are not a positive contributor to society, but they are there, it is a business, and it has changed society

        • sudneo@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 days ago

          I don’t see the change. Sure, there are spam websites with AI content that were not there before, but is this news business at all? All major publishers and newspapers don’t (seem to) use AI as far as I can tell.

          Also I would argue this is no much of a change except maybe in simplicity to generate fluff. All of this existed already for 20 years now, and it’s a byproduct of the online advertisement business (that for sure was a major change in society!). AI pieces are just yet another way to generate content in the hope of getting views.

        • sudneo@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 days ago

          Oh boy…what can possibly go wrong for documents where small minutiae like wording can make a huge difference.

          • figjam@midwest.social
            link
            fedilink
            English
            arrow-up
            0
            ·
            3 days ago

            Creating legal documents, no. Reviewing legal documents for errors and inaccuracies totally.

            • cley_faye@lemmy.world
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 days ago

              No, not that either. Unless you consider “use LLM to summarize the changes/errors/inaccuracies, then have a human read the whole thing again” an improvement over “just have a human read the whole thing”.

              Because LLM will do all these things:

              • point you toward issues
              • point you toward non-issues
              • not point you toward issues
              • change stuff even when “instructed” not to

              If there is one thing you don’t want to throw an LLM at without full, unbiased review, it’s documents where the wording is legally binding. And if you have to do a full, unbiased review to begin with, where you can’t even trust your tool to have highlighted all the important parts, you may as well not bother with the tool.

            • sudneo@lemm.ee
              link
              fedilink
              English
              arrow-up
              0
              ·
              3 days ago

              I really can’t see this being done by any sane person. Why would you have a generator of text reviewing stuff (besides grammar)? Do you have any reference of some companies doing this, perhaps?

              • figjam@midwest.social
                link
                fedilink
                English
                arrow-up
                0
                ·
                3 days ago

                Its complex pattern matching and looking up existing case law online. This work has been outsourced to contracting companies for at least 7 years that I’m aware of. If it is something that can be documented in a run book for non professionals to do for twenty cents on the dollar then there is no reason it can’t be done by a script for .002.

                • cley_faye@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  2 days ago

                  Aside from a handful of business that tried to do that and failed miserably, some of them failing in actual court, you mean?

      • locuester@lemmy.zip
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 days ago

        Computer programming has radically changed. Huge help having llm auto complete and chat built in. IDEs like Cursor and Windsurf.

        I’ve been a developer for 35 years. This is shaking it up as much as the internet did.

        • Nalivai@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          3 days ago

          I quit my previous job in part because I couldn’t deal with the influx of terrible, unreliable, dangerous, bloated, nonsensical, not even working code that was suddenly pushed into one of the projects I was working on. That project is now completely dead, they froze it on some arbitrary version.
          When junior dev makes a mistake, you can explain it to them and they will not make it again. When they use llm to make a mistake, there is nothing to explain to anyone.
          I compare this shake more to an earthquake than to anything positive you can associate with shaking.

          • locuester@lemmy.zip
            link
            fedilink
            English
            arrow-up
            0
            ·
            3 days ago

            This is a problem with your team/project. It’s not a problem with the technology.

            • Nalivai@lemmy.world
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 days ago

              A technology that makes people put bad code is a problematic technology. If your team/project managed to overcome it’s problems so far doesn’t mean it is good or overall helpful. Peoole not seeing the problem is actually the worst part.

              • locuester@lemmy.zip
                link
                fedilink
                English
                arrow-up
                0
                ·
                2 days ago

                Sir, I use it to assist me in programming. I don’t use it to write entire files or functions. It’s a pattern recognizer.

                Your team had people who didn’t review code. That’s a problem.

          • InnerScientist@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            3 days ago

            And so, the problem wasn’t the ai/llm, it was the person who said “looks good” without even looking at the generated code, and then the person who read that pull request and said, again without reading the code, “lgtm”.

            If you have good policies then it doesn’t matter how many bad practice’s are used, it still won’t be merged.

            The only overhead is that you have to read all the requests but if it’s an internal project then telling everyone to read and understand their code shouldn’t be the issue.

            • Nalivai@lemmy.world
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 days ago

              The problem here is that a lot of the time looking for hidden problem is harder than writing good code from scratch. And you will always be at a danger that llm snuck some sneaky undefined behaviour past you. There is a whole plethora of standards, conventions, and good practices that help humans to avoid it, which llm can ignore at any random point.
              So you’re either not spending enough time on review or missing whole lot of bullshit. In my experience, in my field, right now, this review time is more time consuming and more painful than avoiding it in the first place.
              Don’t underestimate how degrading and energy sucking it is for a professional to spend most of the working time sitting through autogenerated garbage, and how inefficient it is.

        • areyouevenreal@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 days ago

          Exactly this. Things have already changed and are changing as more and more people learn how and where to use these technologies. I have seen even teachers use this stuff who have limited grasp of technology in general.

          • AA5B@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 days ago

            My kid’s teachers had what I thought was a fantastic approach - have the kids write an outline. Use an LLM to generate an essay from that outline, then critique the essay

        • sudneo@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 days ago

          I hardly see it changed to be honest. I work in the field too and I can imagine LLMs being good at producing decent boilerplate straight out of documentation, but nothing more complex than that.

          I often use LLMs to work on my personal projects and - for example - often Claude or ChatGPT 4o spit out programs that don’t compile, use inexistent functions, are bloated etc. Possibly for languages with more training (like Python) they do better, but I can’t see it as a “radical change” and more like a well configured snippet plugin and auto complete feature.

          LLMs can’t count, can’t analyze novel problems (by definition) and provide innovative solutions…why would they radically change programming?

          • aleq@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 days ago

            I hardly see it changed to be honest. I work in the field too and I can imagine LLMs being good at producing decent boilerplate straight out of documentation, but nothing more complex than that.

            I think one of the top lists on advent of code this year is a cheater that fully automated the solutions using LLMs. Not sure which LLM though, I use LLMs quite a bit and ChatGPT 4o frequently tells me nonsense like “perhaps subtracting by zero is affecting your results” (issues I thought were already gone in GPT 4, but I guess not, Sonnet 3.5 does a bit better in this regard).

            • sudneo@lemm.ee
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 days ago

              Maybe some postmortem analysis will be interesting. The AoC is also a context in which the domain is self-contained and there is probably a ton of training material on similar problems and tasks. I can imagine LLM might do decently there.

              Also there is no big consequence if they don’t and it’s probably possible to bruteforce (which is how many programming tasks have been solved).

              • aleq@lemmy.world
                link
                fedilink
                English
                arrow-up
                0
                ·
                2 days ago

                I think you’re spot on with LLMs being mostly trained on these kinds of tasks. Can’t say I’m an expert in how to build a training set, but I imagine it’s quite easy to do with these kinds of problems because it’s easy to classify a solution as correct or incorrect. This is in contrast to larger problems which are less guided by algorithmic efficiency and more by sound design/architecture.

                Still, I think it’s quite impressive. You don’t have to go very far back in time to have top of the line LLMs unable to solve these kinds of problems.

                Also there is no big consequence if they don’t and it’s probably possible to bruteforce (which is how many programming tasks have been solved).

                Usually with AoC part 1 is brute-forceable, but part 2 is not. Very often part 1 is to find the 100th number, and part 2 is to find the 1 000 000 000 000th number or something. Last year, out of curiosity, I had a brute-force solution for one problem that successfully completed on ~90% of the input. Solution was multi-threaded and running on a 16 core CPU for about 20 days before I gave up. But the LLMs this year (not sure if this was a problem last year) are in the top list of fastest users to solve the problems.

                • sudneo@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  2 days ago

                  Just to precise, when I said bruteforce I didn’t imagine a bruteforce of the calculation, but a brute force of the code. LLMs don’t really calculate either way, but what I mean is more: generate code -> try to run and see if tests work -> if it doesn’t ask again/refine/etc. So essentially you are just asking code until what it spits out is correct (verifiable with tests you are given).

                  But yeah, few years ago this was not possible and I guess it was not due to the training data. Now the problem is that there is not much data left for training, and someone (Bloomberg?) reported that training chatGPT 5 will cost billions of dollars, and it looks like we might be near the peak of what this technology could offer (without any major problem being solved by it to offset the economical and environmental cost).

                  Just from today https://www.techspot.com/news/106068-openai-struggles-chatgpt-5-delays-rising-costs.html

          • locuester@lemmy.zip
            link
            fedilink
            English
            arrow-up
            0
            ·
            3 days ago

            You’re missing it. Use Cursor or Windsurf. The autocomplete will help in so many tedious situations. It’s game changing.

          • areyouevenreal@lemm.ee
            link
            fedilink
            English
            arrow-up
            0
            ·
            3 days ago

            ChatGPT 4o isn’t even the most advanced model, yet I have seen it do things you say it can’t. Maybe work on your prompting.

            • sudneo@lemm.ee
              link
              fedilink
              English
              arrow-up
              0
              ·
              3 days ago

              That is my experience, it’s generally quite decent for small and simple stuff (as I said, distillation of documentation). I use it for rust, where I am sure the training material was much smaller than other languages. It’s not a matter a prompting though, it’s not my prompt that makes it hallucinate functions that don’t exist in libraries or make it write code that doesn’t compile, it’s a feature of the technology itself.

              GPTs are statistical text generators after all, they don’t “understand” the problem.

              • agamemnonymous@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                0
                ·
                3 days ago

                It’s also pretty young, human toddlers hallucinate and make things up. Adults too. Even experts are known to fall prey to bias and misconception.

                I don’t think we know nearly enough about the actual architecture of human intelligence to start asserting an understanding of “understanding”. I think it’s a bit foolish to claim with certainty that LLMs in a MoE framework with self-review fundamentally can’t get there. Unless you can show me, materially, how human “understanding” functions, we’re just speculating on an immature technology.

                • sudneo@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  2 days ago

                  As much as I agree with you, humans can learn a bunch of stuff without first learning the content of the whole internet and without the computing power of a datacenter or consuming the energy of Belgium. Humans learn to count at an early age too, for example.

                  I would say that the burden of proof is therefore reversed. Unless you demonstrate that this technology doesn’t have the natural and inherent limits that statistical text generators (or pixel) have, we can assume that our mind works differently.

                  Also you say immature technology but this technology is not fundamentally (I.e. in terms of principle) different from what Weizenabum’s ELIZA in the '60s. We might have refined model and thrown a ton of data and computing power at it, but we are still talking of programs that use similar principles.

                  So yeah, we don’t understand human intelligence but we can appreciate certain features that absolutely lack on GPTs, like a concept of truth that for humans is natural.

  • Novamdomum@fedia.io
    link
    fedilink
    arrow-up
    0
    ·
    3 days ago

    “Today’s hype will have lasting effects that constrain tomorrow’s possibilities.”

    Nope. No it won’t. I’d love to have the patience to be more diplomatic but they’re just wrong… and dumb.

    I’m getting so sick of these anti AI cultists who seem to be made up of grumpy tech nerds behaving like “I was using AI before it was cool” hipsters and panicking artists and writers. Everyone needs to calm their tits right down. AI isn’t going anywhere. It’s giving creative and executive options to millions of people that just weren’t there before.

    We’re in an adjustment phase right now and boundaries are being re-drawn around what constitutes creativity. My leading theory at the moment is that we’ll all mostly eventually settle down to the idea that AI is just a tool. Once we’re used to it and less starry eyed about it’s output then individual creativity, possibly supported by AI tools, will flourish again. It’s going to come down to the question of whether you prefer reading something cogitated, written, drawn or motion rendered by AI or you enjoy the perspective of a human being more. Both will be true in different scenarios I expect.

    Honestly, I’ve had to nope out of quite a few forums and servers permanently now because all they do in there is circlejerk about the death of AI. Like this one theory that keeps popping up that image generating AI specifically is inevitably going to collapse in on itself and stop producing quality images. The reverse is so obviously true but they just don’t want to see it. Otherwise smart people are just being so stubborn with this and it’s, quite frankly, depressing to see.

    Also, the tech nerds arguing that AI is just a fancy word and pixel regurgitating engine and that we’ll never have an AGI are probably the same people that were really hoping Data would be classified as a sentient lifeform when Bruce Maddox wanted to dissassemble him in “The Measure of a Man”.

    How’s that for whiplash?

    • sudneo@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      Models are not improving, companies are still largely (massively) unprofitable, the tech has a very high environmental impact (and demand) and not a solid business case has been found so far (despite very large investments) after 2 years.

      That AI isn’t going anywhere is possible, but LLM-based tools might also simply follow crypto, VR, metaverses and the other tech “revolutions” that were just hyped and that ended nowhere. I can’t say it will go one way or another, but I disagree with you about “adjustment period”. I think generative AI is cool and fun, but it’s a toy. If companies don’t make money with it, they will eventually stop investing into it.

      Also

      Today’s hype will have lasting effects that constrain tomorrow’s possibilities

      Is absolutely true. Wasting capital (human and economic) on something means that it won’t be used for something else instead. This is especially true now that it’s so hard to get investments for any other business. If all the money right now goes into AI, and IF this turns out to be just hype, we just collectively lost 2, 4, 10 years of research and investments on other areas (for example, environment protection). I am really curious about what makes you think that that sentence is false and stupid.

      • VoterFrog@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 days ago

        Models are not improving? Since when? Last week? Newer models have been scoring higher and higher in both objective and subjective blind tests consistently. This sounds like the kind of delusional anti-AI shit that the OP was talking about. I mean, holy shit, to try to pass off “models aren’t improving” with a straight face.

        • sudneo@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 days ago

          There is a bunch of research showing that model improvement is marginal compared to energy demand and/or amount of training data. OpenAI itself ~1 month ago mentioned that they are seeing a smaller improvements in Orion (I believe) vs GPT4 than there was between GPT 4 and 3. We are also running out of quality data to use for training.

          Essentially what I mean is that the big improvements we have seen in the past seem to be over, now improving a little cost a lot. Considering that the costs are exorbitant and the difference small enough, it’s not impossible to imagine that companies will eventually give up if they can’t monetize this stuff.

          • theherk@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 days ago

            Surely you can see there is a difference between marginal improvement with respect to energy and not improving.

          • icecreamtaco@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 days ago

            Compare Llama 1 to the current state of the art local AI’s. They’re on a completely different level.

            • sudneo@lemm.ee
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 days ago

              Yes, because at the beginning there was tons of room for improvement.

              I mean take openAI word for it: chatGPT 5 is not seeing improvement compared to 4 as much as 4 to 3, and it’s costing a fortune and taking forever. Logarithmic curve, it seems. Also if we run out of data to train, that’s it.

    • GHiLA@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      It’s fucking fantastic news, tbh.

      Here’s my take, let them dismiss it.

      Let em! Remember Bitcoin at $15k after 2019?

      Let em! And it’s justified! If Ai isn’t important right now, then why should its price be inflated to oblivion? Let it fall. Good! Lower prices for those of us that do see the value down the road.

      That’s how speculative investment works. In no way is this bad. Are sales bad? Sit back and enjoy the show.

      • Voroxpete@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 days ago

        Are sales bad?

        Of AI products? By all available metrics, yes, sales for AI driven products are atrocious.

        Even the biggest name in AI is desperately unprofitable. OpenAI has only succeeded in converting 3% of their free users to paid users. To put that on perspective, 40% of regular Spotify users are on premium plans.

        And those paid plans don’t even cover what it costs to run the service for those users. Currently OpenAI are intending to double their subscription costs over the next five years, and that still won’t be enough to make their service profitable. And that’s assuming that they don’t lose subscribers over those increased costs. When their conversion rate at their current price is only 3%, there’s not exactly an obvious appetite to pay more for the same thing.

        And that’s the headline name. The key driver of the industry. And the numbers are just as bad everywhere else you look, either terrible, or deliberately obfuscated (remember, these companies sank billions of capex into this; if sales were good they’d be talking very openly and clearly about just how good they are).

  • UraniumBlazer@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 days ago

    I have no idea how people can consider this to be a hype bubble especially after the o3 release. It smashed the ARC AGI benchmark on the performance front. It ranks as the 175th best competitive coder in the world on Codeforces’ leaderboard.

    o3 proved that it is possible to have at least an expert AGI if not a Virtuoso AGI (according to Deep mind’s definition of AGI). Sure, it’s not economical yet. But it will get there very soon (just like how the earlier GPTs were a lot dumber and took a lot more energy than the newer, smaller parameter models).

    Please remember - fight to seize the means of production. Do not fight the means of production themselves.

    • r4venw@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      Where, in that position piece, do they mention o3? Who “proved” this?

      Additionally, I’m pretty sure that this “ARC AGI” benchmark is not using the same definition of AGI that you linked to by DeepMind. Conflating them is misleading. There is already so much misinformation out there about “AI”, don’t add to it.

      Lastly, I struggle to take at face value essays written by for-profit companies claiming they have AGI (that DeepMind paper links to OpenAI essays). They only stand to gain monetarily by claiming that their AI is an AGI (to be clear, this is an opinion; I do not have evidence to suggest that OpenAI is being disingenuous).

    • SupraMario@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      Why is it getting an AGI stamp now? I was under the impression humanity has not delivered a sentient AI? Which is what the AGI title was supposed to be used for…has that been pulled back again?

      • Communist@lemmy.frozeninferno.xyz
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        3 days ago

        Agi has nothing to do with sentience, which cannot be measured, openai, I think validly, defines it as a system that can do all intellectual labor.

          • Communist@lemmy.frozeninferno.xyz
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            3 days ago

            It was never about sentience, sentience is a meaningless, unmeasurable term.

            It’s a question of if it can replace humans in the workforce.

            artificial general intelligence means it’s able to generalize its intelligence, not sentience at all.

    • dustyData@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      Unless we invent cold fusion between the next 5 years, they will never be economical. They are the most energy inefficient thing ever invented by humanity and all prediction models state that it will cost more energy, not less, to keep making them better. They will never be energy efficient nor economical in their current state, and most companies are out of ideas on how to shake it up. Even the people who created generative models agree that they have just been brute forcing by making the models larger with more energy consumption. When you try to make them smaller or more energy efficient, they fall off the performance cliff and only produce garbage. I’m sure there are researchers doing cool stuff, but it is neither economical nor efficient.

      • Richard@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 days ago

        Untrue. There are small models that produce better output than the previous “flagships” like GPT-2. Also, you can achieve much more than we currently do with far less energy by working on novel, specialised hardware (neuromorphic computing).

    • Omega_Jimes@lemmy.ca
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      o3 made the high score on ARC through brute force, not by being good. To raise the score from 75% to 87% required 175 times more computing power, but exactly stunning returns.

      • Communist@lemmy.frozeninferno.xyz
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        3 days ago

        Why does it matter?

        If it can through brute force, it can do it. That’s the first step towards true agi, nobody said the first AGI would be economical, this feels like a major goalpost shift if you’re acknowledging it can do it at all, isn’t that insane?

        A little bit ago, everyone would’ve been saying this will never happen, that there was a natural wall simply because all it does is predict the next token, it’s been like, a few years of llm’s and they’re already getting this insane. We’re going to have AGI soon, it might not be a transformer, but billions upon billions of dollars are being thrown at this problem, there are people smart enough in the world to make this work, and this is the earliest sign that it’s coming.

        • Omega_Jimes@lemmy.ca
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 days ago

          I’m not convinced that it’s anywhere near an AGI, I’m convinced after combing through papers and code, that it’s an amazing parlor trick.

          I’d love to be proven wrong, but everything I’ve seen and everything I’ve used in my studies ( using DNN to simulate neurodivergence and spinal disgenesis, which is kinda AI adjacent) leads me to believe that the current part won’t lead to anything but convincing parlor tricks.

          The argument could be made that if a trick is convincing enough, does it matter if it’s intelligent or not.

            • Omega_Jimes@lemmy.ca
              link
              fedilink
              English
              arrow-up
              0
              ·
              3 days ago

              I’m not entirely sure.
              A non-probabilistic algorithm, probably. Something that didn’t rely on the liklihood of association, and instead was capable of context and rationality.
              Something that wouldn’t have a system capable of saying “Put glue on your pizza” because it would know that’s a silly thing to say to a human. A system that, when asked "Whats a good caustic detergent " wouldn’t be able to respond "Any good caustic detergent is a good caustic detergent " because duh. Something that doesn’t require thousands of hours of training to update and instead is capable of ingesting and rationalize new information on the fly.

              • Communist@lemmy.frozeninferno.xyz
                link
                fedilink
                English
                arrow-up
                0
                ·
                edit-2
                3 days ago

                You’re probablistic, you just have an internal verifier, you think things that are silly, and then decide not to say them all the time. A human being often thinks things that they realize are silly before they say them… that’s an entirely unfair goal in the first place from my perspective, why does it have to be non-probablistic?

                Are you not a general intelligence because sometimes your brain thinks silly things?

                o3 currently works precisely that way, by the way, it generates hundreds of possible things, and then uses something that checks if the steps actually work, before it outputs. In fact, they then reinforce it on these correct logical steps, so it becomes better at not outputting illogical answers like you said.

                it’s interesting that you said “not on the probability of the next word, but on context and rationality”

                context IS pricesely that, you know what’s likely to come next because of the context, that’s you understanding context. YOU as a human being don’t even always get this right, you must realize we are not perfect beings, we think of possibilities and choose the right one. I think we’re much better at this right now, but i don’t think that’s a fundamental difference between us and o3.

                Rationality is the internal verifier.

                Something that doesn’t require thousands of hours of training to update and instead is capable of ingesting and rationalize new information on the fly.

                Being able to do this is… exactly what arc-agi was testing. Literally the entire point of the benchmark, it can do that.

                I’ve done the test by the way, I solved it by brute forcing possible solutions in my head, then checking if they were true… did you just divine the answers instantly?

    • Voroxpete@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      It’s a bubble because OpenAI spend $2.35 for every $1.00 they make. Yes, you’re mathing right, that is a net loss.

      It’s a bubble because all of the big players in AI development agree that future models will cost exponentially more money to train, for incremental gains. That means there is no path forward that doesn’t intensely amplify the unprofitability of an already deeply unprofitable industry.

      It’s a bubble because newer models with better capabilities only cost more and more to run.

      It’s a bubble because as far as anyone knows there will never be a solution to the hallucination problem.

      It’s a bubble because despite investments treating it as a trillion dollar industry, no one has yet figured out a trillion dollar problem that AI can solve.

      You’re trying on a new top of the line VR headset and saying “Wow, this is incredible, how can anyone say this is a bubble?” Its not about how cool the tech is in isolation, it’s about its potential to effect widespread change. Facebook went in hard on VR, imagining a future where everyone worked from home while wearing VR headsets. But what they got was an expensive toy that only had niche uses.

      AI performs do well on certain coding tasks because a lot of the individual problems that make up a particular piece of software have already been solved. It’s standard practice to design programs as individual units, each of which performs the smallest task possible, and which can then be assembled to complete more complex tasks. This fits very well into the LLM model of assembling pieces into their most likely expected configurations. But it cannot create truly novel code, except by a kind of trial and error mutation process. It cannot problem solve. It cannot identify a users needs and come up with ideal solutions to them. It cannot innovate.

      This means that, at best, genAI in the software world becomes a tool for producing individual code elements, guided and shepherded by experienced programmers. It does not replace the software industry, merely augments it, and it does so at a cost that many companies simply may not feel is worth paying.

      And that’s its best case scenario. In every other industry AI has been a spectacular failure. But it’s being invested in as if it will be a technological reckoning for every form of intellectual labour on earth. That is the absolute definition of a bubble.

  • Zos_Kia@lemmynsfw.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 days ago

    To have a bubble you need companies with no clear path to monetization, being over-valued to an extreme degree. This leaves me wondering : what company specifically ? Are they talking about nVidia ? OpenAI ? MidJourney ? Or the slew of LLM-powered SaaS products that have started appearing ? How exactly are we defining “over-valuation” here ? Are we talking about the tech industry as a whole ?

    We often invite the comparison to the DotCom bubble but that’s apples to oranges. You had companies making social networks for dogs or similar bullshit, valued in the billions and getting a ticker at the stock market before making a single dime. Or companies with outlandish promises such as delivering to any home in the US, in <1 hour, for a low price, and building warehouses by the hundreds before having a storefront. What would be the 2024 equivalent ? If a bubble is about to deflate then there should be dozens of comparable examples.

    • UraniumBlazer@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      Exactly. There’s a very clear path to monetisation for the bigger tech companies (ofc, not the random startup that screams “AI quantum computing blockchain reeeee”).

      Lemmy is just incredibly biased against AI, as it could replace a shit ton of jobs and lead to a crazy amount of wealth inequality. However, people need to remember that the problem isn’t the tech- it’s the system that the tech is being innovated in.

      Denying AI is just going to make this issue a lot worse. We need to work to make AI be beneficial for all of us instead of the capitalists. But somehow leftist talk surrounding AI has just been about hating on it/ denying it, instead of preparing for a world in which it would be critical infrastructure very soon.

      • sudneo@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 days ago

        What job could possibly replace…? If you can replace a job with LLMs it means either that the job is not needed on the first place (bullshit job) or that you can replace it with a dice (e.g., decision-making processes), since LLMs-output will depend essentially just on what is in the training material -which we don’t know (I.e., the answer is essentially random).

      • Zos_Kia@lemmynsfw.com
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 days ago

        I don’t think it’s just Lemmy, i had similar conversations on Reddit. People don’t realize that the companies they claim are over-valued actually have very strong business fundamentals. That’s why in articles like OP’s they will never mention any names or figures. I guess it’s very convincing for outsiders but it doesn’t stand any amount of scrutiny.

        If you take OpenAI for example, they went from 0 to 3.6B$ annual revenue in just two fucking years. How is that not worth a boatload of money ? Even Uber didn’t have that kind of growth and they burned a LOT more cash than OpenAI is burning right now.

        As for the “AI quantum computing blockchain reeeee” projects… well they have a very hard time raising money right now and when they do, it’s at pretty modest valuations. The market is not as dumb as it is portrayed.

        • Voroxpete@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          3 days ago

          How is that not worth a boatload of money?

          Because they spend $2.35 billion in operating costs for every $1 billion in revenue (gross, not net).

          OpenAI loses money at an absolutely staggering rate, and every indication, even their own openly stated predictions, are that those costs will only increase.

          And keep in mind, right now OpenAI gets a lot of their investment in the form of compute credits from Microsoft, which they get to spend at a massively discounted rate. That means that if they were actually buying their Azure time at market value they’d be closer to spending something like $5bn to make $1bn.

          Again, I really need to be clear here, I’m not saying “to make 1 billion in profit.” I’m saying “revenue”. They lose money every time someone uses their services. The more users they have, the more their losses grow. Even paid users cost them more money than they pay in most cases.

          This is like a store that buys products at $10 and sells them at $4. It is the most insanely unprofitable business plan imaginable.

          And it’s not getting better. Conversions to paid plans are at about 3%. Their enterprise sales are abysmal. Training costs are increasing exponentially with each new generation of models. Attempts to make their models more compute efficient have so far failed utterly.

          OpenAI’s path to profitability is basically “Invent true AGI.” It’s a wild fantasy with zero basis in reality that investors are shovelling money into because investors will shovel money into anything that promises infinite growth.

          • Zos_Kia@lemmynsfw.com
            link
            fedilink
            English
            arrow-up
            0
            ·
            3 days ago

            Because they spend $2.35 billion in operating costs for every $1 billion in revenue

            This is entirely true but also completely normal for a hyperscaler in its first years. At this stage demonstrating a huge demand and your capacity to capture a lot of it is much more important than profitability. You don’t exactly bootstrap a company at this level of CapEx.

            Their revenue is still growing at a staggering rate, showing no signs of slowing down, and enterprise sales are pretty respectable. I don’t know that i’d call 3/4 of a billion in enterprise sales abysmal for a startup in its 2nd year. YMMV i guess.

            Sure there is some uncertainty about their model but that’s what VC backing is for, right ? They’re not building tin can factories with known and predictable business trends, and being valued at 40x your yearly revenue (not profit !) is pretty banal for a successful early stage Deep Tech. We may personally think it’s bullshit and choose not to invest in it but it’s still far from outrageous and very far from the definition of a bubble.

            • Voroxpete@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              0
              ·
              3 days ago

              For the level of investment in and hype around this company? Yes, those enterprise sales are abysmal. When there are major news articles about their product every single week, they should be doing a lot better than that.

              They have demonstrated zero ability at actually “hyperscale”. They have no path to getting those costs down. Their conversion rate from free to paid users is atrocious, and they’re already raising prices on their plans which is only going to worsen those conversion rates. Their costs to build future models are ballooning exponentially, and theres a decent chance that at some point Microsoft will get sick of subsidizing their compute costs.

              Is it possible that they could be successful? Yes. But a lottery ticket would probably be a sounder investment.

              For the record, OpenAI themselves are telling those VCs that they should think of their investment more as a “donation” with no expectation of future profit. Absolutely oozing confidence there.

              • someacnt@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                0
                ·
                edit-2
                3 days ago

                Hypergrowth cult is always “interesting”. Because it happened with X, apparently it will happen with Y. Then, unavoidably, it goes through the boom and bust cycle. I am sick of it.

              • Zos_Kia@lemmynsfw.com
                link
                fedilink
                English
                arrow-up
                0
                ·
                3 days ago

                For the level of investment in and hype around this company? Yes, those enterprise sales are abysmal

                I don’t see the connection. How are enterprise sales specifically relevant here ? Are enterprise customers known for jumping on top of early stage products where you’re from ? Cause where i’m from they’re known for being the last ones to board.

                They have demonstrated zero ability at actually “hyperscale”.

                How would you define hyperscale ? They have one of the biggest GPU fleets around and are likely serving trillions of tokens monthly. That falls well within the range of my personal definition.

                They have no path to getting those costs down. Their conversion rate from free to paid users is atrocious, and they’re already raising prices on their plans which is only going to worsen those conversion rates

                That’s just stuff you say. Atrocious (in your opinion), no path to getting those costs down (in your opinion). Alright, we get it, that’s not a company you’d invest in, but then again your investment thesis seems pretty conservative. If a company has to make billions of enterprise sales in its 2nd year, and have double digits conversion early on, then there’s not that many successful companies you would have invested in. You certainly wouldn’t have put a dime in Uber at 48B valuation 7 years ago - well those who did made a nice return on their investment.

                Is it possible that they could be successful? Yes. But a lottery ticket would probably be a sounder investment.

                Isn’t that the definition of VC-backed startups ? The alternative would be to build a time machine, travel back to the 18th century, and invest in the British textile industry. Sadly they don’t make this kind of predictible, risk-free and quickly profitable enterprises nowadays.

                Absolutely oozing confidence there

                Oh i won’t be the one to contradict you here. Sama is one sleazy motherfucker, that’s just written on his face. Sadly it doesn’t preclude him from building a historical hyperscaler with OpenAI.

  • LenielJerron@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    3 days ago

    A big issue that a lot of these tech companies seem to have is that they don’t understand what people want; they come up with an idea and then shove it into everything. There are services that I have actively stopped using because they started cramming AI into things; for example I stopped dual-booting with Windows and became Linux-only.

    AI is legitimately interesting technology which definitely has specialized use-cases, e.g. sorting large amounts of data, or optimizing strategies within highly restrained circumstances (like chess or go). However, 99% of what people are pushing with AI these days as a member of the general public just seems like garbage; bad art and bad translations and incorrect answers to questions.

    I do not understand all the hype around AI. I can understand the danger; people who don’t see that it’s bad are using it in place of people who know how to do things. But in my teaching for example I’ve never had any issues with students cheating using ChatGPT; I semi-regularly run the problems I assign through ChatGPT and it gets enough of them wrong that I can’t imagine any student would be inclined to use ChatGPT to cheat multiple times after their grade the first time comes in. (In this sense, it’s actually impressive technology - we’ve had computers that can do advanced math highly accurately for a while, but we’ve finally developed one that’s worse at math than the average undergrad in a gen-ed class!)

    • PlexSheep@infosec.pub
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 days ago

      I understand some of the hype. LLMs are pretty amazing nowadays (though closedai is unethical af so don’t use them).

      I need to program complex cryptography code for university. Claude sonnet 3.5 solves some of the challenges instantly.

      And it’s not trivial stuff, but things like “how do I divide polynomials, where each coefficient of that polynomial is an element of GF(2^128).” Given the context (my source code), it adds it seamlessly, writes unit tests, and it just works. (That is important for AES-GCM, the thing TLS relies on most of the time .)

      Besides that, LLMs are good at what I call moving words around. Writing cute little short stories in fictional worlds given some info material, or checking for spelling, or re-formulating a message into a very diplomatic nice message, so on.

      On the other side, it’s often complete BS shoehorning LLMs into things, because “AI cool word line go up”.

    • Brodysseus@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      I’ve ran some college hw through 4o just to see and it’s remarkably good at generating proofs for math and algorithms. Sometimes it’s not quite right but usually on the right track to get started.

      In some of the busier classes I’m almost certain students do this because my hw grades would be lower than the mean and my exam grades would be well above the mean.

    • Voroxpete@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 days ago

      The answer is that it’s all about “growth”. The fetishization of shareholders has reached its logical conclusion, and now the only value companies have is in growth. Not profit, not stability, not a reliable customer base or a product people will want. The only thing that matters is if you can make your share price increase faster than the interest on a bond (which is pretty high right now).

      To make share price go up like that, you have to do one of two things; show that you’re bringing in new customers, or show that you can make your existing customers pay more.

      For the big tech companies, there are no new customers left. The whole planet is online. Everyone who wants to use their services is using their services. So they have to find new things to sell instead.

      And that’s what “AI” looked like it was going to be. LLMs burst onto the scene promising to replace entire industries, entire workforces. Huge new opportunities for growth. Lacking anything else, big tech went in HARD on this, throwing untold billions at partnerships, acquisitions, and infrastructure.

      And now they have to show investors that it was worth it. Which means they have to produce metrics that show people are paying for, or might pay for, AI flavoured products. That’s why they’re shoving it into everything they can. If they put AI in notepad then they can claim that every time you open notepad you’re “engaging” with one of their AI products. If they put Recall on your PC, every Windows user becomes an AI user. Google can now claim that every search is an AI interaction because of the bad summary that no one reads. The point is to show “engagement”, “interest”, which they can then use to promise that down the line huge piles of money will fall out of this pinata.

      The hype is all artificial. They need to hype these products so that people will pay attention to them, because they need to keep pretending that their massive investments got them in on the ground floor of a trillion dollar industry, and weren’t just them setting huge piles of money on fire.

      • rottingleaf@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 days ago

        The answer is that it’s all about “growth”. The fetishization of shareholders has reached its logical conclusion, and now the only value companies have is in growth. Not profit, not stability, not a reliable customer base or a product people will want. The only thing that matters is if you can make your share price increase faster than the interest on a bond (which is pretty high right now).

        As you can see, this can’t go on indefinitely. And also such unpleasantries are well known after every huge technological revolution. Every time eventually resolved, and not in favor of those on the quick buck train.

        It’s still not a dead end. The cycle of birth, growth, old age, death, rebirth from the ashes and so on still works. It’s only the competitive, evolutionary, “fast” model has been killed - temporarily.

        These corporations will still die unless they make themselves effectively part of the state.

        BTW, that’s what happened in Germany described by Marx, so despite my distaste for marxism, some of its core ideas may be locally applicable with the process we observe.

        It’s like a worldwide gold rush IMHO, but not even really worldwide. There are plenty of solutions to be developed and sold in developing countries in place of what fits Americans and Europeans and Chinese and so on, but doesn’t fit the rest. Markets are not exhausted for everyone. Just for these corporations because they are unable to evolve.

        Lacking anything else, big tech went in HARD on this, throwing untold billions at partnerships, acquisitions, and infrastructure.

        If only Sun survived till now, I feel they would have good days. What made them fail then would make them more profitable now. They were planning too far ahead probably, and were too careless with actually keeping the company afloat.

        My point is that Sun could, unlike these corporations, function as some kind of “the phone company”, or “the construction company”, etc. Basically what Microsoft pretended to be in the 00s. They were bad with choosing the right kind of hype, but good with having a comprehensive vision of computing. Except that vision and its relation to finances had schizoaffective traits.

        Same with DEC.

        The point is to show “engagement”, “interest”, which they can then use to promise that down the line huge piles of money will fall out of this pinata.

        Well. It’s not unprecedented for business opportunities to dry out. It’s actually normal. What’s more important, the investors supporting that are the dumber kind, and the investors investing in more real things are the smarter kind. So when these crash (for a few years hunger will probably become a real issue not just in developing countries when that happens), those preserving power will tend to be rather insightful people.

        • AA5B@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          2 days ago

          If only Sun survived till now, I feel they would have good days

          The problem is a lot of what Sun brought to the industry is now in the Linux arena. If Sun survived, would Linux have happened? With such a huge development infrastructure around Linux, would Sun really add value?

          I was a huge fan of Sun also, they revolutionized the industry far above their footprint. However their approach seemed more research or academic at times, and didn’t really work with their business model. Red Hat figured out a balance where they could develop opensource while making enough to support their business. The Linux world figured out a different balance where the industry is above and beyond individual companies and doesn’t require profit

          • rottingleaf@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 days ago

            The problem is a lot of what Sun brought to the industry is now in the Linux arena. If Sun survived, would Linux have happened? With such a huge development infrastructure around Linux, would Sun really add value?

            Linux is not better than Solaris. It was, however, circumstantially more affordable, more attractive, and more exciting than Solaris at the same time. They’ve made a lot of strategic mistakes, but those were in the context of having some vision.

            I mean this to say that the “huge development infrastructure around Linux” is bigger, but much less efficient than that of any of BSDs, and than that of Solaris in the past. Linux people back then would take pride in ability to assemble bigger resources, albeit with smaller efficiency, and call that “the cathedral vs the bazaar”, where Linux is the bazaar. Well, by now one can see that the bazaar approach make development costs bigger long-term.

            IMHO if Sun didn’t make those mistakes, Solaris would be the most prestigious Unix and Unix-like system, but those systems would be targeted by developers similarly. So Linux would be alive, but not much more or less popular than FreeBSD. I don’t think they’d need Solaris to defeat all other Unix systems. After all, in early 00s FreeBSD had SVR4 binary compatibility code, similarly to its Linux compatibility code, which is still there and widely used. Probably commercial software distributed in binaries would be compiled for that, but would run on all of them. Or maybe not.

            It’s hard to say.

            But this

            The Linux world figured out a different balance where the industry is above and beyond individual companies and doesn’t require profit

            is wrong, everything about Linux that keeps going now is very commercial. Maybe 10 years ago one could say it’s not all about profit.

            • AA5B@lemmy.world
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 days ago

              The point is the industry is not a profit driven entity, but has room for many profit driven entities.

              • rottingleaf@lemmy.world
                link
                fedilink
                English
                arrow-up
                0
                ·
                1 day ago

                That’s like saying your body is not a protein driven mechanism (cause there are many other things involved), but has room for proteins.

                If somebody tears out half of your internal organs, you die.

                If profit-driven companies stop participating in Linux, Linux dies. Today’s Linux. Linux of year 1999 wouldn’t.

                That’s how even gifts can be the needle to control you.

                I mean, why is this even a point of contention. BSDs played safe in terms of politics, Linux gambled by not considering the dangers. BSDs grew more slowly, Linux took the bank. But now Linux is confined by the decisions made back then. BSDs are more free.

      • MagicShel@lemmy.zip
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        3 days ago

        I know I’m an enthusiast, but can I just say I’m excited about NotebookLLM? I think it will be great for documenting application development. Having a shared notebook that knows the environment and configuration and architecture and standards for an application and can answer specific questions about it could be really useful.

        “AI Notepad” is really underselling it. I’m trying to load up massive Markdown documents to feed into NotebookLLM to try it out. I don’t know if it’ll work as well as I’m hoping because it takes time to put together enough information to be worthwhile in a format the AI can easily digest. But I’m hopeful.

        That’s not to take away from your point: the average person probably has little use for this, and wouldn’t want to put in the effort to make it worthwhile. But spending way too much time obsessing about nerd things is my calling.

        • drosophila@lemmy.blahaj.zone
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          3 days ago

          Being able to summarize and answer questions about a specific corpus of text was a use case I was excited for even knowing that LLMs can’t really answer general questions or logically reason.

          But if Google search summaries are any indication they can’t even do that. And I’m not just talking about the screenshots people post, this is my own experience with it.

          Maybe if you could run the LLM in an entirely different way such that you could enter a question and then it tells you which part of the source text statistically correlates the most with the words you typed; instead of trying to generate new text. That way in a worse case scenario it just points you to a part of the source text that’s irrelevant instead of giving you answers that are subtly wrong or misleading.

          Even then I’m not sure the huge computational requirements make it worth it over crtrl-f or a slightly more sophisticated search algorithm.

          • anomnom@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            2 days ago

            you could enter a question and then it tells you which part of the source text statistically correlates the most with the words you typed; instead of trying to generate new text. That way in a worse case scenario it just points you to a part of the source text that’s irrelevant instead of giving you answers that are subtly wrong or misleading.

            Isn’t this what the best search engines were doing before the AI summaries?

            The main problem now is the proliferation of AI “sources” that are really just keyword stuffed junk websites that take over the first page of search results. And that’s apparently a difficult or unprofitable problem for the search algorithms to solve.

          • Flying Squid@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            3 days ago

            Multiple times now, I’ve seen people post AI summaries of articles on Lemmy which miss out really, really important points.

          • MagicShel@lemmy.zip
            link
            fedilink
            English
            arrow-up
            0
            ·
            3 days ago

            Well an example of something I think it could solve would be: “I’m trying to set this application up to run locally. I’m getting this error message. Here’s my configuration files. What is not set up correctly, or if that’s not clear, what steps can I take to provide more helpful information?”

            ChatGPT is always okay at that as long as you have everything set up according to the most common scenarios, but it tells you a lot of things that don’t apply or are wrong in the specific case. I would like to get answers that are informed by our specific setup instructions, security policies, design standards, etc. I don’t want to have to repeat “this is a Java spring boot application running on GCP integrating with redis on docker… blah blah blah”.

            I can’t say whether it’s worth it yet, but I’m hopeful. I might do the same with ChatGPT and custom GPTs, but since I use my personal account for that, it’s on very shaky ground to upload company files to sobering like that, and I couldn’t share with the team anyway. It’s great to ask questions that don’t require specific knowledge, but I think I’d be violating company policy to upload anything.

            We are encouraged to use NotebookLLM, however.

        • Voroxpete@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 days ago

          From a nerdy perspective, LLMs are actually very cool. The problem is that they’re grotesquely inefficient. That means that, practically speaking, whatever cool use you come up with for them has to work in one of two ways; either a user runs it themselves, typically very slowly or on a pretty powerful computer, or it runs as a cloud service, in which case that cloud service has to figure out how to be profitable.

          Right now we’re not being exposed to the true cost of these models. Everyone is in the “give it out cheap / free to get people hooked” stage. Once the bill comes due, very few of these projects will be cool enough to justify their costs.

          Like, would you pay $50/month for NotebookLM? However good it is, I’m guessing it’s probably not that good. Maybe it is. Maybe that’s a reasonable price to you. It’s probably not a reasonable price to enough people to sustain serious development on it.

          That’s the problem. LLMs are cool, but mostly in a “Hey this is kind of neat” way. They do things that are useful, but not essential, but they do so at an operating cost that only works for things that are essential. You can’t run them on fun money, but you can’t make a convincing case for selling them at serious money.

          • MagicShel@lemmy.zip
            link
            fedilink
            English
            arrow-up
            0
            ·
            3 days ago

            Totally agree. It comes down to how often is this thing efficient for me if I pay the true cost. At work, yes it would save over $50/mo if it works well. At home it would be difficult to justify that cost, but I’d also use it less so the cost could be lower. I currently pay $50/mo between ChatGPT and NovelAI (and the latter doen’t operate at a loss) so it’s worth a bit to me just to nerd out over it. It certainly doesn’t save me money except in the sense that it’s time and money I don’t spend on some other endeavor.

            My old video card is painfully slow for local LLM, but I dream of spending for a big card that runs closer to cloud speeds even if the quality is lower, for easier tasks.

            • xavier666@lemm.ee
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 days ago

              but I dream of spending for a big card that runs closer to cloud speeds

              Nvidia’s new motto: “An A100 at every home”

        • FarceOfWill@infosec.pub
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 days ago

          You’re using the wrong tool.

          Hell, notepad is the wrong tool for every use case, it exists in case you’ve broken things so thoroughly on windows that you need to edit a file to fix it. It’s the text editor of last resort, a dumb simple file editor always there when you need it.

          Adding any feature (except possibly a hex editor) makes it worse at its only job.

          • MagicShel@lemmy.zip
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            3 days ago

            … I don’t use Notepad. For anything. Hell, I don’t even use Windows.

            Not sure where the wires got crossed here.

            • FarceOfWill@infosec.pub
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 days ago

              Yes as others said, the op mentioned notepad and you said notebookllm.

              I thought you were talking about notepad and it’s new ai features.

              • MagicShel@lemmy.zip
                link
                fedilink
                English
                arrow-up
                0
                ·
                2 days ago

                I had no idea notepad + AI was a thing. It sounds farcical, so I assumed wrongly it was a reference to NotebookLLM. My mistake. I shouldn’t have assumed OP was just being dismissive.

            • TJA!@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              0
              ·
              3 days ago

              Then either you replied with your first post to the wrong post or you misread “windows putting AI into notepad” as notebookLLM? Because if not there is nothing obvious connecting your post to the parent