• hendrik@palaver.p3x.de
    link
    fedilink
    English
    arrow-up
    0
    ·
    10 hours ago

    Keep in mind this isn’t about open-weight vs other AI models at all. This is about how training data can be collected and used.

    • Grimy@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      6 hours ago

      Because of the vast amount of data needed, there will be no competitive viable open source solution if half the data is kept in a walled garden.

      This is about open weights vs closed weights.

      • hendrik@palaver.p3x.de
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        2 hours ago

        I agree that we need open-source and emancipate ourselves. The main issue I see is: The entire approach doesn’t work. I’d like to give the internet as an example. It’s meant to be very open, connect everyone and enable them to share information freely. It is set up to be a level playing field… Now look what that leads to. Trillion dollar mega-corporations, privacy issues everywhere and big data silos. That’s what the approach promotes. I agree with the goal. But in my opinion the approach will turn out to lead to less open source and more control by rich companies. And that’s not what we want.

        Plus nobody even opens the walled gardes. Last time I looked, Reddit wanted money for data. Other big platforms aren’t open either. And there’s kind of a small war going on with the scrapers and crawlers and anti-measures. So it’s not as if it’s open as of now.

        • Grimy@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          1 hour ago

          A lot of our laws are indeed obsolete. I think the best solution would be to force copy left licenses on anything using public created data.

          But I’ll take the wild west we have now with no walls then any kind of copyright dystopia. Reddit did successfully sell it’s data to Google for 60 million. Right now, you can legally scrape anything you want off reddit, it is an open garden in every sense of the word (even if they dont like it). It’s a lot more legal then using pirated books, but Google still bet 60 million that copyright laws would swing broadly in their favor.

          I think it’s very foolhardy to even hint at a pro copyright stance right now. There is a very real chance of AI getting monopolized and this is how they will do it.

      • JcbAzPx@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        4 hours ago

        They haven’t dewalled the garden yet. The copyright infringement part of the case will continue.

    • bob_omb_battlefield@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      10 hours ago

      If you aren’t allowed to freely use data for training without a license, then the fear is that only large companies will own enough works or be able to afford licenses to train models.

      • hendrik@palaver.p3x.de
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        8 hours ago

        Yes. But then do something about it. Regulate the market. Or pass laws which address this. I don’t really see why we should do something like this then, it still kind of contributes to the problem as free reign still advantages big companies.

      • Nomad Scry@lemmy.sdf.org
        link
        fedilink
        English
        arrow-up
        0
        ·
        9 hours ago

        If they can just steal a creator’s work, how do they suppose creators will be able to afford continuing to be creators?

        Right. They think we have enough original works that the machines can just make any new creations.

        😠

        • Grimy@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          6 hours ago

          The companies like record studio who already own all the copyrights aren’t going to pay creators for something they already owned.

          All the data has already been signed away. People are really optimistic about an industry that has consistently fucked everyone they interact with for money.

        • MudMan@fedia.io
          link
          fedilink
          arrow-up
          0
          ·
          9 hours ago

          It is entirely possible that the entire construct of copyright just isn’t fit to regulate this and the “right to train” or to avoid training needs to be formulated separately.

          The maximalist, knee-jerk assumption that all AI training is copying is feeding into the interests of, ironically, a bunch of AI companies. That doesn’t mean that actual authors and artists don’t have an interest in regulating this space.

          The big takeaway, in my book, is copyright is finally broken beyond all usability. Let’s scrap it and start over with the media landscape we actually have, not the eighteenth century version of it.

          • hendrik@palaver.p3x.de
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            8 hours ago

            I’m fairly certain this is the correct answer here. Also there is a seperation between judicative and legislative. It’s the former which is involved, but we really need to bother the latter. It’s the only way, unless we want to use 18th century tools on the current situation.

          • Grimy@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            6 hours ago

            Yes precisely.

            I don’t see a situation where the actual content creators get paid.

            We either get open source ai, or we get closed ai where the big ai companies and copyright companies make bank.

            I think people are having huge knee jerk reactions and end up supporting companies like Disney, Universal Music and Google.