Given how Reddit now makes money by selling its data to AI companies, I was wondering how the situation is for the fediverse. Typically you can block AI crawlers using robot.txt (Verge reported about it recently: https://www.theverge.com/24067997/robots-txt-ai-text-file-web-crawlers-spiders). But this only works per domain/server, and the fediverse is about many different servers interacting with each other.

So if my kbin/lemmy or Mastodon server blocks OpenAI’s crawler via robot.txt, what does that even mean when people on other servers that don’t block this crawler are boosting me on Mastodon, or if I reply to their posts. I suspect unless all the servers I interact with block the same AI crawlers, I cannot prevent my posts from being used as AI training data?

  • originalucifer@moist.catsweat.com
    link
    fedilink
    arrow-up
    0
    ·
    9 months ago

    the fact that google has to pay for the data proves the walled garden you claim is public.

    the fediverse is public, by default. it publicly distributes information to other publicly accessible servers… by default.

    its public information on publicly accessible servers that are opt-out by default. publicly.

    im baffled how people can have some expectation of privacy in such a clearly defined public space.

    • cecep@fedia.ioOP
      link
      fedilink
      arrow-up
      0
      ·
      9 months ago

      You don’t need to explain to me how the Fediverse works and I never said I have any expectation of privacy. But generally speaking, you’re overlooking the fact that there always have been rules for what can, and cannot be done with information that is publicly available. Just because someone publicly posts his Facebook profile picture doesn’t mean it’s legal to use in an ad without permission, for example. People might break the rules, yes, but then they might face consequences, and that alone prevents many from breaking them in the first place. Not perfect, but better than nothing. And I’m saying we’re in a process where rules are being renegotiated when it comes to using public information for AI training

      • originalucifer@moist.catsweat.com
        link
        fedilink
        arrow-up
        0
        ·
        9 months ago

        fair points, but i still posit that its a waste of time to attempt to regulate what can be viewed anonymously.

        personally, i could not possibly care less about any of my data being ingested by ‘ai’. not a battle i care to fight, or even find worthy of fighting.

        • cecep@fedia.ioOP
          link
          fedilink
          arrow-up
          0
          ·
          9 months ago

          That’s fair, but I think if AI companies would be legally required to disclose the sources of their training data and if you make some successor to robot.txt legally binding as well (both is being discussed in the EU for example), at least the “bigger players” in the AI industry would respect the rules. Better than nothing