Popular iPad design app Procreate is coming out against generative AI, and has vowed never to introduce generative AI features into its products. The company said on its website that although machine learning is a “compelling technology with a lot of merit,” the current path that generative AI is on is wrong for its platform.

Procreate goes on to say that it’s not chasing a technology that is a threat to human creativity, even though this may make the company “seem at risk of being left behind.”

Procreate CEO James Cuda released an even stronger statement against the technology in a video posted to X on Monday.

  • Mossy Feathers (They/Them)@pawb.social
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    Wow, I’m actually kinda impressed. I’m not sure I’m 100% behind their stance, but it’s better than companies that blindly chase profits.

    Tbh I think generative AI can be used creatively and artistically, but knowing how to use generative AI doesn’t automatically make you creative or artistic. It’s like making someone paint a picture for you. Just making someone paint a picture for you doesn’t make you an artist, but an artist could say something by making someone paint for them. To put it another way, the AI element has to be more than just a means to an end; it has to justify itself somehow.

    “But normal artists don’t have to justify themselves!”

    You’re right! That’s because it’s assumed that the amount of time, effort and practice that is required to create art “manually” leads to the artist thinking deeply about their artwork before and during its creation; and 99% of the time, that’s completely true (the other 1% is “eye candy” like Kinkade; which is what AI is 99% of the time). Most people don’t understand this because they have never truly attempted to make “art”, however artists obsess over the details. You think that red truck in the bottom corner was “just there”? No, the artist probably put it there for a reason. Hell, the truck being red likely has a reason behind it. Maybe the artist wanted to say something about red trucks, or maybe the truck just looked better in red. Either way, that was a decision the artist was required to make.

    That said, AI can do some really cool stuff that would take humans years to reproduce, or would be extremely tedious and mind-numbing. A good example I recently came across is using AI to split music into stems or even into individual instruments. This makes it a lot easier for DJs, musicians and producers to get clean samples. It also makes it significantly easier for people to make custom tracks for Fuser (that’s how I found out about it).

    I guess what I’m trying to say is that I don’t think they should write-off AI entirely, but instead try and think of areas where AI would help artists. Maybe you use it to allow people to rescale their artwork without potentially having to redraw blurry lines. Maybe it’s AI that’s designed to separate photographs into individual pieces for the purpose of collages. Maybe it’s an AI designed to interpolate animation frames better than human-written algorithms. AI can do a lot of stuff other than just making eye candy.

    That said, I think rejecting generative AI entirely is better than blindly chasing the money, so good on you.

    • FaceDeer@fedia.io
      link
      fedilink
      arrow-up
      0
      ·
      1 month ago

      They’re chasing profit too, though. “Taking a stand” means they’re advertising, trying to differentiate themselves from their competitors and draw in people who hold anti-AI views.

      That will last until that segment of users becomes too small to be worth trying to base their business on.

      • mke@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        1 month ago

        Well, sounds great. I almost wish more companies would advertise to that market, really.

        It’s like… I know you’re lying, and I know you probably don’t actually care, but some of your competitors couldn’t even be bothered to do this much. Those companies thought shitting on things I care about to maximize profits was the better strategy. I’ll take that into consideration in my future decisions.

        And if the situation changes, if they turn around and go full in on generative AI, we’ll just have to consider that too. That’s life.

        Of course, I believe using alternatives that are more resistant to these kinds of market trends (community built software, perhaps?) would be ideal, but it’s not always an option.

    • pycorax@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      They specifically called out generative AI though. Stuff like separating photographs to individual pieces doesn’t require generative AI specifically. Machine learning models that fall into the general umbrella of AI already exist for object segmentation.

  • li10@feddit.uk
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    Ironically, I think AI may prove to be most useful in video games.

    Not to outright replace writers, but so they instead focus on feeding backstory to AI so it essentially becomes the characters they’ve created.

    I just think it’s going to be inevitable and the only possible option for a game where the player truly chooses the story.

    I just can’t be interested in multiple choice games where you know that your choice doesn’t matter. If a character dies from option a, then option b, c, and d kill them as well.

    Realising that as a kid instantly ruined telltale games for me, but I think AI used in the right way could solve that problem, to at least some degree.

    • schizo@forum.uncomfortable.business
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      Something like using a LLM to make actually unique side quests in a Skyrim-esque game could be interesting.

      The side quest/bounty quest shit in something like Starfield was fucking awful because it was like, 5 of the same damn things. Something capable of making at least unique sounding quests would be a shockingly good use of the tech.

    • brucethemoose@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      1 month ago

      Yeah, ultimately a lof of devs are trying to make “story generators” relying on the user’s imagination to fill in the blanks, hence rimworld is so popular.

      There’s a business/technical model where “local” llms would kinda work for this too, if you set it up like the Kobold Horde. So the dev hosts a few GPU instances for GPUs that can’t handle the local LLM, but users with beefy PCs also generate responses for other users (optionally, with a low priority) in a self hosted horde.

    • Mossy Feathers (They/Them)@pawb.social
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      Honestly, I think that…

      • AI is going to revolutionize the game industry.

      • AI is going to kill the game industry as it currently exists.

      • Generative AI will lead to a lot of real-time effects and mechanics that are currently impossible, like endless quests that don’t feel hollow, realistic procedural generation that can convincingly create everything from random clutter to entire galaxies, true photorealistic graphics (look up gaussian splatting, it’s pretty cool), convincing real-time art filters (imagine a 3d game that looks like an animated Van Gogh painting), and so on.

      • Generative AI is going to result in a hell of a lot of layoffs and will likely ruin people’s lives.

      • Generative AI will eventually open the door to small groups of devs being able to compete with AAA releases on all metrics.

      • Generative AI will make studios with thousands of employees obsolete. This is a double-edged sword. Fewer employees means fewer ideas; but on the other side, you get a more accurate vision of what the director originally intended. Fewer employees also will also mean that you will likely have to be a genuinely creative person to get ahead, instead of someone who knows how to use Maya or Photoshop but is otherwise creatively bankrupt. Your contribution matters far more in a studio of <50 than it does in a studio of >5,000; as such, your creative skill will matter more.

      • A lot of people will have to be retrained because they will no longer be creative enough to make a living off of making games.

      Tbh, I think game development is one of the few places that generative AI will actually have a significant benefit; however I also think it will completely scramble the industry once it starts being widely adopted, and it’ll be a long time before the dust settles.

      • mke@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 month ago

        I’ve no idea where you’re getting these predictions from. I think some of them are fundamentally flawed if not outright incorrect, and don’t reflect real life trends of generative AI development and applications.

        Gonna finish this comment in a few, please wait.

        • Mossy Feathers (They/Them)@pawb.social
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 month ago

          I think the big difference is that you seem to think that AI has peaked or is near its peak potential, while I think AI is still just getting started. Will generative AI ever progress beyond being a gimmick? I don’t know, but I suspect it will eventually.

          For example, indies do not have the budget to license expensive actors (e.g. Call of Duty, Cyberpunk 2077), brands (e.g. racing games), and so on. GenAI will not change this. Hell, GenAI will certainly not pay for global advertising.

          Admittedly I had not thought about the licensing and advertising aspect. That’s a bit of a blind spot for me because it’s not something I tend to care about. You’re correct there.

          If hiring them is no longer advantageous due to financial incentives to adopt AI, that’s not their fault for being insufficiently creative.

          I mean, maybe I could have phrased it better, but what else are you gonna do? They have to make a living somehow and if they can’t get hired in the game industry anymore, you gotta help them find somewhere else they can work.

          • mke@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            1 month ago

            I think the big difference is that you seem to think that AI has peaked or is near its peak potential, while I think AI is still just getting started.

            That’s a fair assessment. I’m still not sure if popular AI tech is on an exponential or a sigmoid curve, but I tend towards the latter. Note, however, that the industry at large is starting to believe it’s just not worth it. Even worse, the entities at the forefront of AI are unsustainable—they’re burning brightly right now, but the cash flow required to keep a reaction on this scale going is simply too large. If you’ve got time and are willing, please check the linked article by Ed (burst damage).

            I mean, maybe I could have phrased it better, but what else are you gonna do?

            My bad, I try to trim down the fat while editing, but I accidentally removed things I shouldn’t. As I said, it’s a nitpick, and I understand the importance of helping those who find themselves unhirable. Maybe it’s just me, but I thought it came across a little mean, even if it wasn’t your intent. I try to gently “poke” folks when I see stuff like this because artists get enough undeserved crap already.

            • Mossy Feathers (They/Them)@pawb.social
              link
              fedilink
              English
              arrow-up
              0
              ·
              1 month ago

              Yeaaahh… artists are really hard on themselves too. I make art myself, albeit not professionally; and I tend to be extremely hard on myself. I think maybe some of my comments about creativity are the result of my own self-negativity bleeding through.

  • RobotToaster@mander.xyz
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    1 month ago

    Didn’t krita say the same thing at one time?

    It’s currently one of the best programs to generate AI art using self hosted models.

    • Mossy Feathers (They/Them)@pawb.social
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      I think that’s kinda comparing apples to oranges. Krita is FOSS, and FOSS developers can be just as affected by community pressure as proprietary developers; possibly moreso. I dunno the circumstances around Krita’s decision to walk back and include AI, but I speculate it may have come from community pressure. Procreate isn’t FOSS so the community has a much harder time forcing their hand (the community can’t exactly fork the code and push everyone to migrate to a pro-AI version of Procreate). The other side of this, however, is that as proprietary developers, they feel more pressure from money.

      My prediction is that they’ll stick to this as long as it’s profitable. If they break away from it then it’s either because the CEO was replaced with a more profit-hungry CEO, they’re no longer profitable and they believe adding AI would fix that, or they believe they’ve found a use for AI that wouldn’t sacrifice creativity.

      • GreyBeard@lemmy.one
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 month ago

        Generate images with self hosted models, or integrate it with art programs? Because yes to both.

  • ReCursing@lemmings.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    a technology that is a threat to human creativity

    Perhaps the most stupid take on this subject I have seen. Nothing will stop humans creating, definitely not a new creative medium! That’s all it is, by the way, a new medium, like photography a hundred and some years ago, or digital painting more recently. Most of the same arguments were made against pre mixed paints - Turner was dragged for using them, for example!

    • Ibuthyr@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      It is problematic though. People start relying on content generation more and more and stop learning how to do it properly. Once they start relying on AI shit, that’s when capitalism does its thing and locks you into monthly subscription costs. Just look at what Adobe is doing. They create a dependency and then start changing their business model. Cloud this and cloud that is the same kind of problem.

      Plus, ai generated content often looks alike. You kind of take away signature looks of creators.

      I’m not entirely against AI generated content. A friend of mine hates social media but his small business relies on it. Most of his posts are ai generated just so he doesn’t have to deal with that cancer.

      • linearchaos@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 month ago

        It is problematic just like the current IP laws are non congruent with AI generation.

        But The cat’s out of the bag nothing’s going to stop it from happening.

        We’re going to need to figure out how to make it work

      • sunbeam60@lemmy.one
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 month ago

        The “they stop learning how to do it properly” is as old as time itself!

        How many of today’s Illlustrator artists know how to blend oil colours and layer them on cloth? How many software developers could build what they do in pure assembler?

        We stand on the shoulders of giants, have been since the Stone Age. Specialisation and advancement has meant we don’t need to start from first principle. You could argue that is what “progress” is; being able to get a little bit further because your parents got a little bit further because their parents got a little bit further.

        • Ibuthyr@discuss.tchncs.de
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          1 month ago

          That argument goes together with the whole money trap thing though. That’s my main problem. And people still blend colors digitally. Devs still program themselves and merely use ai as a starting point. They’d never trust ai generated code blindly. Entering a prompt to have something generated or using an ai brush just isn’t really the same as actually creating something. It’s literally letting a (cloud of) computer (s) create your art for you.

          Edit: dumbing down the internet and computers has lead to a colossal amount of imbeciles who aren’t able to do anything beyond scrolling on their phones. This is a huge problem as they’re being exploited by companies and politicians in a very efficient manner. All because people forgot how to use and troubleshoot a computer.

    • Mirodir@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      Does it? I worked on training a classifier and a generative model on freely available galaxy images taken by Hubble and labelled in a citizen science approach. Where’s the theft?

      • unsignedbit@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 month ago

        Hard to say. Training models is generative; training a model from scratch is costly. Your input may not infringe copyright but the input before or after may have.

        • Mirodir@discuss.tchncs.de
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 month ago

          I trained the generative models all from scratch. Pretrained models are not that helpful when it’s important to accurately capture very domain specific features.

          One of the classifiers I tried was based on zoobot with a custom head. Assuming the publications around zoobot are truthful, it was trained exclusively on similar data from a multitude of different sky surveys.

    • ribhu@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      That’s a blanket statement. While I understand the sentiment, what about the thousands of “AIs” trained on private, proprietary data for personal or private use by organizations that own the said data. It’s the not the technology but the lack of regulation and misaligned incentives.

    • BumpingFuglies@lemmy.zip
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      Can you explain how you came to that conclusion?

      The way I understand it, generative AI training is more like a single person analyzing art at impossibly fast speeds, then using said art as inspiration to create new art at impossibly fast speeds.

      • gap_betweenus@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 month ago

        With this logic photography is a painting, painted at an impossible high speed - but for some reasons we make a difference between something humans make and machines make.

        • ReCursing@lemmings.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 month ago

          Amusingly, every argument against ai art was made against photography over a hundred years ago, and I bet you own a camera - possibly even on the device you wrote your stupid comment on!

          • gap_betweenus@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            1 month ago

            Sure, I even do photography professionally form time to time - I just don’t consider it to be a painting.

              • gap_betweenus@lemmy.world
                link
                fedilink
                English
                arrow-up
                0
                ·
                30 days ago

                Photography can be art as well as AI generated images can be art as well. AI is a tool and people can create art with it. But also what is art is completely subjective to the viewer.

      • Clasm@ttrpg.network
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 month ago

        The art isn’t being made btw so much as being copy and pasted in a way that might convince you it was new.

        Since the AI cannot create a new style or genre on its own, without source material that already exists to train it, and that source material is often scraped up off of databases, often against the will and intent of the original creators, it is seen as theft.

        Especially if the artists were in no way compensated.

        • paw@feddit.org
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 month ago

          To add to your excellent comment:

          It does not ask if it can copy the art nor does it attribute its generated art with: “this art was inspired by …”

          I can understand why creators unhappy with this situation.

          • ReCursing@lemmings.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            1 month ago

            Do you go into a gallery and scream “THIS ART WAS INSPIRED BY PICASSO. WHY DOESN’T IT SAY THAT! tHIS IS THEFT!” - no, I suspect you don’t because that would be stupid. That’s what you sound like here

        • FatCrab@lemmy.one
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 month ago

          This is absolutely wrong about how something like SD generates outputs. Relationships between atomic parts of an image are encoded into the model from across all training inputs. There is no copying and pasting. Now whether you think extracting these relationships from images you can otherwise access constitutes some sort of theft is one thing, but characterizing generative models as copying and pasting scraped image pieces is just utterly incorrect.

          • Clasm@ttrpg.network
            link
            fedilink
            English
            arrow-up
            0
            ·
            1 month ago

            While, yes it is not copy and paste in the literal sense, it does still have the capacity to outright copy the style of an artist’s work that was used to train it.

            If teaching another artist’s work is already frowned upon when trying to pass the trace off as one’s own work, then there’s little difference when a computer does it more convincingly.

            Maybe a bit off tangent here, since I’m not even sure if this is strictly possible, but if a generative system was only trained off of, say, only Picasso’s work, would you be able to pass the outputs off as Picasso pieces? Or would they be considered the work of the person writing a prompt or built the AI? What if the artist wasn’t Picasso but someone still alive, would they get a cut of the profits?

            • FatCrab@lemmy.one
              link
              fedilink
              English
              arrow-up
              0
              ·
              1 month ago

              The outputs would be considered no one’s outputs as no copyright is afforded to AI general content.

              • Eccitaze@yiffit.net
                link
                fedilink
                English
                arrow-up
                0
                ·
                29 days ago

                That feels like it’s rather besides the point, innit? You’ve got AI companies showing off AI art and saying “look at what this model can do,” you’ve got entire communities on Lemmy and Reddit dedicated to posting AI art, and they’re all going “look at what I made with this AI, I’m so good at prompt engineering” as though they did all the work, and the millions of hours spent actually creating the art used to train the model gets no mention at all, much less any compensation or permission for their works to be used in the training. Sure does seem like people are passing AI art off as their own, even if they’re not claiming copyright.

                • FatCrab@lemmy.one
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  28 days ago

                  I’m not sure how it could be besides the point, though it may not be entirely dispositive. I take ownership to be a question of who has a controlling and exclusionary right to something–in this case thats copyright. Copyright allows you to license these things and extract money for their use. If there is no copyright, there is no secure monetization (something companies using AI generated materials absolutely keep high in mind). The question was “who would own it” and I think it’s pretty clear cut who would own it. No one.

    • TheGrandNagus@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      That’s a very reactionary take, IMO.

      There’s plenty of AI out there that’s not built on theft. You can train them solely on your own data if you want them to. There’s open source models out there trained only on data they were expressly given consent to use.

      You can get machine learning algorithms to learn how to play basic games completely on their own, etc.

  • demesisx@infosec.pub
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    Procreate is amazing. I bought it for my neurodivergent daughter and used it as a non-destructive coloring book.

    I’d grab a line drawing of a character that she wanted to color from a google image search, add it to the background layer, lock the background so she can’t accidentally move or erase it, then have her color on the layer above it using the multiply so the black lines can’t be painted over. She got the point where she prefers to have the colorized version alongside the black and white so she can grab the colors from the original and do fun stuff like mimic its shading and copy paste in elements that might have been too difficult for her to render. Honestly, she barely speaks but on that program, she’s better than most adults already even at age 8. Her work looks utterly perfect and she knows a lot of advanced blending and cloning stuff that traditional media artists don’t usually know.

  • mindlight@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    While a honorable move, “never” doesn’t exist in a world based on quarterly financials…

  • daniskarma@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    As with everything the problem is not AI technology the problem is capitalism.

    End capitalism and suddenly being able to share and openly use everyone’s work for free becomes a beautiful thing.

    • exocortex@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      1 month ago

      I agree, but as long as we still have capitalism I support measures that at least slow down the destructiveness of capitalism. AI is like a new powertool in capitalism’s arsenal to dismantle our humanity. Sure we can use it for cool things as well. But right now it’s used mostly to automate stuff that makes us human - art, music and so on. Not useful stuff like loading the dishwasher for me. More like writing a letter for me to invite my friends to my birthday. Very cool. But maybe the work I put in doing this myself is making my friends feel appreciated?

      Edit: It’s also nice to at least have an app that takes this maximalist approach. Then people can choose. If they’re half-assing it there will be more and more ai-features creeping in over time. One compromise after the next until it’s like all the other apps. It’s also important to have such a maximalist stand in order to gauge the scale in a way.

    • sunbeam60@lemmy.one
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      1 month ago

      I sort of agree. But…

      … In a world of adequate distribution and a form of universal income, we should all relish automation.

      That doesn’t preclude capitalism (investing for profit, the use of currency, interest rates etc), however, just needs a state with guts and capability to force redistribution.

    • Allero@lemmy.today
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      This, over and over again.

      Going against AI is being a luddite, not aware of the core underlying issue.

  • Aceticon@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    1 month ago

    No doubt his decision was helped by the fact that you can’t really fit full image generation AI on iPads - for example Stable Diffusion needs at the very least 6GB of GPU memory to work.

    That said, since what they sell is a design app, I applaud him for siding with the interests of at least some of his users.

    PS: Is it just me that finds it funny that the guy’s last name is “Cuda” and CUDA is the Nvidia technology for running computing on their GPUs and hence widelly used for this kind of AI?

    • TheRealKuni@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      you can’t really fit full image generation AI on iPads - for example Stable Diffusion needs at the very least 6GB of GPU memory to work.

      You can currently run Stable Diffusion and Flux on iPads and iPhones with the Draw Things app. Including LoRAs and TIs and ControlNet and a whole bunch of other options I’m too green to understand.

      Technically the app even runs on relatively old devices, though I imagine only at lower resolutions and probably takes ages.

      But in my limited experience it works quite well on an iPad Pro and an iPhone 13 Pro.

        • TheRealKuni@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 month ago

          Honestly most of what I’ve learned about how to use SD comes from seeing what other people have done and trying to tweak or adjust to get a feel for the tool and its various models. Spend some time on a site like CivitAI to both see what can be done and to find models. I’m very much a noob and cannot produce results nearly as impressive as a good chunk of what I find on there.

          The most important thing I’ve learned is how much generative AI, especially SD, is just a tool. And people with more creativity and a better understanding of the tool use it better, just like every other tool.

          I do like the idea of using it in GIMP as an answer to Adobe’s Firefly.

  • flop_leash_973@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    29 days ago

    That stance will change if they ever get acquired. Might even get the chance to see James Cuda try and walk back this stance in a few years.