AI has become as a deeply polarizing issue on the left, with many people having concerns regarding its reliance on unauthorized training data, displacement of workers, lack of creativity, and environmental costs. I’m going to argue that while these critiques warrant attention, they overlook the broader systemic context. As Marxists, our focus should not be on rejecting technological advancement but on challenging the capitalist framework that shapes its use. By reframing the debate, we can recognize AI’s potential as a tool for democratizing creativity and accelerating the contradictions inherent in capitalism.

Marxists have never opposed technological progress in principle. From the Industrial Revolution to the digital age, we have understood that technological shifts necessarily proletarianize labor by reshaping modes of production. AI is no exception. What distinguishes it is its capacity to automate aspects of cognitive and creative tasks such as writing, coding, and illustration that were once considered uniquely human. This disruption is neither unprecedented nor inherently negative. Automation under capitalism displaces workers, yes, but our critique must target the system that weaponizes progress against the workers as opposed to the tools themselves. Resisting AI on these grounds mistakes symptoms such as job loss for the root problem of capitalist exploitation.

Democratization Versus Corporate Capture

The ethical objection to AI training on copyrighted material holds superficial validity, but only within capitalism’s warped logic. Intellectual property laws exist to concentrate ownership and profit in the hands of corporations, not to protect individual artists. Disney’s ruthless copyright enforcement, for instance, sharply contrasts with its own history of mining public-domain stories. Meanwhile, OpenAI scraping data at scale, it exposes the hypocrisy of a system that privileges corporate IP hoarding over collective cultural wealth. Large corporations can ignore copyright without being held to account while regular people cannot. In practice, copyright helps capitalists far more than it help individual artists. Attacking AI for “theft” inadvertently legitimizes the very IP regimes that alienate artists from their work. Should a proletarian writer begrudge the use of their words to build a tool that, in better hands, could empower millions? The true conflict lies not in AI’s training methods but in who controls its outputs.

Open-source AI models, when decoupled from profit motives, democratize creativity in unprecedented ways. They enable a nurse to visualize a protest poster, a factory worker to draft a union newsletter, or a tenant to simulate rent-strike scenarios. This is no different from fanfiction writers reimagining Star Wars or street artists riffing on Warhol. It’s just collective culture remixing itself, as it always has. The threat arises when corporations monopolize these tools to replace paid labor with automated profit engines. But the paradox here is that boycotting AI in grassroots spaces does nothing to hinder corporate adoption. It only surrenders a potent tool to the enemy. Why deny ourselves the capacity to create, organize, and imagine more freely, while Amazon and Meta invest billions to weaponize that same capacity against us?

Opposing AI for its misuse under capitalism is both futile and counterproductive. Creativity critiques confuse corporate mass-production with the experimental joy of an individual sketching ideas via tools like Stable Diffusion. Our task is not to police personal use but to fight for collective ownership. We should demand public AI infrastructure to ensure that this technology is not hoarded by a handful of corporations. Surrendering it to capital ensures defeat while reclaiming it might just expand our arsenal for the fights ahead.

Creativity as Human Intent, Not Tool Output

The claim that AI “lacks creativity” misunderstands both technology and the nature of art itself. Creativity is not an inherent quality of tools — it is the product of human intention. A camera cannot compose a photograph; it is the photographer who chooses the angle, the light, the moment. Similarly, generative AI does not conjure ideas from the void. It is an instrument wielded by humans to translate their vision into reality. Debating whether AI is “creative” is as meaningless as debating whether a paintbrush dreams of landscapes. The tool is inert; the artist is alive.

AI has no more volition than a camera. When I photograph a bird in a park, the artistry does not lie in the shutter button I press or the aperture I adjust, but in the years I’ve spent honing my eye to recognize the interplay of light and shadow, anticipating the tilt of a wing, sensing the split-second harmony of motion and stillness. These are the skills that allow me to capture images such as this:

Hand my camera to a novice, and it is unlikely they would produce anything interesting with it. Generative AI operates the same way. Anyone can type “epic space battle” into a prompt, but without an understanding of color theory, narrative tension, or cultural symbolism, the result is generic noise. This is what we refer to as AI slop. The true labor resides in the human ability to curate and refine, to transform raw output into something resonant.

AI tools like ComfyUI are already being used by artists to collaborate and bring their visions to life, particularly for smaller studios. These tools streamline the workflow, allowing for a faster transition from the initial sketch to a polished final product. They also facilitate an iterative and dynamic creative process, encouraging experimentation and leading to unexpected, innovative results. Far from replacing artists, AI expands their creative potential, enabling smaller teams to tackle more ambitious projects.

People who attack gen AI on the grounds of it being “soulless” are recycling a tired pattern of gatekeeping. In the 1950s, programmers derided high-level languages like FORTRAN as “cheating,” insisting real coders wrote in assembly. They conflated suffering with sanctity, as if the drudgery of manual memory allocation were the essence of creativity. Today’s artists, threatened by AI, make the same error. Mastery of Photoshop brushes or oil paints is not what defines art, it’s a technical skill developed for a particular medium. What really matters is the capacity to communicate ideas and emotions through a medium. Tools evolve, and human expression adapts in response. When photography first emerged, painters declared mechanical reproduction the death of art. Instead, it birthed new forms such as surrealism, abstraction, cinema that expanded what art could be.

The real distinction between a camera and generative AI is one of scope, not substance. A camera captures the world as it exists while AI visualizes worlds that could be. Yet both require a human to decide what matters. When I shot my bird photograph, the camera did not choose the park, the species, or the composition. Likewise, AI doesn’t decide whether a cyberpunk cityscape should feel dystopian or whimsical. That intent, the infusion of meaning, is irreplaceably human. Automation doesn’t erase creativity, all it does is redistribute labor. Just as calculators freed mathematicians from drudgery of arithmetic, AI lowers technical barriers for artists, shifting the focus to concept and critique.

The real anxiety over AI art is about the balance of power. When institutions equate skill with specific tools such as oil paint, Python, DSLR cameras, they privilege those with the time and resources to master them. Generative AI, for all its flaws, democratizes access. A factory worker can now illustrate their memoir and a teenager in Lagos can prototype a comic. Does this mean every output is “art”? No more than every Instagram snapshot is a Cartier-Bresson. But gatekeepers have always weaponized “authenticity” to exclude newcomers. The camera did not kill art. Assembly lines did not kill craftsmanship. And AI will not kill creativity. What it exposes is that much of what we associate with production of art is rooted in specific technical skills.

Finally, the “efficiency” objection to AI collapses under its own short-termism. Consider that just a couple of years ago, running a state-of-the-art model required data center full of GPUs burning through kilowatts of power. Today, DeepSeek model runs on a consumer grade desktop using mere 200 watts of power. This trajectory is predictable. Hardware optimizations, quantization, and open-source breakthroughs have slashed computational demands exponentially.

Critics cherry-pick peak resource use during AI’s infancy. Meanwhile, AI’s energy footprint per output unit plummets year-over-year. Training GPT-3 in 2020 consumed ~1,300 MWh; by 2023, similar models achieved comparable performance with 90% less power. This progress is the natural arc of technological maturation. There is every reason to expect that these trends will continue into the future.

Open Source or Oligarchy

To oppose AI as a technology is to miss the forest for the trees. The most important question is who will control these tools going forward. No amount of ethical hand-wringing will halt development of this technology. Corporations will chase AI for the same reason 19th-century factory owners relentlessly chased steam engines. Automation allows companies to cut costs, break labor leverage, and centralize power. Left to corporations, AI will become another privatized weapon to crush worker autonomy. However, if it is developed in the open then it has the potential to be a democratized tool to expand collective creativity.

We’ve seen this story before. The internet began with promises of decentralization, only to be co-opted by monopolies like Google and Meta, who transformed open protocols into walled gardens of surveillance. AI now stands at the same crossroads. If those with ethical concerns about AI abandon the technology, its development will inevitably be left solely to those without such scruples. The result will be proprietary models locked behind corporate APIs that are censored to appease shareholders, priced beyond public reach, and designed solely for profit. It’s a future where Disney holds exclusive rights to generate “fairytale” imagery, and Amazon patents “dynamic storytelling” tools for its Prime franchises. This is the necessary outcome when technology remains under corporate control. Under capitalism, innovation always serves monopoly power as opposed to the interests of the public.

On the other hand, open-source AI offers a different path forward. Stable Diffusion’s leak in 2022 proved this: within months, artists, researchers, and collectives weaponized it for everything from union propaganda to indigenous language preservation. The technology itself is neutral, but its application becomes a tool of class warfare. To fight should be for public AI infrastructure, transparent models, community-driven training data, and worker-controlled governance. It’s a fight for the means of cultural production. Not because we naively believe in “neutral tech,” but because we know the alternative is feudalistic control.

The backlash against AI art often fixates on nostalgia for pre-digital craftsmanship. But romanticizing the struggle of “the starving artist” only plays into capitalist myths. Under feudalism, scribes lamented the printing press; under industrialization, weavers smashed looms. Today’s artists face the same crossroads: adapt or be crushed. Adaptation doesn’t mean surrender, it means figuring out ways to organize effectively. One example of this model in action was when Hollywood writers used collective bargaining to demand AI guardrails in their 2023 contracts.

Artists hold leverage that they can wield if they organize strategically along material lines. What if illustrators unionized to mandate human oversight in AI-assisted comics? What if musicians demanded royalties each time their style trains a model? It’s the same solidarity that forced studios to credit VFX artists after decades of erasure.

Moralizing about AI’s “soullessness” is a dead end. Capitalists don’t care about souls, they care about surplus value. Every worker co-op training its own model, every indie game studio bypassing proprietary tools, every worker using open AI tools to have their voice heard chips away at corporate control. It’s materialist task of redistributing power. Marx didn’t weep for the cottage industries steam engines destroyed. He advocated for socialization of the means of production. The goal of stopping AI is not a realistic one, but we can ensure its dividends flow to the many, not the few.

The oligarchs aren’t debating AI ethics, they’re investing billions to own and control this technology. Our choice is to cower in nostalgia or fight to have a stake in our future. Every open-source model trained, every worker collective formed, every contract renegotiated is a step forward. AI won’t be stopped any more than the printing press and the internet before it. The machines aren’t the enemy. The owners are.

    • ☆ Yσɠƚԋσʂ ☆@lemmygrad.mlOP
      link
      fedilink
      arrow-up
      2
      ·
      8 hours ago

      Exactly, the profession is becoming proletarianized, and the mode of production is changing. Traditional art isn’t going to go away as a result, but it’s not going to be the way illustration is produced in the industry. I expect this will have similar effect on other industries such as software development, writing, etc. Automation is encroaching on lot of the jobs that previously could only be done by a human previously. It doesn’t remove the need for human labor, but it does change the nature of the work significantly.

  • Muad'Dibber@lemmygrad.ml
    link
    fedilink
    arrow-up
    4
    ·
    1 day ago

    Just got around to reading this, excellent post. It’s unfortunate but necessary work to continually debunk the neo-luddites every time new technology gets introduced.

    You hit it on the head that these tools are neutral, and their mis-use under capitalism is not an argument to reject them wholesale.

    • ☆ Yσɠƚԋσʂ ☆@lemmygrad.mlOP
      link
      fedilink
      arrow-up
      1
      ·
      1 day ago

      Thanks, and I think the messaging we should focus on building alternatives. The AI boycott movement is peak liberal idealism. It’s the political equivalent of shaking a fist at the sky. Unhappy with the tech? Can’t imagine real action? Just vote it away! As if moral posturing ever stopped a single corporation.

      Real action involves building instead of pleading. Open development where artists, engineers, and communities co-create tools is the only path to wrest control from Silicon Valley’s black boxes. You don’t vote out exploitation, you have to put in work to replace it.

  • GreatSquare@lemmygrad.ml
    link
    fedilink
    arrow-up
    2
    ·
    3 days ago

    AI won’t be stopped any more than the printing press and the internet before it. The machines aren’t the enemy. The owners are.

    There’s a hype machine that is trying to shoehorn AI into places where it isn’t mature enough currently. Which is why the rush to market is a problem.

    The owners of the machine think the machine can do more than it is capable of. This leads to the enshittification of the product. It is the customer that suffers.

    In terms of art content, there’s already a tonne of it. WAY MORE than audiences even want to consume. If anything, AI is pouring more shit into an overflowing cup.

    The machine has been abused because the owner doesn’t understand where the value resides.

    In the meantime it’s burning a lot of fossil fuels to make shit.

  • Kras Mazov@lemmygrad.ml
    link
    fedilink
    English
    arrow-up
    1
    ·
    3 days ago

    Last time I discussed this with you didn’t really end well, but I genuinely think this is a pretty good post that I have some critiques of. I also want to clarify that some of the stuff you’re saying here was not made this clear in that previous discussion which led to some of our disagreements, which I’ll touch on briefly.

    That being said, I believe you aren’t weighting much the issues of the way current models are constructed. And I’m not talking about the “copyright issue” we see plastered everywhere when it comes to the current discourse surrounding AI, because this discourse wants to simply maintain and enforce copyright as a system. I’m talking about how artists should have a say in how their work is used and should be compensated for their labor if they are to be used in these training models.

    I can 100% see how this would not be an issue at all under a sufficiently advanced socialist society, since the way we would view and interact with our labor would be completely different than how it is under the current capitalist societies around the world. But the fact of the matter is that right now we live in these capitalist societies and these artists are being screwed over. If they don’t want their art to be part of the training model of these genAI tools, that should be enough for their work to not be used.

    I strongly believe that caving in to how it is done right now and simply using these models that were built on the unpaid labor of countless non consenting artists is a huge slap to their faces. We should instead be on their side and educate people about these issues, while also advocating and participating in the constructions of new models that do not have these issues. The clarification you made in this post shows that your stance on this seen to largely be the same as mine, which was not something I got from your previous comments on this.

    Also, as it is right now, workers are divided in a way where a lot of techbros and the general public just don’t give a shit about how this tech works or how it is constructed, and act actively antagonizing artists instead of uniting with them against these corpo controlled AIs.

    No amount of ethical hand-wringing will halt development of this technology.

    I don’t think anyone believes it will honestly. As far as the discourse on genAI goes, I haven’t seen many people that actually believe you can simply halt the development of the tech. What I have seen a lot of however is protesting against its use as it currently stands.

    The technology itself is neutral, but its application becomes a tool of class warfare. To fight should be for public AI infrastructure, transparent models, community-driven training data, and worker-controlled governance. It’s a fight for the means of cultural production. Not because we naively believe in “neutral tech,” but because we know the alternative is feudalistic control.

    This is a good point and I completely agree with your point here, but I have a question, isn’t the tech itself not neutral because of the conditions that resulted in it’s development? It was created for a political reason, and that is not inseparable from itself. To give an extreme example, nuclear bombs are not neutral just because it exists. It’s whole existence is political in nature. You can argue that it is a tool today, sure, but what uses would nuclear bombs have in a communist society?

    The backlash against AI art often fixates on nostalgia for pre-digital craftsmanship.

    Digital art exists, I don’t think it has anything to do with pre-digital craftsmanship, but rather there is a fixation on craftsmanship itself. Tho I wouldn’t say this is the only reason there is backlash against AI. The discourse surrounding it have in some ways advanced, and there are better arguments being made today.

    One thing I never see discussed, maybe because it’s still unknown territory, is the plateau of the tech. It can never completely replace non-generated art, as people will still learn the skill or their own reasons, but can it completely replace drawing/painting as a job? And even if it could, would it actually replace it? Also, would it still make sense to exist and be used on a society with a different mode of production where capital and profit are not the main motors of society?

    • ☆ Yσɠƚԋσʂ ☆@lemmygrad.mlOP
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      2 days ago

      Also, as it is right now, workers are divided in a way where a lot of techbros and the general public just don’t give a shit about how this tech works or how it is constructed, and act actively antagonizing artists instead of uniting with them against these corpo controlled AIs.

      Right, educating the public on how corporations rip artists off is important. However, messaging is also important. My impression is that outside niche circles like the Fediverse, the current messaging is simply not resonating with the public at large. It’s important to recognize when a particular approach is not achieving its goals, and to step back and think about how to improve it.

      I don’t think anyone believes it will honestly. As far as the discourse on genAI goes, I haven’t seen many people that actually believe you can simply halt the development of the tech. What I have seen a lot of however is protesting against its use as it currently stands.

      The approach of protesting against the use of the tech would work if there was a critical mass of people refusing to use it that would cut in the profit of the corporations and force them to change their approach. It does not appear that such a critical mass exists. I’d go as far as to argue that it’s a form of voting which stems from liberal model of political participation most people in the west are indoctrinated into. People are unhappy this tech exists, they can’t think of any meaningful action to take, and so they just want to vote it out of existence.

      On the other hand, there is a lot of support for the open development model, and that seems to be like a far more viable approach towards materially improving the current situation.

      isn’t the tech itself not neutral because of the conditions that resulted in it’s development? It was created for a political reason, and that is not inseparable from itself

      I think this applies to corporate closed models, but I don’t see how it applies to open community developed models or models developed in countries like China where different political conditions exist. Surely you don’t believe that DeepSeek exists to promote right wing political agenda in the US. Similarly, there are lots of Chinese stable diffusion models. Given that this technology exists, and it’s being actively developed outside the west, I don’t really think this argument holds water.

      What you could argue is that we should promote use models from China, or build new models that are community driven from the start. I think that would be very good messaging. Realistically, we could even build and train new models entirely from scratch and only train them exclusively on materials that are not copyrighted which would avoid the whole issue of using work of artists without permission. This is also something that should be advocated as well for in my opinion.

      If we’re not happy with the way this technology is being developed, then the solution has to be to do community driven development where artists are part of the process from the start and have a voice. While corporations aren’t going to stop developing this tech for ethical reasons, open models can cut into their profits. DeepSeek basically destroyed the whole OpenAI business model overnight. That’s a huge victory over corporate driven development, and something that should be celebrated and encouraged.

      One thing I never see discussed, maybe because it’s still unknown territory, is the plateau of the tech.

      This is a very good question, and it will be interesting to watch where this tech peaks in the coming years. We’re already starting to see some limitations in terms of context size, and it turns out simply making models bigger does not lead to better outputs. At the same time this tech is still in its infancy, and people will probably be figuring out a lot of interesting techniques for improving it for a while yet. DeepSeek approach of using mixture of agents and reinforcement learning produced massive gains, showing that there was low hanging fruit waiting to be picked.

      Also, would it still make sense to exist and be used on a society with a different mode of production where capital and profit are not the main motors of society?

      I think it would because it would just mean that everyone would be able to take an idea in their head and make it real. For example, we already see people using AI to make silly things like memes. They’re not doing it for profit or personal gain. They just had a funny thought they wanted to share with their friends.

  • woodenghost [comrade/them]@hexbear.net
    link
    fedilink
    arrow-up
    1
    ·
    edit-2
    3 days ago

    Yes, new technologies in capitalism are adapted, because they shift the organic composition of capital towards fixed capital, initially increasing profits for individual firms but inevitably contributing to the tendency of the rate of profit to fall, leading to crisis eventually. A technology, that doesn’t lower the socially necessary labor time will inevitably fail. So any technology that does succeed in capitalism is a technology that has the potential to be freeing us from labor under communism. Until then, it’s impact depends not on the specific technology, but on how it changes the power dynamic between capital and labor. I agree, that for AI, this could go both ways.

    To recycle my own comment from another thread: Take software development for example. It’s a a field with unusually high wages despite almost no unionization. That’s because it’s organic composition of capital leans towards variable capital. The tools of the trade are cheap. Like a skilled artesian, a software developer can just take their laptop and walk, if their wage is too low. An engineer in a car factory might be just as skilled, but can’t take the robots and assembly lines and walk out, their field has much more fixed capital. So labor in the field of software development has high individual bargaining power, even without collective bargaining.

    But like almost every technical innovation ever, AI will shift the organic composition of capital towards fixed capital. This could lower the bargaining power of workers and drive down wages. That’s why they push it. For example, if huge server farms to drive closed source, centralized AI models become the norm, software engineers won’t be able to just take those with them and walk out as easily as before. On the other hand, small, cheap, specialized, easy to train, open source models (like China develops) might actually benefit labor power. It will be necessary to fight for democratic control over AI to decide whether it’s a blessing or a curse.

  • CarlMarks@lemmygrad.ml
    link
    fedilink
    arrow-up
    1
    ·
    3 days ago

    I will offer a few critiques.

    The first is that “AI”, which is really LLMs and more advanced CNNs, is oversold as a technology and does not deliver what is promised. It can make a 90% as good stocj image, sure, but it does not code well. For coding, at best, it functions as a template creator that often makes mistakes. And it does not do the actually important part of coding, which is conceptualizing and solving the challenges of the actual problem domain. This is because all it can really do is regurgitate patterns based on patternes inputs. There is not any actual understanding underneath it. In this sense, it is a parlor trick compared to what the concept of actual AI evokes. We are essentially in a tech bubble phase of overpromising and overinvestment by two camps of capital: one that buys the hype as a technology and wants to get ahead of “the curve” to cut costs (your example of automation under capitalism) and the cynical crowd that understands its limits but uses it anyways to discipline labor. Large companies already wanted to do layoffs, this just provides an excuse. And it’s just that: an excuse, not really much of a variable cost capital saver for capitalists. Its best corporate use is by workers to help them craft diplomatic emails to unreasonable managers. Or maybe the 90%-as-good stock images thing.

    The second critique is of the enclosure of “AI”, which amounts to capitalist enclosure of publicly funded research combined with vast swaths of data from all sources. These large models require substantial up-front capital investment. They are like building a power plant: to democratize it, you need democratic control over their production, not just a free output that can run on consumer hardware. “AI” companies spend huge amounts on scraping data and training their models, two things that “the consumer” does not experience in use value and arguably does not see in value, either, as with other Silicon Valley tech schemes this bubble is fueled by financial capital pumping companies up in the hopes of getting in on monopolies later. Like with Uber being convenient and cheap for the first few years and now being much more expensive, these companies undercharge and take losses now in the hopes of gainjng “marketshare”. The likely outcome if American “AI” took off in isolation would be 2-3 big companies controlling the entire process and selling an exoensive sunscription, its expense depending on whether they killed off the non-“AI” competition or not. For example, stock image generation prices would be kept below Getty stock image prices right up until they killed the actually-photos market, at which point they would charge whatever they wanted, having skipped ahead to the monopoly phase of pricing. And they would not make the models free or open source, they would use their positions to crush all such endeavors and make them illegal. The only hope there is a country like China, who could nationalize the production of these models and direct their production to be towards lower power and free/open end use. You can see this dynsmic already and the threat it poses for US “AI” companies in their response to DeepSeek: they are trying to get it and all Chinese “AI” banned under claims of national security and intellectual property. This is just their response to a lower power LLM produced by a private group, which highlights their mechanism of enclosure is the high capital barrier to entry. You mention this, but I want to emphasize that it is the entire game.

    I do also want to note that part of the hype is itself disrespectful to artists. As it exists, most “AI” visual productions are a fun parlor trick but not actually good enough to even convince people they are part of, say, a narrative work you would want to personally enjoy, let alone pay for. There is a lot of “searching for the niche” where people are taking what it is already almost convincing at reproducing and trying to shoehorn it in simply because it’s “AI” and cheaper than hiring artists, not because anyone actually prefers it (or doesn’t notice it) aesthetically. We kind of all feel the vibes of an “AI” generated image, let alone movie. When they have to form a narrative structure with consistent character designs and a sense of place, with unique scenarios, it currently all falls apart outside of a few highly skilled individuals that can work these systems (even then you can always recognize it as “AI”). So in this context, think about what it means to artists when people around them say “AI” slop is just as good as their work, or good enough to do the job. I mean, we all know it isn’t either of those things. 99% of the time it’s slop and it doesn’t make any interesting creative “decisions” (it literally can’t). So it is really, in these cases, a way to demean the value of art and artists by exaggerating the value of “AI” “art” to equate them, even on the aesthetic output value.

    To summarize: “AI” has silicon valley financial capital dynamics that threaten any open/democratic use, it is oversold, and overselling it does harm to those whose jobs and interests are meant to be displaced per techbro marketing.

    With that said, there are entirely valid use cases and situations where it performs very well, usually when highly constrained by heuristic modsls that only use it for a subset of aspects. For example, text-to-speech that relies on “AI” for speech generation and only part of translation but not all of it. In that sense it operates like a traditional technology, though still with high cost capital inputs. Though in this same case, there are many slop examples out there, usually when too much is offloaded to the “AI” models, like auto-translation devices for tourists (they do NOT work well).

  • inferiorjc@lemmygrad.ml
    link
    fedilink
    arrow-up
    1
    ·
    4 days ago

    Another Yogthos masterpiece and best position I’ve read yet on AI. Does anyone else have any bangers or good historical quotes on this moment?

    If Lenin thought “a large scale machine industry that is also capable of reorganizing agriculture is the only material basis that is possible for socialism”, I’m thinking the same is true to reorganize industry itself to create the possibility for communism.

    I’m seeing massive AI-driven changes across all the sectors my industry touches, and regressive attempts to erase this momentum rather than seize it for liberation astound me.

  • Commiejones@lemmygrad.ml
    link
    fedilink
    arrow-up
    0
    ·
    edit-2
    3 days ago

    Well written. It hits all the points I have come to in debating with neo-luddites. This whole argument happened in miniature when computer illustration and digital music took off in the mid 00-10s. People said that it wasn’t real art because people weren’t making the sounds with mechanical tools or because the digital painters hadn’t spent years practicing how to hold a brush.

    The cries of “theft” are silly because every artist is influenced by other artists. Nothing is created in a vacuum and no artist can trace every influence on their work even if they pay lip service to the obvious influences. Nobody cries Plagiarism when a painter does cubism that looks exactly like something Picasso would do.

    Only corporate “artists” stood up on the side of the music industry when they tried sued rappers for using samples. That was directly lifting someone else’s identifiable product and putting it on loop with only minimal modification if any.

    Automating image generation isn’t destroying art. Image creation for cash is what kills art. Commoditizing creativity was the sin not the automation of image generation. The existence of LLM image generations doesn’t stop people from doing illustrations. The existence of cameras didn’t stop people from painting portraits. The invention of flash videos didn’t kill hand drawn cartoons. There is no threat to “art” it is a threat to occupations.

    The issue is that people who generate images for money are losing their control over the means of production. Their investment in those means is depreciating rapidly due to new tools that make any untrained worker capable of the what was once skilled labor.

    Just like the Luddites who feared losing their occupation after investing years perfecting their skills these image generations workers fear automation and are trying to destroy the machines instead of destroying the rich people who are profiting by their destruction.