It reminds me of that bit from Its Always Sunny in Philadelphia where Dennis is talking about Hollywood movies.
I mean, it used to be only, like, the hard-line conservatives, like the pearl-clutching types, were the only ones that were overly vocal and extreme in their policing of sexuality. But now you got this, like, liberal wave of moral authority sweeping the nation. You know, it's nuts. I mean, think about it. If the conservatives had always run Hollywood, movies would have sucked. You know what I mean? The art would have suffered. So I guess the question we're asking is how will art fare under the oppressive thumb of this new liberal Hollywood moral PC elite?
It's just so silly and yet so accurate. Whether it's social values, politics or even just the opinion of AI and it's capabilities vs. it's potential vs. how people actually use it, there's this pervading idea that restrictions en masse are a viable solution. I feel almost the opposite, like to some extent the oversaturation of it intrinsically lowers the negative reception of it. Prohibition philosophy - when it's not allowed people will work even harder to use it in those ways, when it's not only allowed but widely used and even encouraged, people just inherently care less over time.
We're at a point right now where we are getting some pretty poor quality oversaturation of AI content and the tool alone is what is being blamed, to the point where copyright is being touted as this saving grace despite it consistently having been used against us smaller artists when corporate money is involved. Copyright isn't promoting small artists, rarely has, nor is it preventing AI, but it's somehow suddenly meant to ensure that the art you uploaded isn't reproduced? That seems not only unlikely, but like it's a scapegoat for a larger issue. Generative art isn't a problem because Ms. Jane working two 40-hour jobs uses it to make art featuring existing characters. That circumstance was and never will be a problem because Jane very likely would never have the money to commission an artist in the first place. What Jane makes is 100% irrelevant, so long as she's not claiming it as her original creation and trying to sell it - beyond that? I don't think anyone should care or fault her, because she is doing the amount of art that her circumstances allow her.
What I absolutely agree is an issue is businesses and corporations using AI, cutting staff further overworking employees that remain. However, that Secret Invasion intro that seemed likely AI generated? I can't in good faith try to argue "they should be tried for infringement" but I can fully support the fact that they should have hired an artist who would at least try to better use the tools at their disposal. I can simultaneously feel that the fact that Deforum may have been used is absolutely awesome, while also being annoyed and frustrated that they didn't utilize artists who deserve it.
There is a very large difference between Ms. Jane making AI images, even movies, and any corporate product - or that AI generated rat for the science journal. For the former, it is something that IMO is fully necessary in order for Jane to be able to enjoy the experience of a creative process under the bullshit system we've worked out. The latter is a completely unnecessary replacement used to cut costs. And yet, for neither does the concept of infringement actually matter that much, because copyright isn't the fundamental issue of AI, it's just the one people are latching on to. Without realizing that the likelihood of copyright laws helping someone like us is nil. Especially since there's probably an overlap of people who laugh at NFT's and pirate files because bits of data aren't a physical commodity that runs out, but a generative Imaging tool that does it is... Too far?
I think AI's issues are separate from what I've mentioned here. What people blame AI for is something else entirely. AI is still just the tool that speeds up the process. We have the concept of safeguards utilized as signs, barriers, and nets, so that if someone wants to use a bridge for the wrong purpose there are some measures in place to prevent them. We don't blame bridges for what the person is trying to do - we recognize that there is some reasonable level of safeguard and beyond that we just have to trust the person to do the right thing. And when it does show to be a pervasive issue, even still there is pretty much a bare minimum done - add another layer and a net and call it a day - instead of focusing on maybe why people in society are so inclined to jump.
The issue is always us. Yes AI makes evils job easier, like so many tools have. But trying to safeguard AI to the point of non-existence is just absurd from every angle, given that the bad stuff is likely going to happen in abundance regardless. I don't particularly see AI as the evil so much as the humans creating the meaningless AI generated articles.
Keep in mind that thinner paper is more likely to jam whatever printer you have, so you may want to consider workarounds such as smaller text, wider margins, and thinner line breaks.