AI CEOs be like
Online Communism 😃
Real Life Communism 😠
Communism for fascists and fascism for the commons. It’s the american way.
Online communism unless you actually want something for free.
More like theft of data = “communism”
I work as an AI engineer and any company that’s has a legal department that isn’t braindead isn’t stealing content. The vast majority of content posted online is done so under a creative commons licence, which allows other people to do basically whatever they want with that content, so it’s in no way shape or form stealing.
Yeah I know, but the press pics some rare cases and says, that this goes for all.
You know how it is.
It really depends on what the site’s terms of service/usage agreement says about the content posted on the site.
For example a site like Art Station has users agree to a usage license that lets them host and transmit the images in their portfolios. However there is no language saying the visitors are also granted these rights. So you may be able to collect the site text for fair use, but the art itself requires reaching out for anything other than personal/educational use.
I remember when copying data wasn’t theft, and the entire internet gave the IP holders shit for the horrible copyright laws…
If its published online its not theft Thats like saying that if i publish a book and someone uses it to learn a language then they are stealing my book
Ok then picture this: A webscrapper is copying code that has copyright that indicates, that it is forbidden to modify this code, to publish it under a different name or to sell this code (for example a method that calculates the inverse square root really fast).
By using the code as training data, most language models may actually paste this code or write it with little change, because most language models are based not on writing something that has a purpose that is given by the user, like for example AI’s that are supposed to evaluate pictures of dogs and cats and is supposed to decide which is which, but they are based on the following schematic:
-
Read previous text
-
Predict, what letter will follow
-
Repeat until user interferes.
Because language models work this way, if I would for example only train it on the novsl “Alice in Wonderland”, then there is a high possibly, that the model will reproduce parts of it.
But there is a way to fix this problem: If we broaden the training data very much, the chance the output would be considered plagiarism will narrow down.
Upon closer inspection there is another problem though, because AI (at this point in time), don’t have an influence from outside in a sense like humans do: A human is experiencing every day of their life with there being a chance of something happening, that modifies their brain structure through emotion, like for example a chronic depression. This influences the output of a person not only in their symptoms, but also in the way they would write text for example.
The consequence is that the artist may use this emotional change to express it with their art.
Every day influences the artist differently and inspires them with unseen and new thoughts.
The AI (today) has the problem that they definitely retell the stories it has heard again and again like Aristotle (?) says.
The outer influence is missing to the AI at this point in time.
If you want to have a model that can give you things, that never have been written before or don’t even seem like anything that there has ever been, you have to give it these outer influences.
And there is a big problem coming up, because, yes this process could be implemented by training the AI even further after it already was launched, by reinforced learning, but this process would still need data input from humans, which is really annoying.
A way to make it easier would be to give an AI a device on which it can run and sensors as well as output devices, so that it can learn from its sensors and use this information in its post-training training phase to gather more data and make current events, that it perceives, relevant.
As you can see, if we would do that, then we would have an AI that could do anything, and learn from anything, which both makes it really really fucking dangerous, but also really really fucking interesting.
lol that’s funny my machine learning course used Alice in wonderland as an example for us to train our projects on
That wasn’t a coincidence, I watched a video some time ago, where they used tensor flow to do exactly that.
(I’m doing my CS bachelors btw)
Oh OK nice! Good luck!!
-
Well, theft of the labor of contributors to give to non contributors is communism … So, your statement is true, it’s just more broad than that.
Intellectual property is fake lmao. Train your AI on whatever you want
deleted by creator
You are allowed to use copyrighted content for training. I recommend reading this article by Kit Walsh, a senior staff attorney at the EFF if you haven’t already. The EFF is a digital rights group who most recently won a historic case: border guards now need a warrant to search your phone.
deleted by creator
The comparision doesn’t work. Because the AI is replacing the pencil or other drawing tool. And we aren’t saying pencil companies are selling you Mario pics because you can draw a Mario picture with a pencil either. Just because the process of how the drawing is made differs, doesn’t change the concept behind it.
An AI tool that advertises Mario pcitures would break copyright/trademark laws and hear from Nintendo quickly.
deleted by creator
I don’t think how you interact with a tool matters. Typing what you want, drawing it yourself, or clicking through options is all the same. There are even other programs that allow you to draw by typing. They are way more difficult but again, I don’t think the difficulty matters.
There are other tools that allow you to recreate copyrighted material fairly easily. Character creators being on the top of the list. Games like Sims are well known for having tons of Sims that are characters from copyrighted IP. Everyone can recreate Barbie or any Disney Princess in the Sims. Heck, you can even download pre made characters on the official mod site. Yet we aren’t calling out the Sims for selling these characters. Because it doesn’t make sense.
deleted by creator
I don’t buy the pencil comparison. If I have a painting in my basement that has a distinctive style, but has never been digitized and trained upon, I’d wager you wouldn’t be able to recreate neither that image nor it’s style. What gives? Because AI is not a pencil but more like a data mixer you throw complete works in into and it spews out colllages. Maybe collages of very finely shredded pieces, to the point you could even tell, but pieces of original works nontheless. If you put any non-free works in it, they definitely contaminate the output, and so the act of putting them in in the first place should be a copyright violation in itself. The same as if I were to show you the image in question and you decided to recreate it, I can sue you and I will win.
That is a fundamental misunderstanding of how AI works. It does not shred the art and recreate things with the pieces. It doesn’t even store the art in the algorithm. One of the biggest methods right now is basically taking an image of purely random pixels. You show it a piece of art with a whole lot of tags attached. It then semi-randomly changes pixel colors until it matches the training image. That set of instructions is associated with the tags, and the two are combined into a series of tiny weights that the randomizer uses. Then the next image modifies the weights. Then the next, then the next. It’s all just teeny tiny modifications to random number generation. Even if you trained an AI on only a single image, it would be almost impossible for it to produce it again perfectly because each generation starts with a truly (as truly as a computer can get, an unweighted) random image of pixels. Even if you force fed it the same starting image of noise that it trained on, it is still only weighting random numbers and still probably won’t create the original art, though it may be more or less undistinguishable at a glance.
AI is just another tool. Like many digital art tools before it, it has been maligned from the start. But the truth is what it produces is the issue, not how. Stealing others’ art by manually reproducing it or using AI is just as bad. Using art you’re familiar with to inspire your own creation, or using an AI trained on known art to make your own creation, should be fine.
As a side note because it wasn’t too clear from your writing, but the weights are only tweaked a tiny tiny bit by each training image. Unless the trainer sees the same image a shitload of times (Mona Lisa, that one stock photo used to show off phone cases, etc) then the image can’t be recreated by the AI at all. Elements of the image that are shared with lots of other images (shading style, poses, Mario’s general character design, etc) could, but you’re never getting that one original image or even any particular identifiable element from it out of the AI. The AI learns concepts and how they interact because the amount of influence it takes from each individual image and its caption is so incredibly tiny but it trains on hundreds of millions of images and captions. The goal of the AI image generation is to be able to create vast variety of images directed by prompts, and generating lots of images which directly resemble anything in the training set is undesirable, and in the field it’s called over-fitting.
Anyways, the end result is that AI isn’t photo-bashing, it’s more like concept-bashing. And lots of methods exist now to better control the outputs, from ControlNet, to fine-tuning on a smaller set of images, to Dalle-3 which can follow complex natural language prompts better than older methods.
Regardless, lots of people find that training generative AI using a mass of otherwise copyrighted data (images, fan fiction, news articles, ebooks, what have you) without prior consent just really icky.
I agree.
The problem is that you might technically be allowed to, but that doesn’t mean you have the funds to fight every court case from someone insisting that you can’t or shouldn’t be allowed to. There are some very deep pockets on both sides of this.
Let them come.
I like your moxy, soldier
What a tough guy you are, based AI artist fighting off all the artcucks like a boss 😎
deleted by creator
Nah all information should be freely available to as many people as practically possible. Information is the most important part of being human. All copyright is inherently immoral.
deleted by creator
If copyright was vigorously enforced a lot more people would starve than would be fed
deleted by creator
Go on etsy search mickey mouse and go report all the hundreds of artists whose whole career is violating the copyright of the most litigious company on the planet.
deleted by creator
deleted by creator
deleted by creator
deleted by creator
deleted by creator
“Artists don’t deserve to profit off their own work” is stupid as shit. Complain about copyright abuse and lobbying a la Disney and I’ll be right there with you, but people shouldn’t have the right to take your work and profit off it without either your consent or paying you for it.
Artists and other creatives who actually do work to create art (not shitting out text into an image generator) should take every priority over AI “creators.”
Equating training AI to not being able to profit is stupid as shit and the same bullshit argument big companies use to say “we lost a bazillion dollars to people pursuing out software” someone training their AI on an art work (that is probably under a creative commons licence anyway) does suck money out of an artists pocket they would have otherwise made.
Artists and other creatives who actually do work to create art (not shitting out text into an image generator) should take every priority over AI “creators.”
Why are you the one that gets to decide what is “work” to create art? Should digital artists not count because they are computer assisted, don’t require as much skill and technique as “traditional” artists and use tools that are based on the work of others like, say, brush makers?
And the language you use shows that you’re vindictive and angry.
They said IP, IP protects artists from having their work stolen. The fact AI guzzlers are big mad that IP might apply to them too is irrelevant.
Digital artists do exactly as much work as traditional artists, comparing it to AI “art” from an AI “artist” is asinine. Do you actually think digital artists just type shit in and a 3D model appears or something?
And yeah I’m angry when my friends and family who make their living as actual artists, digital and traditional, have their work stolen or used without their permission. They aren’t fucking corporations making up numbers about lost sales, they’re spending weeks trying to get straight up stolen art mass printed on tshirts and mugs removed from online sale. They’re going outside and seeing their art on shit they’ve never sold. Almost none of them own a home or even make enough to not have a regular job, it’s literally taking money out of their pockets to steal their work. This is the shit you’re endorsing by shitting on the idea of IP.
Do you actually think artists using AI tools just type shit into the input and output decent art? It’s still just a new, stronger digital tool. Many previous tools have been demonized, claiming they trivialize the work and people who used them were called hacks and lazy. Over time they get normalized.
And as far as training data being considered stealing IP, I don’t buy it. I don’t think anyone who’s actually looked into what the training process is and understands it properly would either. For IP concerns, the output should be the only meaningful measure. It’s just as shitty to copy art manually as it is to copy it with AI. Just because an AI used an art piece in training doesn’t mean it infringed until someone tries to use it to copy it. Which, agreed, is a super shitty thing to do. But again, it’s a tool, how it’s used is more important than how it’s made.
Lmao, I’ve used AI image generation, you’re not going to be able to convince me any skill was involved in what I made. The fact some people type a lot more and keep throwing shit at the wall to see what sticks doesn’t make it art or anything they’ve done with their own skill. The fact none of them can control what they’re making every time the sauce updates is proof of that.
If it’s so obviously not IP violating to train with it then I’m sure it’ll be totally fine if they train them without using artists’ work without permission, since it totally wasn’t relying on those IP violating images. Yet for some reason they fight this tooth and nail. 🤔
“I can” and “you can’t” are not opposites.
I have no idea what you’re trying to convey here.
“Artists don’t deserve to profit off their own work” is not what anyone said.
Even if people can just take your shit and profit off it - so can you.
This is not a complete rebuttal, but if you need the core fallacy spelled out, let’s go slowly.
Why are you entitled to profit off the labor of someone else?
Why am I not entitled to profit from my own work building on stories I enjoy?
Isn’t this a completely different conversation than the one we were having and kind of missing the point? Yes, imo you should be allowed to do that. Still, AI Companies are using the labor of millions of artist for free to train their AIs, which are then threatening to eliminate ways of these artist to gather income.
How is that related in any way to the ways that copyright has been exploited against fanmade art?
Even if people can just take your shit and profit off it - so can you.
What entitles someone to take another person’s work and profit off it?
Fuck your sophistry.
No. Fuck that. I don’t consent to my art or face being used to train AI. This is not about intellectual property, I feel my privacy violated and my efforts shat on.
Unless you have been locked in a sensory deprivation tank for your whole life, and have independently developed the English language, you too have learned from other people’s content.
Well my knowledge can’t be used as a tool of surveillance by the government and the corporations and I have my own feelings intent and everything in between. AI is artifical inteligence, Ai is not an artificial person. AI doesn’t have thoughts, feelings or ideals. AI is a tool, an empty shell that is used to justify stealing data and survelience.
This very comment is a resource that government and corporations can use for surveillance and training.
AI doesn’t have thoughts? We don’t even know what a thought is.
We may not know what comprises A thought, but I think we know it’s not matrix math. Which is basically all an LLM is
Hard disagree, the neural connections in the brain can be modeled with matrix math as well. Sure some people will be uncomfortable with that notion, especially if they believe in spiritual stuff outside physical reality. But if you’re the type that thinks consciousness is an emergent phenomenon from our underlying biology then matrix math is a reasonable approach to representing states of our mind and what we call thoughts.
Shut the fuck up tech bro xD
Yet we live in a world where people will profit from the work and creativity of others without paying any of it back to the creator. Creating something is work, and we don’t live in a post-scarcity communist utopia. The issue is the “little guy” always getting fucked over in a system that’s pay-to-play.
Donating effort to the greater good of society is commendable, but people also deserve to be compensated for their work. Devaluing the labor of small creators is scummy.
I’m working on a tabletop setting inspired by the media I consumed. If I choose to sell it, I’ll be damned if I’m going to pay royalties to the publishers of every piece of media that inspired me.
If you were a robot that never needed to eat or sleep and could generate 10,000 tabletop RPGs an hour with little to no creative input then I might be worried about whether or not those media creators were compensated.
The efficiency something can be created with should have no bearing on whether someone gets paid royalties.
It absolutely should, especially when the “creator” is not a person. AI is not “inspired” by training data, and any comparisons to human artists being inspired by things they are exposed to are made out of ignorance of both the artistic process and how AI generates images and text.
When, not if, such a robot exists, do you imagine we’ll pay everyone who’d ever published a roleplaying game? Like a basic income exclusively for people who did art before The Training?
Or should the people who want things that couldn’t exist without magic content engines be denied, to protect some prior business model? Bear in mind this is the literal Luddite position. Weavers smashed looms. Centuries later, how much fabric in your life do you take for granted, thanks to that machinery?
‘We have to stop this labor-saving technology because of capitalism’ is a boneheaded way to deal with any conflict between labor-saving technology… and capitalism.
I know this is shocking to people who have never picked up a brush or written anything other than internet arguments after the state stopped mandating they do so when they graduated high school, but you can just create shit without AI. Literal children can do it.
And there is no such thing as something that “couldn’t exist” without the content stealing algorithm. There is nothing it can create that humans can’t, but humans can create things it can’t.
There’s also something hilarious about the idea that the real drag on the artistic process was the creating art part. God almighty, I’d rather be a Luddite than a Philistine.
Those 10000 tabletop RPGs will almost certainly be completely worthless on their own, but might contain some novel ideas. Until a human comes by and scours it for ideas and combines it. It could very well be that in the same time it could only create 1 coherent tabletop RPG idea.
Should be mentioned though, AIs don’t run for free either, they cost quite a lot of electricity and require expensive hardware.
what a shitty argument. AI isn’t people
Then don’t post your art or face publicly, I agree with you if it’s obtained through malicious ways, but if you post it publicly than expect it to be used publicly
If you post your art publicly why should it be legal for Amazon to take it and sell it? You are deluding yourself if you believe AI having a get out of jail free card on IP infringement won’t be just one more source of exploitation for corporations.
Taking it and selling it is obviously not legal, but taking it and using it for training data is a whole different thing.
Once a model has been trained the original data isn’t used for shit, the output the model generates isn’t your artwork anymore it isn’t really anybody’s.
Sure, with some careful prompts you can get the model to output something in your style fairly closely, but the outputting image isn’t yours. It’s whatever the model conjured up based on the prompt in your style. The resulting image is still original of the model
It’s akin to someone downloading your art, tracing it over and over again till they learn the style and then going off to draw non-traced original art just in your style
“You don’t understand, it’s not infringement because we put it in a blender first” is why AI “art” keeps taking Ls in court.
Care to point to those Ls in court? Because as far as I’m aware they are mostly still ongoing and no legal consensus has been established.
It’s also a bit disingenuous to boil his argument down to what you did, as that’s obviously not what he said. You would most likely agree that a human would produce transformative works even when basing most of their knowledge on that of another (such as tracing).
Ideas are not copyrightable, and neither are styles. You could use that understanding of another’s work to create forgeries of their work, but hence why there exist protections against forgeries. But just making something in the style of another does not constitute infringement.
EDIT:
This was pretty much the only update on a currently running lawsuit I could find: https://www.reuters.com/legal/litigation/us-judge-finds-flaws-artists-lawsuit-against-ai-companies-2023-07-19/
“U.S. District Judge William Orrick said during a hearing in San Francisco on Wednesday that he was inclined to dismiss most of a lawsuit brought by a group of artists against generative artificial intelligence companies, though he would allow them to file a new complaint.”
"The judge also said the artists were unlikely to succeed on their claim that images generated by the systems based on text prompts using their names violated their copyrights.
“I don’t think the claim regarding output images is plausible at the moment, because there’s no substantial similarity” between images created by the artists and the AI systems, Orrick said."
Hard to call that an L, so I’m eagerly awaiting them.
Biggest L: that shit ain’t copyrightable. A doodle I fart out on a sticky pad has more legal protection.
A court saying “nobody owns this” is closer to the blender argument than the ripoff argument.
Go fuck yourself AI bitch :3
If the large corporations can use IP to crush artists, artists might as well try to milk every cent they can from their labor. I dislike IP laws as well, and you can never use the masters’ tools to dismantle their house, but you can sure as shit do damage and get money for yourself.
Luckily, AI aren’t the master’s tools, they’re a public technology. That’s why they’re already trying their had at regulatory capture. Just like they’re trying to destroy encryption. Support open source development, It’s our only chance. Their AI will never work for us. John Carmack put it best.
AI algorithms aren’t the masters tools in the sense that anyone can set up a model using free code on cheap tech, but they do require money to improve and produce quality products. The training data requires labor of artists to produce, the computers the model runs on require labor to make, and the electricity that allows the model to continuously improve requires, you guessed it, labor. Labor will always, and should always, be expensive, making AI most useful to those with money to spend.
IP is only one tool in the masters’ tool box, but capitalism is their box, and the other forms of capital can be used to swing the public’s tools far harder than a wage laborer.
Hence why it’s important that open source models should keep the support of the public. If they have to close shop, what remains will be those made by tech giants. They will be censored, they will be neutered, except if you pay enough for them. The power should remain with the artists, and not those with the money.
Here is an alternative Piped link(s):
They have the plant, but we have the power
Piped is a privacy-respecting open-source alternative frontend to YouTube.
I’m open-source; check me out at GitHub.
Based take.
Me, literally training a neutral net to generate pictures of carrot cakes right now:
WHERE DID YOU GET THE DATA?
I feel the current AI crawling bots + “opt-out your data” tactic is ingeniously evil.
It’s hilarious really
Companies have been stealing data for so long, and then another company comes and steals their data by scraping it they go surprised Pikachu
The best part is the end result, not where the data comes from. Tired of hearing about AI models “stealing” data. I put all my art, designs and code online and assume it’ll be used to train models (which I’ll be able to use later on)
You’re fine with a corporation making money off your copyrighted work? Without seeing a cent of it?
Why not? Same as a person being inspired to reuse certain aspects. Artists reuse other artists work constantly and usually more blatantly than what AI does.
I wouldn’t want to copyright every visual pattern conceivable, everything would be a copyright violation of some sort.
Unfathomably based
Disney doesn’t agree
Now I’m pretty sure they won’t see a conflict of interest in not using their IP to train AI but advocating for your data for training.
The inverse is also true. Disney will make their own AI regardless of being able to use anyone else’s data for training. Because they have a ton of data already. The only ones that will be shafted if that freedom is restricted is that those without a library of data to train will have access to AI.
So long as it’s correctly attributed it’s no more immoral than what a corpo inherently does.
I interpreted this less as being about art models and more about predictive models trained off of data with racial prejudice (i.e. crime prediction models trained off of old police data)
You’re putting your stuff out there so some cryptobro cracker can steal it, claim credit for it, and then call you a peasant for helping him make it.
It’s sort of an accelerationistic take on broken IP law I’m sure of it.
One thing I’ve started to think about for some reason is the problem of using AI to detect child porn. In order to create such a model, you need actual child porn to train it on, which raises a lot of ethical questions.
Cloudflare says they trained a model on non-cp first and worked with the government to train on data that no human eyes see.
It’s concerning there’s just a cache of cp existing on a government server, but it is for identifying and tracking down victims and assailants, so the area could not be more grey. It is the greyest grey that exists. It is more grey than #808080.
well, many governments had no issue taking over a cp website and hosting it for montha to come, using it as a honeypot. Still they hosted and distributed cp, possibly to thousands of unknown customers who can redistribute it.
I’m pretty sure those AI models are trained on hashes of the material, not the material directly, so all you need to do is save a hash of the offending material in the database any time that type of material is seized
That wouldn’t be ai though? That would just be looking up hashes.
You’re almost there…
Nah, flipping the image would completely bypass a simple hash map
From my very limited understanding it’s some special hash function that’s still irreversible but correlates more closely with the material in question, so an AI trained on those hashes would be able to detect similar images because they’d have similar hashes, I think
Perceptual hashes, I think they’re called
could you provide a source for this? that spunds very counterintuive and bad for the hash functions. especially as the whole point of AI training in this case is detecting new images. And say a small boy at the beach wearing speedos has a lot of similiarity to a naked boy. So looking by some resemblance in the hash function would require the hashes to practically be reversible.
Not all hashes are for security. They’re called perceptual hashes
Probably a case of definitional drift of the word, because it probably should be just for the security kind.
I’m no expert, but we use those kind of hashes at my company to detect fraudulent software Here’s a Wikipedia link: https://en.m.wikipedia.org/wiki/Locality-sensitive_hashing
You absolutely do not real CSAM in the dataset for an AI to detect it.
It’s pretty genius actually: just like you can make the AI create an image with prompts, you can get prompts from an existing image.
An AI detecting CSAM would have to be trained on nudity and on children separately. If an image-to-prompts conversion results in “children” AND “nudity”, it is very likely the image was of a naked child.
This has a high false positive rate, because non-sexual nude images of children, which quite a few parents have (like images of their child bathing) would be flagged by this AI. However, the false negative rate is incredibly low.
It therefore suffices for an upload filter for social media but not for reporting to law enforcement.
This dude isn’t even whining about the false positives, they’re complaining that it would require a repository of CP to train the model. Which yes, some are certainly being trained with the real deal. But with law enforcement and tech companies already having massive amounts of CP for legal reasons, why the fuck is there even an issue with having an AI do something with it? We already have to train mods on what CP looks like, there is no reason its more moral to put a human through this than a machine.
This is a stupid comment trying to hide as philosophical. If your website is based in the US (like 80 percent of the internet is), you are REQUIRED to keep any CSAM uploaded to your website and report it. Otherwise, you’re deleting evidence. So all these websites ALREADY HAVE giant databases of child porn. We learned this when Lemmy was getting overran with CP and DB0 made a tool to find it. This is essentially just using shit any legally operating website would already have around the office, and having a computer handle it instead of a human who could be traumatized or turned on by the material. Are websites better for just keeping a database of CP and doing nothing but reporting it to cops who do nothing?
Yeah, a real fuckin moral quandary there, I bet this is the question that killed Kant.
deleted by creator
For Stable Diffusion, it comes from images on Common Crawl through the LAION 5b dataset.
Never forget: businesses do not own data about you. The data belongs to the data subject, businesses merely claim a licence to use it.
Legally, businesses very much own the data about you unfortunately.
No, they very explicitly don’t. They claim a licence in perpetuity to nearly all the same rights as the data owner, but the data subject is still the owner.
Also, that licence may not be so robust. A judge should see that the website has no obligation to continue hosting the website, and they offer nothing in return for the data, so the perpetual licence is not a reasonable term in the contract and should be struck down to something the data subject can rescind. In some respects we do have this with “the right to be forgotten” and to have businesses delete your data, however the enforcement of this is sorely lacking.
Laws change over time, though. Everyone is the victim of this practice, so eventually the law should catch up.
What do you think owning is? If I pull your strings, you’re my puppet. Ownership is control.
not necessarily. Your landlord is the owner of your flat, but he ceedes all rights to control over it to you, within the limits of the contract and can only take back co trol if he has evidence of you violating the contract.
Ironically your metaphor works really well as an example of how that’s not true - just because you’re allowed to pull the strings, doesn’t mean you’re the owner of the puppet
You know, that’s kinda fair. That applies pretty well for media licensing in particular.
Jesus Christ who called in the tech bro cavalry? Get a fucking life losers you’re not artists and nobody is proud of you for doing the artistic equivalent of commissioning an artist (which you should be doing instead of stealing their art and mashing it into a shitty approximation of art)
It’s like photography. Photography + photoshop for some workflows. There’s a low barrier to entry.
Would you say the same thing to someone proud of how their tracing came out?
These are not comparable to AI image generation.
Even tracing has more artistic input than typing “artist name cool thing I like lighting trending on artstation” into a text box.
So about the same as a photograph then
Removed by mod
Aside from the fact that your comment applies to photography as well, I think it’s fair to point out image generation can also be a complex pipeline instead of a simple prompt.
I use ComfyUI on my own hardware and frequently include steps for control net, depth maps, canny edge detection, segmentation, loras, and more. The text prompts, both positive and negative, are the least important parts in my workflow personally.
Hell sometimes I use my own photos as one of the dozens of inputs for the workflow, so in a sense photography was included.
Annoying and aggressive about it to the point where you’d like to wring their necks? Yeah, that’s exactly what that’s like.
it’s funny, but still, where did you get the data? XD
From Reddit comments.
Content platforms could start to poison their data with random AI generated garbage
found it