I know a lot of people want to interpret copyright law so that allowing a machine to learn concepts from a copyrighted work is copyright infringement, but I think what people will need to consider is that all that’s going to do is keep AI out of the hands of regular people and place it specifically in the hands of people and organizations who are wealthy and powerful enough to train it for their own use.
If this isn’t actually what you want, then what’s your game plan for placing copyright restrictions on AI training that will actually work? Have you considered how it’s likely to play out? Are you going to be able to stop Elon Musk, Mark Zuckerberg, and the NSA from training an AI on whatever they want and using it to push propaganda on the public? As far as I can tell, all that copyright restrictions will accomplish to to concentrate the power of AI (which we’re only beginning to explore) in the hands of the sorts of people who are the least likely to want to do anything good with it.
I know I’m posting this in a hostile space, and I’m sure a lot of people here disagree with my opinion on how copyright should (and should not) apply to AI training, and that’s fine (the jury is literally still out on that). What I’m interested in is what your end game is. How do you expect things to actually work out if you get the laws that you want? I would personally argue that an outcome where Mark Zuckerberg gets AI and the rest of us don’t is the absolute worst possibility.
So what does that mean? Do you not believe that AIs like ChatGPT and Stable Diffusion have neural networks that are made up of simulated neurons? Or are you saying that we haven’t simulated an actual human brain? Because the former is factually incorrect, and I never claimed the latter. Please explain exactly what “hype” you believe I’m buying into? Because I don’t think you have any clue what it is you think I’m wrong about. You just really don’t want me to be right.
I think they simulate what some people think neurons are like. I mean, I guess you can get the binary neurons fine but there are analog neurons (and that is something that has just now been proven). But there are so many value inputs in the human brain that we haven’t isolated, so much about it we haven’t mapped. We don’t even know how the electricity is encoded. So no, I don’t think what you’re calling a “neural network” is ACTUALLY simulating the human brain.
The hype you’re buying into is that AI will improve our lives just by existing. Thing is, any new tech is a weapon in the hands of the rich whether it’s available to the common man or not. We need to focus on setting the rules for the rich and enforcing the rules we have. Copyright, which is also a weapon in the hands of the rich yes, has aspects which are made to protect the common man and we need to enforce those to keep the rich in line while we have them. If someday we junk copyright, it needs to be as a whole. We can’t go chucking copyright for small time authors while the courts are still allowing Disney to keep Mickey Mouse out of the public domain, which is what you suggest doing when you suggest copyright should be ignored so that the common man can make their own AI.
I think I’ve softened on quite a bit with your arguments, honestly. It’s unfair to say I just don’t want you to be right. My position remains that I think copyright is a fair place for limitation on AI training.
So most of my opinions about what AI can do aren’t about hype at all, but what I’ve personally experienced with it firsthand. The news, frankly, is just as bad a source about AI is marketing departments of AI companies, because the news is primarily written by people who feel threatened by its existence and are rationalizing reasons that it’s bad, as well as amplifying bad things that they hear and, in the best case, reporting on it without really understanding what it actually does. The news is partly why you’re connecting what’s happening with that WGA/SAG-AFTRA contract; nothing I’ve said here supports people losing their existing rights to their own likenesses, and the reason they’re trying to slip it into the contracts is because even under existing copyright law, AI isn’t a get out of jail free card to produce copyrighted works despite the fact that you can train it on them.
At any rate, here are a few of my personal experiences with using AI:
This kind of thing is why I’m excited about AI – it’s improving my life in a big way right now. None of what I’ve done with it is “hype”. I don’t care that Elon Musk’s dumb ass is starting his own AI company, or what tech company marketing divisions have to say about it, or what some MBA CEO’s wild guess about what we’ll be using it for in 5 years is.
It’s nice that your life is better, but that doesn’t change that these AIs were trained by being fed the work of creatives who were never compensated for that work.
And it doesn’t change that on the high level and in the real world, they’re pushing to put AI in places AI isn’t ready to be, because they don’t want to pay humans to do those jobs.
I mean, yeah, you don’t care… but the rest of us do.
Waaaaait a minute. How is it you thinking AI is good because it makes your life a bit better leisure-wise different from me thinking it’s a problem because it will make my life worse work-wise? You threw that at me saying I was worried about a small group and here you are basing your excitement on it helping your niche hobbies?
Are you sure you’re not projecting here? In this entire thread, have you budged an inch based on all the people arguing against your original post? Or are you just refusing to admit that it could cause trouble in the world for people’s livelihoods because you get to have fun with it?
When did I refuse to admit automation causes problems for people?