They are experimenting and tuning. Apparently without any correction there is significant racist bias. Basically the AI reflects the long term racial bias in the training data. According to this BBC article it was an attempt to correct this bias but went a bit overboard.
PS: I find it hilarious. If anything it elevates the AI system to art, since it now provides an emotionally provoking mirror about white identity.
Significant racist bias is an understatement.
I asked a generator to make me a “queen monkey in a purple gown sitting on a throne” and I got maybe two pictures of actual monkeys. I even tried rewording it several times to be a real monkey, described the hair and everything.
The rest were all women of color.
Very disturbing. Pretty ladies, but very racist.
Stable diffusion online version, several weeks ago. Might not be the same situation anymore, idk how often that stuff gets updated, and I’m not able to test it at the moment.
It’s also possible that some sort of “sticky idea” got into its head and made it start generating it that way after it did one like that. I’ve heard that sort of thing isn’t uncommon.
Apparently without any correction there is significant racist bias.
This doesn’t make it any less ridiculous. This is a central pillar of this kind of AI tech, and they’re trying to shove a band aid over the most obvious example of it. Clearly, that doesn’t work. It’s also only even attempting to fix one of the “problems” - they’re never going to be able to “band aid” every single place where the AI exhibits this problem, so it’s going to leave thousands of others un-fixed. Even if their band aid works, it only continues to mask the shortcomings of this tech and makes it less obvious to people that it’s horrendously inacurrate with the other things it does.
Basically the AI reflects the long term racial bias in the training data. According to this BBC article it was an attempt to correct this bias but went a bit overboard.
Exactly. This is a core failing of LLM tech. It’s just going to repeat all the shit it was fed to it. You’re never going to fix that. You can attempt to steer it in different directions, but the reason this tech was used was because it is otherwise impossible for us to trudge through all the info that was fed to it. This was the only way to get it to “understand” everything. But all of it’s understandings are going to have these biases, and it’s going to be just as impossible to run through and fix all of these. It’s like you didn’t have enough metal to build the titanic so you just built it out of Swiss cheese and are trying to duct tape one hole closed so it doesn’t sink. It’s just never going to work.
This being pushed as some artificial INTELLIGENCE is the problem here. This shit doesn’t understand what it’s doing, it’s just regurgitating the things it’s consumed. It’s going to be exactly as flawed as whatever was put into it, and you can’t change that. The internet media it was trained on is racist, biased, full of undeniably false information, and massively swayed by propaganda on all sides of the fence. You can’t expect LLMs to do anything different when trained on that data. They’re going to have all the same problems. Asking these things to give you any information is like asking the average internet user what the answer is. And the average internet user is not very intelligent.
These are just amped up chat bots with data being sourced from random bits of the internet. Calling them artificial INTELLIGENCE misleads people into thinking these bots are smart of have some sort of understanding of what they’re doing. They don’t. They’re just fucking internet parrots, and they don’t have the architecture to be “fixed” from having these problems. Trying to patch these problems out is a fools errand and only masks their underlying failings.
Would it be possible to create a kind of “formula” to express the abstract relationship of ethical makeup, location, year and field? Like convert a table of population, country, ethnicity mix per year and then train the model on that. It’s clear that it doesn’t understand the meaning or abstract concept, but it can associate and extrapolate things. So it could “interpret” what the image description says while training and then use the prompt better. So if you’d prompt “english queen 1700” it would output white queen, if you input year 2087 it would be ever so slightly less pasty.
I don’t know, maybe that would work, for this one particular problem. My point is it’s more than that. Even if you go through the trouble of fixing this one particular issue with LLMs, there are literally thousands of other problems to solve before it’s all “fixed”. At some point, when you’ve built and maintained thousands of workarounds, they start conflicting with each other and making a giant spider web of issues to juggle.
And so you’re right back at the problem that you were trying to solve by building the LLM in the first place. This approach is just futile and nonsensical.
None of this has been pushed, by any researcher, by any company, by any open source group even, as “intelligence” In fact, it was unanimously disliked as a term by everyone working with the models and transformers, but media circus combined with techbros laymen hard on hype have won. Since then everyone has given up trying to be semantically correct on this front.
I didn’t say any researcher or anything had named it intelligence. Nor am I trying to be semantically correct.
Read the guys comments. He’s trying to push the idea that we can “change” it’s “understanding” about the things it’s discussing. He is one of the people who has fallen for the tech bros etc convincing people it is intelligent. I’m not fighting semantics, I’m trying to explain to him that it’s not intelligent. Because he himself clearly doesn’t understand that.
For example, a prompt seeking images of America’s founding fathers turned up women and people of colour.
“A bit” overboard yeah
We all expected the AIs to launch nukes, and they simply held up a mirror.