An Asian MIT student asked AI to turn an image of her into a professional headshot. It made her white with lighter skin and blue eyes.::Rona Wang, a 24-year-old MIT student, was experimenting with the AI image creator Playground AI to create a professional LinkedIn photo.

197 points

Look, I hate racism and inherent bias toward white people but this is just ignorance of the tech. Willfully or otherwise it’s still misleading clickbait. Upload a picture of an anonymous white chick and ask the same thing. It’s going go to make a similar image of another white chick. To get it to reliably recreate your facial features it needs to be trained on your face. It works for celebrities for this reason not a random “Asian MIT student” This kind of shit sets us back and makes us look reactionary.

permalink
report
reply
150 points
*

It’s less a reflection on the tech, and more a reflection on the culture that generated the content that trained the tech.

Wang told The Globe that she was worried about the consequences in a more serious situation, like if a company used AI to select the most “professional” candidate for the job and it picked white-looking people.

This is a real potential issue, not just “clickbait”.

permalink
report
parent
reply
34 points

If companies go pick the most professional applicant by their photo that is a reason for concern, but it has little to do with the image training data of AI.

permalink
report
parent
reply
5 points

Especially ones that are still heavily in development

permalink
report
parent
reply
1 point
Removed by mod
permalink
report
parent
reply
28 points

A company using a photo to choose a candidate is really concerning regardless if they use AI to do it.

permalink
report
parent
reply
20 points
*

Some people (especially in business) seem to think that adding AI to a workflow will make obviously bad ideas somehow magically work. Dispelling that notion is why articles like this are important.

(Actually, I suspect they know they’re still bad ideas, but delegating the decisions to an AI lets the humans involved avoid personal blame.)

permalink
report
parent
reply
3 points

Hiring practices are broken from its very basics. The vast majority of businesses consistently discriminate against people who deviate from the norm in presentation, even if the candidate meets the technical requirements or would otherwise be productive, which results in millions of people who are capable of contributing to society being pushed aside.

permalink
report
parent
reply
8 points
*

Again, that’s not really the case.

I have Asian friends that have used these tools and generated headshots that were fine. Just because this one Asian used a model that wasn’t trained for her demographic doesn’t make it a reflection of anything other than the fact that she doesn’t understand how MML models work.

The worst thing that happened when my friends used it were results with too many fingers or multiple sets of teeth 🤣

permalink
report
parent
reply
2 points

No company would use ML to classify who’s the most professional looking candidate.

  1. Anyone with any ML experience at all knows how ridiculous this concept is. Who’s going to go out there and create a dataset matching “proffesional looking scores” to headshots?
  2. The amount of bad press and ridicule this would attract isn’t worth it to any company.
permalink
report
parent
reply
7 points

Companies already use resume scanners that have been found to bias against black sounding names. They’re designed to feedback loop successful candidates, and guess what shit the ML learned real quick?

permalink
report
parent
reply
36 points
Removed by mod
permalink
report
parent
reply
-14 points
Removed by mod
permalink
report
parent
reply
24 points
*
Removed by mod
permalink
report
parent
reply
10 points
*
Removed by mod
permalink
report
parent
reply
2 points
Removed by mod
permalink
report
parent
reply
1 point
Removed by mod
permalink
report
parent
reply
20 points

It still perfectly and visibly demonstrates the big point of criticism in AI: The tendencies the the training material inhibits.

permalink
report
parent
reply
10 points

The AI might associate lighter skin with white person facial structure. That kind of correlation would need to be specifically accounted for I’d think, because even with some examples of lighter skinned Asians, the majority of photos of people with light skin will have white person facial structure.

Plus it’s becoming more and more apparent that AIs just aren’t that good at what they do in general at this point. Yes, they can produce some pretty interesting things, but they seem to be the exception rather than the norm, and in hindsight, a lot of my being impressed with results I’ve seen so far is that it’s some kind of algorithm that is producing that in the first place when the algorithm itself isn’t directly related to the output but is a few steps back from that.

I bet for the instances where it does produce good results, it’s still actually doing something simpler than what it looks like it’s doing.

permalink
report
parent
reply
8 points

This is like a demonstration of lack of self awareness

permalink
report
parent
reply
3 points

Yup

permalink
report
parent
reply
2 points

Almost like we’re looking for things to get mad about.

Also what are these 50 people downvoting you for? Too much nuance I suppose.

permalink
report
parent
reply
-8 points

You said yourself you hate inherent bias yet attempt to justify the result by saying if used again it’s just going to produce another white face.

that’s the problem

It’s a racial bias baked into these AIs based on their training models.

permalink
report
parent
reply
20 points

I doubt it is concious racial bias, it’s most likely that the training data is made up of mostly white people and labeled poorly.

permalink
report
parent
reply
4 points

I also wouldn’t say it was conscious bias either. I don’t think it’s intentionally developed in that way.

The fact still remains though whether conscious or unconscious, it’s potentially harmful to people of other races. Sure it’s an issue with just graphic generation now. What about when it’s used to identify criminals? When it’s used to filter between potential job candidates?

The possibilities are virtually endless, but if we don’t start pointing out and addressing any type of bias, it’s only going to get worse.

permalink
report
parent
reply
14 points

They aren’t justifying anything, they literally said it was about the training data.

permalink
report
parent
reply
99 points

Meanwhile every trained model on Civit.ai produces 12/10 Asian women…

Joking aside, what you feed the model is what you get. Model is trained. You train it on white people, it’s going to create white people, you train it on big titty anime girls it’s not going to produce WWII images either.

Then there’s a study cited that claims Dall-e has a bias when producing images of CEO or director as cis-white males. Think of CEOs that you know. Better yet, google them. It’s shit but it’s the world we live in. I think the focus should be on not having so many white privileged people in the real world, not telling AI to discard the data.

permalink
report
reply
8 points

Yeah there are a lot of cases of claims being made of AI “bias” which is in fact just a reflection of the real world (from which it was trained). Forcing AI to fake equal representation is not fixing a damn thing in the real world.

permalink
report
parent
reply
2 points
*

I recall the being a study of the typical CEO. 6+ feet tall, white males.

But yeah, the output she was getting really depends heavily on the data that whatever model she used was trained on. For someone who is a computer science major, I’m surprised she simply cried “racial bias” rather than investigating the why, and how to get the desired results. Like cranking down the denoising strength.

To me it just seems like she tried messing around with those easy to use, baity websites without really understanding the technology.

permalink
report
parent
reply
1 point
*
Deleted by creator
permalink
report
parent
reply
1 point
Deleted by creator
permalink
report
parent
reply
1 point
*

How did you get from what I wrote to “tearing down” anyone is a bit puzzling. It’s simply about striving to change the status quo and not the AI model representing it. I’m not advocating guillotining Bezos or Musk, hope that’s clear.

permalink
report
parent
reply
1 point
*
Deleted by creator
permalink
report
parent
reply
1 point

Cool let’s just focus on skin color. If you’re white you shouldn’t be in power cause my racism is better than your racism. How about we judge people by their quality of work instead of skin color. I thought that was the whole point.

permalink
report
parent
reply
1 point

Also sure, let’s judge male white CEOs on merit. Let’s start with Elon Musk…

Also I can’t understand why there are people here assuming that the only way to “focus on having less white male CEOs” == eliminating them. This shit is done organically. Eliminating wage gap, providing equal opportunities in education etc.

permalink
report
parent
reply
93 points
*

They should just call AIs “confirmation bias amplifiers”.

permalink
report
reply
27 points

AI learns what is in the data.

permalink
report
parent
reply
-1 points

The AI we have isn’t “learning”. They are pre-trained.

permalink
report
parent
reply
12 points

The “pre-training” is learning, they are often then fine-tuned with additional training (that’s the training that isn’t the ‘pre-training’), i.e. more learning, to achieve specific results.

permalink
report
parent
reply
15 points

Stereotype machines

permalink
report
parent
reply
2 points

Humans will identify sterotypes in AI generated materials that match the dataset.

Assume the dataset will grow and eventually mimic reality.

How will the law handle discrimination based on data supported sterotypes?

permalink
report
parent
reply
3 points

Assume the dataset will grow and eventually mimic reality.

How would that happen, exactly?

Stereotypes themselves and historical bias can bias data. And AI trained on biased data will just learn those biases.

For example, in surveys, white people and black people self-report similar levels of drug use. However, for a number of reasons, poor black drug users are caught at a much higher rate than rich white drug users. If you train a model on arrest data, it’ll learn that rich white people don’t use drugs much but poor black people do tons of drugs. But that simply isn’t true.

permalink
report
parent
reply
-1 points

That’s just stupid and shows a lack of understanding of how this all works.

permalink
report
parent
reply
69 points

This is not surprising if you follow the tech, but I think the signal boost from articles like this is important because there are constantly new people just learning about how AI works, and it’s very very important to understand the bias embedded into them.

It’s also worth actually learning how to use them, too. People expect them to be magic, it seems. They are not magic.

If you’re going to try something like this, you should describe yourself as clearly as possible. Describe your eye color, hair color/length/style, age, expression, angle, and obviously race. Basically, describe any feature you want it to retain.

I have not used the specific program mentioned in the article, but the ones I have used simply do not work the way she’s trying to use them. The phrase she used, “the girl from the original photo”, would have no meaning in Stable Diffusion, for example (which I’d bet Playground AI is based on, though they don’t specify). The img2img function makes a new image, with the original as a starting point. It does NOT analyze the content of the original or attempt to retain any features not included in the prompt. There’s no connection between the prompt and the input image, so “the girl from the original photo” is garbage input. Garbage in, garbage out.

There are special-purpose programs designed for exactly the task of making photos look professional, which presumably go to the trouble to analyze the original, guess these things, and pass those through to the generator to retain the features. (I haven’t tried them, personally, so perhaps I’m giving them too much credit…)

permalink
report
reply
23 points

If it’s stable diffusion img2img, then totally, this is a misunderstanding of how that works. It usually only looks at things like the borders or depth. The text based prompt that the user provides is otherwise everything.

That said, these kinds of AI are absolutely still biased. If you tell the AI to generate a photo of a professor, it will likely generate an old white dude 90% of the time. The models are very biased by their training data, which often reflects society’s biases (though really more a subset of society that created whatever training data the model used).

Some AI actually does try to counter bias a bit by injecting details to your prompt if you don’t mention them. Eg, if you just say “photo of a professor”, it might randomly change your prompt to “photo of a female professor” or “photo of a black professor”, which I think is a great way to tackle this bias. I’m not sure how widespread this approach is or how effective this prompt manipulation is.

permalink
report
parent
reply
3 points

I’ve taken a look at the website for the one she used and it looks like a cheap crap toy. It’s free, which is the first clue that it’s not going to be great.

Not a million miles from the old “photo improvement” things that just run a bunch of simple filters and make over-processed HDR crap.

permalink
report
parent
reply
58 points

Garbage in = Garbage out

ML training data sets are only as good as their data, and almost all data is inherently flawed. Biases are just more pronounced in these models because they scale the bias with the size of the model, becoming more and more noticeable.

permalink
report
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 18K

    Monthly active users

  • 12K

    Posts

  • 553K

    Comments