Our results show that women’s contributions tend to be accepted more often than men’s [when their gender is hidden]. However, when a woman’s gender is identifiable, they are rejected more often. Our results suggest that although women on GitHub may be more competent overall, bias against them exists nonetheless.

You are viewing a single thread.
View all comments
51 points
*

Anyone found the specific numbers of acceptance rate with in comparison to no knowledge of the gender?

On researchgate I only found the abstract and a chart that doesn’t indicate exactly which numbers are shown.

edit:

Interesting for me is that not only women but also men had significantly lower accepance rates once their gender was disclosed. So either we as humans have a really strange bias here or non binary coders are the only ones trusted.

edit²:

I’m not sure if I like the method of disclosing people’s gender here. Gendered profiles had their full name as their user name and/or a photography as their profile picture that indicates a gender.

So it’s not only a gendered VS. non-gendered but also a anonymous VS. indentified individual comparison.

And apparantly we trust people more if we know more about their skills (insiders rank way higher than outsiders) and less about the person behind (pseudonym VS. name/photography).

permalink
report
reply
2 points
*
9 points

Thank you. Unfortunately, your link doesn’t work either - it just leads to the creative commons information). Maybe it’s an issue with Firefox Mobile and Adblockers. I’ll check it out later on a PC.

permalink
report
parent
reply
5 points

Page 15 of the pdf has this chart

(note the vertical axis starts at 60% acceptance rate)

permalink
report
parent
reply
13 points

Looking at their comment history they seem to allways include that link to the CC license page in some attempt to prevent the comments from being used with AI.

I have no idea of if that is actually a thing or just a fad, but that was the link.

permalink
report
parent
reply
2 points

Their link wasn’t to the paper but to the license to poison possible AIs training their models on our posts. Idk if that actually is of any use though

permalink
report
parent
reply
3 points
*

We ass-u-me too much based on people’s genders/photographs/ideas/etc., which taints our objectivity when assessing the quality of their code.

For a close example on Lemmy: people refusing to collaborate with “tankie” devs, with no further insight on whether the code is good or not.

There also used to be code licensed “not to be used for right wing purposes”, and similar.

permalink
report
parent
reply
2 points

That’s not an excellent example. If they refuse to collaborate with them, and also don’t make any claims about the quality of code, then the claim that their objectivity in reviewing code is tainted doesn’t hold.

permalink
report
parent
reply
1 point

Their objectivity is preempted by a subjective evaluation, just like it would be by someone’s appearance or any other perception other than the code itself.

permalink
report
parent
reply
18 points

…or the research is flawed. Gender identity was gained from social media accounts. So maybe it’s a general bias against social media users (half joking).

permalink
report
parent
reply
2 points
*

I think (unless I misunderstood the paper), they only included people who had a Google+ profile with a gender specified in the study at all (this is from 2016 when Google were still trying to make Google+ a thing).

permalink
report
parent
reply
18 points

Thanks for grabbing the chart.

My Stats 101 alarm bells go off whenever I see a graph that does not start with 0 on the Y axis. It makes the differences look bigger than they are.

The ‘outsiders, gendered’ which is the headline stat, shows a 1% difference between women and men. When their gender is unknown there is a 3% difference in the other direction (I’m just eyeballing the graph here as they did not provide their underlying data, lol wtf ). So, overall, the sexism effect seems to be about 4%.

That’s a bit crap but does not blow my hair back. I was expecting more, considering what we know about gender pay gaps, etc.

permalink
report
parent
reply
11 points
*

Perhaps ppl who keep their githubs anonymous simply tend to be more skilled. Older, more senior devs grew up in an era where social medias were anonymous. Younger, more junior devs grew up when social medias expect you to put your real name and face (Instagram, Snapchat, etc.).

More experienced devs are more likely to have their commits accepted than less experienced ones.

permalink
report
parent
reply
1 point

How does a photograph disclose someone’s gender?

permalink
report
parent
reply
2 points

Through gender stereotypes. The observer’s perception is all that counts in this case anyways.

permalink
report
parent
reply
6 points

The study differentiates between male and female only and purely based on physical features such as eye brows, mustache etc.

I agree you can’t see one’s gender but I would say for the study this can be ignored. If you want to measure a bias (‘women code better/worse than men’), it only matters what people believe to see. So if a person looks rather male than female for a majority of GitHub users, it can be counted as male in the statistics. Even if they have the opposite sex, are non-binary or indentify as something else, it shouldn’t impact one’s bias.

permalink
report
parent
reply

Free and Open Source Software

!foss@beehaw.org

Create post

If it’s free and open source and it’s also software, it can be discussed here. Subcommunity of Technology.


This community’s icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

Community stats

  • 677

    Monthly active users

  • 878

    Posts

  • 13K

    Comments