Amazon Reportedly Killed an AI Recruitment System

SOS

Is alive.
Wackbag Staff
Aug 14, 2000
48,309
8,909
1,038
USA
#1
http://fortune.com/2018/10/10/amazon-ai-recruitment-bias-women-sexist/

Amazon Reportedly Killed an AI Recruitment System Because It Couldn't Stop the Tool from Discriminating Against Women
By David Meyer October 10, 2018

Machine learning, one of the core techniques in the field of artificial intelligence, involves teaching automated systems to devise new ways of doing things, by feeding them reams of data about the subject at hand. One of the big fears here is that biases in that data will simply be reinforced in the AI systems—and Amazon seems to have just provided an excellent example of that phenomenon.

According to a new Reuters report, Amazon spent years working on a system for automating the recruitment process. The idea was for this AI-powered system to be able to look at a collection of resumes and name the top candidates. To achieve this, Amazon fed the system a decade’s worth of resumes from people applying for jobs at Amazon.

The tech industry is famously male-dominated and, accordingly, most of those resumes came from men. So, trained on that selection of information, the recruitment system began to favor men over women.

According to Reuters’ sources, Amazon’s system taught itself to downgrade resumes with the word “women’s” in them, and to assign lower scores to graduates of two women-only colleges. Meanwhile, it decided that words such as “executed” and “captured,” which are apparently deployed more often in the resumes of male engineers, suggested the candidate should be ranked more highly.

The team tried to stop the system from taking such factors into account, but ultimately decided that it was impossible to stop it from finding new ways to discriminate against female candidates. There were apparently also issues with the underlying data that led the system to spit out rather random recommendations.

And so, Amazon reportedly killed the project at the start of 2017.

“This was never used by Amazon recruiters to evaluate candidates,” Amazon said in a statement.

Amazon isn’t the only company to be alert to the problem of algorithmic bias. Earlier this year, Facebook said it was testing a tool called Fairness Flow, for spotting racial, gender or age biases in machine-learning algorithms. And what was the first target for Facebook’s tests of the new tool? Its algorithm for matching job-seekers with companies advertising positions.

This article was updated to include Amazon’s statement.
 

Bobobie

Registered User
Oct 1, 2005
11,883
2,155
651
Bay City Michigan
#3
Also that pesky science thing with genetics, Race, and sex. You have to ignore all that cuz racism, misogyny, sexism, and other 'isms. Isn't strange that ideologues , both the religious and the Lefties, seem to dislike science?
 

THE FEZ MAN

as a matter of fact i dont have 5$
Aug 23, 2002
43,380
10,020
848
#6

Bobobie

Registered User
Oct 1, 2005
11,883
2,155
651
Bay City Michigan
#8
This isn't AI, this just software with algorithms doing " if this then that " stuff . It's the same shit that drives google search and Echo.
 

THE FEZ MAN

as a matter of fact i dont have 5$
Aug 23, 2002
43,380
10,020
848
#9
This isn't AI, this just software with algorithms doing " if this then that " stuff . It's the same shit that drives google search and Echo.
And flags “mean words” on hugbook