Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It makes sense to me? The algorithm specialises in distinguishing between the faces in its training set. It works by dimensionality reduction. If there aren't many black faces there it can just dedicate a few of its dimensions to "distinguishing black face features".

Then if you give it a task that only contains black faces, most of the dimensions will go unused.



Are black faces overrepresented or underrepresented? According to AI researchers, we're faced with Schrodinger's Mugshot--there's simultaneously too many and too few!


It's phrased accurately if confusingly. The bigger and un-fixable problem is that people are more apt to believe that a computer has calculated the correct answer when by its very nature popping oft bad images into a facial recognition search is almost always going to produce results even if most are fake and the real ID may not even be among the results.

Without additional leads police are strongly incentivized to pick one of the results and run with it and in many cases with enough data you have enough to get a plea or conviction even if they didn't do it especially if the person so selected was in the database in the first place because they have a record.

Convictions/pleas are obtained all the time with similar levels of proof.

This is fundamentally the same problem as dragnet searches of phone GPS to see who was in a space in a range of time. It could be a valuable investigative tool but its also a great way to "solve" a crime by finding someone to pin it on.


What makes you think mugshots were used as the training data?


Because models are trained and validated on real data. Given a training set of crimes and corresponding surveillance footage, arrestee info is a (not noisy) label for “who is the guy in the movie.”


There's two datasets but your conviction that there's some political undertone to this is leaving you unable to process basic logic.


With a moment's thought, even the most emotive amongst us should see that the mugshots will be part of the training set--the photographed individuals are, after all, the class of true positives.


You train a model on a bunch of photos of white people, and a few photos of black people.

You then deploy that model, and use the model to match black person detained by racist officers against a database of photos that the police have from before. In that database the majority of people are black.

Shitty AI that was not properly taught what black people look like because most of the people in the training data were white, says that it found a probable match for detained black person.

Racist officers do not attempt to second guess the computer, so they throw innocent black person into their car and drive off to the police station.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: