‘Racist’ AI can only be stopped if tech bosses share secrets, warns viral selfie app creator

Trevor Paglen and Kate Crawford exposed racist and sexist labels used to train AI
Trevor Paglen and Kate Crawford exposed racist and sexist labels used to train AI Credit: Kate Crawford/Twitter

The spread of "racist" artificial intelligence can only be stopped if Silicon Valley giants share the secret databases used to train them, the creator of a viral selfie app has claimed. 

Trevor Paglen, an artist and one of the pair behind an app that exposed racist and sexist flaws in a colossal database used to train AI, has warned that these same terms could be present in systems developed by big technology companies. 

The flaws could have spread to companies including Google, Microsoft, Facebook and Huawei if they used it as a "seed" database, he has claimed.

"We can assume similar things are going on in the databases of Google and Facebook or whatever, but we can't see that happening," he said.

"For many companies these are the crown jewels of what they can do. They are often trade secrets, and so I think that is a huge problem for the field of machine learning in general, especially in applications that touch people's everyday lives."

He called for "a lot more transparency" from companies on how machine learning systems are being used and how they're classifying people to stop this from making biased decisions.

Mr Paglen's app, created with the AI researcher Kate Crawford and called ImageNet Roulette, exposed that pictures of black and ethnic minority people generated race labels such as “negroid” or “black person”, while results from caucasian faces varied more widely, such as “researcher”, “scientist” or “singer”.

In other words, white people were more likely to be categorised as a specific profession or character type, whereas non-white people were more likely to be categorised by their race alone, sometimes in pejorative terms. 

The app, which was "trained" using a popular image recognition database called ImageNet, was described as “a peek into the politics of classifying humans in machine learning systems and the data they’re trained on”.

ImageNet, created by Stanford University scientists, has been credited with kickstarting the modern AI boom and has become a benchmark against which new image recognition systems are measured.

The team led by Stanford professor Fei-Fei Lin has committed to removing over 600,000 images of people from the database since the challenge went viral earlier this week. 

They welcomed the scrutiny of ImageNet, and claimed it "deserves to be critically examined, in order for the research community to design better collection methods and build better datasets".

Mr Paglen's comments follow plans launched by the UK government to pilot diversity regulations for staff working on artificial intelligence to reduce the risk of sexist and racist computer programs.

License this content