The Guardian view on computers and language: reproducing bias | Editorial

The English language is full of value judgments. These are taken over by the computer algorithms that use it. What can we do about these unconscious biases?

“Reason is and ought only to be the slave of the passions,” wrote David Hume. Language, our instrument of reason, is saturated with value judgments. So what happens when computers – apparently the embodiment of pure mathematical rationality – start to use human language? They reproduce the traces of our passions, of course.

A thorough and elegant experiment reported in the journal Science this week shows this clearly. Researchers analysed a gigantic collection of English texts – more than 840bn instances of 2.3m words – and expressed mathematically how likely different words are to appear in the same contexts. This captures the largely unconscious web of associations around any given word with greater subtlety and fidelity than dictionary definitions can do, since people use words with much greater confidence than they can define them.

This technique also shows the value judgments embedded into ordinary language. Flowers are consistently easier to associate with pleasant concepts than insects; musical instruments than weapons. But others of these unconscious judgments are more worrying and in some cases reflect attitudes we would consciously reject. White-sounding names appear more pleasant than African American ones; young people’s names more than old people’s (a rather weaker effect). Male names are associated with careers, maths, and sciences, while female names are associated with family and with humanities. Bias of this sort both reflects and amplifies existing tendencies in society. This does not mean that it corresponds to underlying, essential inequalities. Different languages encode different systems of prejudice, as does the same language as it changes over time. But changes on that scale happen slowly, and the question meanwhile is what is to be done? Machine learning of this sort is becoming ubiquitous and at the same time much less visible. There is prejudice involved in the way these decisions are made by humans, too, but shuffling responsibility off onto computer programs is, in the phrase of one researcher, “money laundering for bias”.

Some people have proposed that the algorithms become self-correcting. If sexism is expressible as a mathematical relation between words in the system, and so transmitted by invisible algorithms, it could just as well be corrected algorithmically. Something similar already happens with the virtual keyboards on phones, which will write “duck” in place of a more common word. But to extend the principle would involve social engineering on a very large scale, carried out by private companies, not democratic governments. And differing societies, even different governments, could demand that the system reflect very different valuations. One can imagine a Russian government demanding that all the algorithmic associations of a concept like “gay” be tweaked to make them more negative.

The EU has taken a different approach. Next year a law will come into operation in member states which gives everyone a right to an explanation of any decision affecting them that has been reached algorithmically. There are considerable technical and even philosophical difficulties involved in realising any such aim. But computer systems are in some ways easier to interrogate for bias than human beings. We must take advantage of this fact.

Contributor

Editorial

The GuardianTramp

Related Content

Article image
The Guardian view on killer robots: on the loose | Editorial
Editorial: Lethal autonomous weapons are a reality, but the campaign to prevent their use is ours to win

Editorial

29, Aug, 2017 @6:40 PM

Article image
The Guardian view on artificial intelligence: look out, it’s ahead of you | Editorial
Editorial: There is a tendency to see intelligence where it does not exist. But it is just as wrong to fail to see where it is emerging

Editorial

08, May, 2016 @7:41 PM

Article image
The Guardian view on machine learning: people must decide | Editorial
Editorial: Each advance in artificial intelligence increases the power of computer networks, but the responsibility for their use remains with human beings

Editorial

23, Oct, 2016 @6:37 PM

Article image
The Guardian view on digital injustice: when computers make things worse | Editorial
Editorial: Software makes bigger mistakes and faster than humans can. It should not be trusted with vital decisions

Editorial

10, Jun, 2019 @5:32 PM

Article image
AI programs exhibit racial and gender biases, research reveals
Machine learning algorithms are picking up deeply ingrained race and gender prejudices concealed within the patterns of language use, scientists say

Hannah Devlin Science correspondent

13, Apr, 2017 @6:00 PM

Article image
The Guardian view on data protection: a vital check on power | Editorial
Editorial: The UK government plans to harmonise our data protection laws with the EU’s. This is necessary and sensible, too

Editorial

07, Aug, 2017 @7:05 PM

Article image
The Guardian view on the ethics of AI: it’s about Dr Frankenstein, not his monster | Editorial
Editorial: Google’s ethical principles for the use of artificial intelligence are little more than a smokescreen, but they show that many engineers are rightly worried by the possible uses of the technology they’re developing

Editorial

12, Jun, 2018 @5:10 PM

Article image
The Guardian view on the automated future: fewer shops and fewer people | Editorial
Editorial: Low-paid and unskilled jobs in retail will soon be automated away. What will happen to the people?

Editorial

29, Feb, 2016 @6:59 PM

Article image
The Guardian view on robots and humanity: passing Go | Editorial
Editorial: Machines can’t feel or do many of the things that make us human. Sadly that doesn’t dispel the concerns about them putting people out of business

Editorial

09, Mar, 2016 @7:20 PM

Article image
The Guardian view on AI in the NHS: not the revolution you are looking for | Editorial
Editorial: Computer systems may not replace doctors or nurses. But even to replace support staff would be a huge change

Editorial

06, Jan, 2017 @7:02 PM