Researchers criticise AI software that predicts emotions

Reading Time: 2 minutes

A prominent group of researchers alarmed by the harmful social effects of artificial intelligence called for a ban on automated analysis of facial expressions in hiring and other major decisions. The AI Now Institute at New York University said action against such software-driven “affect recognition” was its top priority because science doesn’t justify the technology’s use and there is still time to stop widespread adoption.

The group of professors and other researchers cited as a problematic example the company HireVue, which sells systems for remote video interviews for employers such as Hilton and Unilever. It offers AI to analyse facial movements, tone of voice and speech patterns, and doesn’t disclose scores to the job candidates.

The non-profit Electronic Privacy Information Centre has filed a complaint about HireVue to the United States. Federal Trade Commission, and AI Now has criticised the company before.

HireVue said it had not seen the AI Now report and did not answer questions on the criticism or the complaint.

“Many job candidates have benefited from HireVue’s technology to help remove the very significant human bias in the existing hiring process,” said spokesperson Kim Paone.

AI Now, in its fourth annual report on the effects of artificial intelligence tools, said job screening is one of many ways in which such software is used without accountability and typically favoured privileged groups.

The report cited a recent academic analysis of studies on how people interpret moods from facial expressions.

“How people communicate anger, disgust, fear, happiness, sadness, and surprise varies substantially across cultures, situations, and even across people within a single situation,” wrote a team at North-eastern University and Massachusetts General Hospital.

Companies including Microsoft Corp are marketing their ability to classify emotions using software, the study said. Microsoft did not respond to a request for comment Wednesday evening.

AI Now also criticised Amazon.com Inc, which offers analysis on expressions of emotion through its Rekognition software. Amazon told Reuters that its technology only makes a determination on the physical appearance of someone’s face and does not claim to show what a person is actually feeling.

In a conference call ahead of the report’s release, AI Now founders Kate Crawford and Meredith Whittaker said that damaging uses of AI are multiplying despite broad consensus on ethical principles because there are no consequences for violating them.