0.7 C
New York
Wednesday, December 25, 2024
HomeTechnologyResearchers criticize AI software that predicts emotions

Researchers criticize AI software that predicts emotions

Date:

Related stories

Understanding Instagram’s Role in Rising Youth Anxiety

In the digital age, social media platforms like Instagram...

WhatsApp to bring voice and video calls to desktop next year

WhatsApp will introduce voice and video calling feature to...

UK’s sole hydrogen car maker bets on green revolution

Hydrogen-powered car manufacturer Riversimple is hoping to steal a...

Twitter plans to bring back ‘blue tick’ next year

Twitter on Tuesday released plans for its new policy...

A prominent group of researchers alarmed by the harmful social effects of artificial intelligence called Thursday for a ban on automated analysis of facial expressions in hiring and other major decisions.

The AI Now Institute at New York University said action against such software-driven “affect recognition” was its top priority because science doesn’t justify the technology’s use and there is still time to stop widespread adoption.

The group of professors and other researchers cited as a problematic example the company HireVue, which sells systems for remote video interviews for employers such as Hilton and Unilever. It offers AI to analyze facial movements, tone of voice and speech patterns, and doesn’t disclose scores to the job candidates.

The nonprofit Electronic Privacy Information Center has filed a complaint about HireVue to the U.S. Federal Trade Commission, and AI Now has criticized the company before.

HireVue said it had not seen the AI Now report and did not answer questions on the criticism or the complaint.

- Advertisement -

“Many job candidates have benefited from HireVue’s technology to help remove the very significant human bias in the existing hiring process,” said spokeswoman Kim Paone.

AI Now, in its fourth annual report on the effects of artificial intelligence tools, said job screening is one of many ways in which such software is used without accountability, and typically favored privileged groups.

The report cited a recent academic analysis of studies on how people interpret moods from facial expressions. That paper found that the previous scholarship showed such perceptions are unreliable for multiple reasons.

“How people communicate anger, disgust, fear, happiness, sadness, and surprise varies substantially across cultures, situations, and even across people within a single situation,” wrote a team at Northeastern University and Massachusetts General Hospital.

Companies including Microsoft Corp are marketing their ability to classify emotions using software, the study said. Microsoft did not respond to a request for comment Wednesday evening.

AI Now also criticized Amazon.com Inc, which offers analysis on expressions of emotion through its Rekognition software. Amazon told Reuters that its technology only makes a determination on the physical appearance of someone’s face and does not claim to show what a person is actually feeling.

In a conference call ahead of the report’s release, AI Now founders Kate Crawford and Meredith Whittaker said that damaging uses of AI are multiplying despite broad consensus on ethical principles because there are no consequences for violating them.

Subscribe

- Never miss a story with notifications

- Gain full access to our premium content

- Browse free from up to 5 devices at once

Latest stories

LEAVE A REPLY

Please enter your comment!
Please enter your name here