New York City aims to be first to rein in AI hiring tools


  • AI
  • Monday, 22 Nov 2021

Polli demonstrates AI technology used to assess job skills during an interview with The Associated Press at the Pymetrics headquarters, in New York. — AP

Job candidates rarely know when hidden artificial intelligence tools are rejecting their resumes or analysing their video interviews. But New York City residents could soon get more say over the computers making behind-the-scenes decisions about their careers.

A bill passed by the city council in early November would ban employers from using automated hiring tools unless a yearly bias audit can show they won’t discriminate based on an applicant’s race or gender. It would also force makers of those AI tools to disclose more about their opaque workings and give candidates the option of choosing an alternative process – such as a human – to review their application.

Proponents liken it to another pioneering New York City rule that became a national standard-bearer earlier this century – one that required chain restaurants to slap a calorie count on their menu items.

Instead of measuring hamburger health, though, this measure aims to open a window into the complex algorithms that rank the skills and personalities of job applicants based on how they speak or what they write. More employers, from fast food chains to Wall Street banks, are relying on such tools to speed up recruitment, hiring and workplace evaluations.

“I believe this technology is incredibly positive but it can produce a lot of harms if there isn’t more transparency,” said Frida Polli, co-founder and CEO of New York startup Pymetrics, which uses AI to assess job skills through game-like online assessments. Her company lobbied for the legislation, which favours firms like Pymetrics that already publish fairness audits.

But some AI experts and digital rights activists are concerned that it doesn’t go far enough to curb bias, and say it could set a weak standard for federal regulators and lawmakers to ponder as they examine ways to rein in harmful AI applications that exacerbate inequities in society.

“The approach of auditing for bias is a good one. The problem is New York City took a very weak and vague standard for what that looks like,” said Alexandra Givens, president of the Centre for Democracy & Technology. She said the audits could end up giving AI vendors a “fig leaf” for building risky products with the city's imprimatur.

Givens said it’s also a problem that the proposal only aims to protect against racial or gender bias, leaving out the trickier-to-detect bias against disabilities or age. She said the bill was recently watered down so that it effectively just asks employers to meet existing requirements under US civil rights laws prohibiting hiring practices that have a disparate impact based on race, ethnicity or gender. The legislation would impose fines on employers or employment agencies of up to US$1,500 per violation – though it will be left up to the vendors to conduct the audits and show employers that their tools meet the city's requirements.

The City Council voted 38-4 to pass the bill on Nov 10, giving a month for outgoing Mayor Bill De Blasio to sign or veto it or let it go into law unsigned. De Blasio’s office says he supports the bill but hasn’t said if he will sign it. If enacted, it would take effect in 2023 under the administration of Mayor-elect Eric Adams.

Julia Stoyanovich, an associate professor of computer science who directs New York University’s Center for Responsible AI, said the best parts of the proposal are its disclosure requirements to let people know they’re being evaluated by a computer and where their data is going.

“This will shine a light on the features that these tools are using,” she said.

But Stoyanovich said she was also concerned about the effectiveness of bias audits of high-risk AI tools – a concept that’s also being examined by the White House, federal agencies such as the Equal Employment Opportunity Commission and lawmakers in Congress and the European Parliament.

“The burden of these audits falls on the vendors of the tools to show that they comply with some rudimentary set of requirements that are very easy to meet,” she said.

The audits won’t likely affect in-house hiring tools used by tech giants like Amazon. The company several years ago abandoned its use of a resume-scanning tool after finding it favored men for technical roles – in part because it was comparing job candidates against the company’s own male-dominated tech workforce.

There’s been little vocal opposition to the bill from the AI hiring vendors most commonly used by employers. One of those, HireVue, a platform for video-based job interviews, said in a statement this week that it welcomed legislation that “demands that all vendors meet the high standards that HireVue has supported since the beginning”.

The Greater New York Chamber of Commerce said the city’s employers are also unlikely to see the new rules as a burden.

“It’s all about transparency and employers should know that hiring firms are using these algorithms and software, and employees should also be aware of it,” said Helana Natt, the chamber’s executive director. – AP

Article type: metered
User Type: anonymous web
User Status:
Campaign ID: 1
Cxense type: free
User access status: 0
Subscribe now to our Premium Plan for an ad-free and unlimited reading experience!
   

Next In Tech News

Opinion: AI voices for audiobooks is a strange new trend
How 'crunch' time and low pay are fuelling a union drive among video game workers
Preview: 'Like a Dragon: Ishin!’ brings an old formula into the past
Elon Musk says to attempt Starship launch in March
Advisory firm Innisfree sues Musk's Twitter for $1.9 million in unpaid bills
ChatGPT makers launch tool to spot AI texts – but it's not very good
U.S. judge denies FTC request to stop Meta from acquiring VR firm Within
Two minutes a day to be authentic: Does BeReal show real life?
Opinion: Junk emails can be maddening, but ignoring them is the best option
Lies, racism and AI:�IT experts point to serious flaws in ChatGPT

Others Also Read