‘Racist’ facial recognition sparks ethical concerns in Russia


Developer guides and code examples unearthed by video surveillance research firm IPVM show software advertised by four of Russia’s biggest facial analytics firms can use AI to classify faces based on their perceived ethnicity or race. — Reuters

TBILISI: From scanning residents’ faces to let them into their building to spotting police suspects in a crowd, the rise of facial recognition is accompanied by a growing chorus of concern about unethical uses of the technology.

A report published on July 5 by US-based researchers showing that Russian facial recognition companies have built tools to detect a person’s race has raised fears among digital rights groups, who describe the technology as “purpose-made for discrimination”.

Developer guides and code examples unearthed by video surveillance research firm IPVM show software advertised by four of Russia’s biggest facial analytics firms can use artificial intelligence (AI) to classify faces based on their perceived ethnicity or race. There is no indication yet that Russian police have targeted minorities using the software developed by the firms – AxxonSoft, Tevian, VisionLabs and NtechLab – whose products are sold to authorities and businesses in the country and abroad.

But Moscow-based AxxonSoft said the Thomson Reuters Foundation’s enquiry prompted it to disable its ethnicity analytics feature, saying in an emailed response it was not interested “in promoting any technologies that could be a basis for ethnic segregation”.

The IPVM findings – seen exclusively by the Thomson Reuters Foundation – have sparked concern among civil rights groups, who say racial profiling is common in the country and note that authorities have already used AI to identify and detain anti-government protesters.

“The findings underline the ugly racism baked into these systems,” said Edin Omanovic, advocacy director at Privacy International, a rights charity based in London.

“Far from being benign security tools which can be abused, such tools are deeply rooted in some of humanity’s most destructive ideas and purpose-made for discrimination.”

The Russian interior ministry and NtechLab did not reply to requests for comment.

The other three companies expressed scepticism about their technology’s current capacity to enable abuse, but said they are aware of ethical concerns related to its use.

Vadim Konushin, CEO of Tevian, also known as Video Analysis Technologies, denied current uses of its tool by police could entrench discrimination, and VisionLabs said its ethnicity analytics software was developed for internal research purposes only.

Ethnicity analytics

Race detection is part of a broadening range of analytics offered by facial recognition companies that allow clients to detect physical features like hair colour and facial expressions, and deduce information such as a person’s age, gender and emotions.

The software developed by AxxonSoft, Moscow-based Tevian and VisionLabs, a Russia-founded firm headquartered in the Netherlands, all categorise people walking past their cameras into roughly the same groups: Asian, Black, white and Indian.

The categories for Axxonsoft’s Face Intellect tool included outdated and offensive terms, which the company put down to a translation error. The terms have now been removed from the company literature. NtechLab, which is partially funded and owned by the Russian government, lists “race” among the features its software can detect.

The findings come as facial recognition firms are feeling increased pressure across the globe over the technology’s potential to undermine human rights and civil liberties.

In June, a group of 50 investors managing more than US$4.5 trillion (RM18.70 trillion) in assets urged companies to make sure their facial recognition products are developed and used in an ethical way.

That same month, more than 170 rights groups signed an open letter calling for a global ban on the use of facial recognition and remote biometric recognition tools that enable mass and discriminatory targeted surveillance.

While ethnicity analytics are developed by dozens of firms worldwide – most prominently in China – Russian companies have a significant presence in the facial recognition market, with reported revenues of up to US$40mil (RM166.24mil) and expansion abroad, according to IPVM.

Rights removed

In Russia, where rights groups say migrants, particularly from Central Asia, are often subjected to racial profiling, arbitrary detentions and violence, authorities’ potential use of race-detection was worrying, said Mahmoudi.

“Minorities who might have been equated with a level of deviance or criminality may fall subject to having their rights, for example their freedom to assemble, removed as a result of the extensive usage of this technology,” he said.

A spokesperson for Moscow’s Department of Technology, which manages the city’s surveillance system, said via email that video analytics were used to improve safety and find offenders.

They did not reply to questions on race detection software, only saying the city uses algorithms from various independent developers.

‘Frightful’ future

AxxonSoft, which is known as ITV in Russia, and Tevian said law enforcement agencies used their products for criminal investigations, usually to find suspects based on descriptions.

NtechLab’s website says its FindFace software underpins Moscow’s facial recognition system, which authorities say has helped cut crime and enforce coronavirus lockdown restrictions.

Some of the Russian firms said they were aware of how their tech could be used to lead to discrimination.

Anton Nazarkin, global sales director at VisionLabs, said that while the tool was marketed on the firm’s website, the company has never licensed it because of ethical and legal concerns, as it might contravene European data protection laws.

Azret Teberdiev, head of marketing for AxxonSoft, whose website says it has more than two million cameras installed worldwide, said ethnicity analytics were included in its product “inadvertently” when integrating third-party software.

The company has pledged to disable the feature after being contacted by the Thomson Reuters Foundation.

“We consider such functionality as not aligning with the ethical standards of our company,” Teberdiev said.

Tevian’s Konushin said while facial recognition technology is still too rudimentary to target specific ethnic minority groups, its potential is worrisome.

“What frightens (me) the most is a theoretical use case, when all people are automatically being racially profiled, and law enforcement officers are signalled to check up on ethnic minorities,” he said.

“If algorithms improve a lot, it could theoretically become possible – which is frightful.” – Thomson Reuters Foundation

Article type: free
User access status:
Join our Telegram channel to get our Evening Alerts and breaking news highlights

Racism

   

Next In Tech News

Why Facebook is really holding off on kids’ Instagram
US researchers aim to identify autism earlier using artificial intelligence
‘The Big Delete’: Inside Facebook’s crackdown in Germany
CD Projekt boss says lesson learned from Cyberpunk debacle
Billionaire behind Europe’s top builder wants to disrupt housing
Database containing personal info of four million Malaysians allegedly being sold online
Alibaba apps start offering WeChat Pay option after government orders
Netflix’s gaming push is coming together, but lacks release date
‘I still can’t understand how he did that’: US teen finds second Apple security bug
Pandemic pushes Chinese tech giants to roll out more courier robots

Stories You'll Enjoy


Vouchers