Lawmakers want US intelligence assessment on fake videos

  • TECH
  • Friday, 14 Sep 2018

This technology uses facial mapping and artificial intelligence to produce videos that appear so genuine it's hard to spot the phonies.

A bipartisan group of lawmakers asked for a US intelligence assessment of the threat posed by technology that lets anyone make fake, but realistic, videos of real people saying things they’ve never said.

The rising capabilities of the technology are fueling concerns it could be used to make a bogus video, for example, of an American politician accepting a bribe or of a US or an adversarial foreign leader warning of an impending disaster.

Three lawmakers wrote a letter to National Intelligence Director Dan Coats asking his office to assess how these bogus, high-tech videos – known as deepfakes – could threaten US national security.

“By blurring the line between fact and fiction, deepfake technology could undermine public trust in recorded images and videos as objective depictions of reality,” wrote Adam Smith, Stephanie Murphy, and Carlos Curbelo.

“We are deeply concerned that deepfake technology could soon be deployed by malicious foreign actors.”

Deepfakes are not lip-syncing videos that are obvious spoofs. This technology uses facial mapping and artificial intelligence to produce videos that appear so genuine it’s hard to spot the phonies. Republicans and Democrats predict this high-tech way of putting words in someone’s mouth will become the latest weapon in disinformation wars against the United States and other Western democracies.

The lawmakers asked the intelligence agencies to submit a report to Congress by mid-December describing the threat and possible counter measures the US can develop or employ to protect the nation.

Realising the implications of the technology, the US Defense Advanced Research Projects Agency is already two years into a four-year programme to develop technologies that can detect fake images and videos. Right now it takes extensive analysis to separate phony videos from the real thing. It’s unclear if new ways to weed out the fakes will keep pace with technology used to make them.

Deepfakes are so named because they utilise deep learning, a form of artificial intelligence. They are made by feeding a computer an algorithm, or set of instructions, lots of images and audio of a certain person. The computer program learns how to mimic the person’s facial expressions, mannerisms, voice and inflections. If you have enough video and audio of someone, you can combine a fake video of the person with a fake audio and get them to say anything you want. – AP

Article type: metered
User Type: anonymous web
User Status:
Campaign ID: 1
Cxense type: free
User access status: 3
Join our Telegram channel to get our Evening Alerts and breaking news highlights

Next In Tech News

Brazil's Nubank, last valued at $30 billion, files for U.S. IPO
EBay revenue outlook disappoints as pandemic-led boom fades
Facebook asks employees to preserve internal documents for legal inquiries
This electric multicopter can fly at speeds of over 100 km/h
Justice Department probes Visa's relationships with fintech companies - WSJ
White House not surprised Facebook knew about vaccine misinformation on its platform
U.S. senators call for passing laws to protect online privacy for kids
U.S. consumer watchdog to review Big Tech data, promote lending competition
Enterprise software firm Informatica valued at $7.5 billion in stock market return
WinkyVerse, the metaverse teaching kids about the technologies of the future

Others Also Read