+

Cookies on the Business Insider India website

Business Insider India has updated its Privacy and Cookie policy. We use cookies to ensure that we give you the better experience on our website. If you continue without changing your settings, we\'ll assume that you are happy to receive all cookies on the Business Insider India website. However, you can change your cookie setting at any time by clicking on our Cookie Policy at any time. You can also see our Privacy Policy.

Close
HomeQuizzoneWhatsappShare Flash Reads
 

Clearview AI scraped billions of photos from social media to build a facial recognition app that can ID anyone - here's everything you need to know about the mysterious company

Mar 6, 2020, 23:09 IST
Alex Wong/Getty Images
  • A software startup that scraped billions of images from major web services - including Facebook, Google, and YouTube - is selling its tool to law enforcement agencies across the United States.
  • The app can be loaded onto smartphones, and used to instantly identify unknown people. The point of the software is to match unknown faces with publicly available photos, thus identifying crime suspects.
  • But the startup, Clearview AI, has faced major criticism for the way it obtains images: By taking them without permission from major services like Facebook, Twitter, and YouTube.
  • Moreover, despite Clearview's stated goal of working with law enforcement, several reports point to a far wider clientele - including a string of billionaire investors, the founder's friends, and retailers ranging from Walmart to Macy's.
  • Visit Business Insider's homepage for more stories.

Police departments across the United States are paying tens of thousands of dollars apiece for access to software that identifies faces using images scraped from major web platforms like Google, Facebook, YouTube, and Twitter.

The software is produced by a company named Clearview AI - a relatively unknown tech startup supported by a slew of somewhat better-known investors: From early Facebook backer Peter Thiel to Texas-based investor Hal Lambert, who's most notable for running an investment fund with the "MAGA" ticker symbol.

So, what does the software do? It identifies people using images scraped from the web and social media platforms, without permission, to create a searchable database. If you want to identify someone, you simply upload a photo or snap a new one, and Clearview's software attempts to make a match.

Photos of you, photos of friends and family - all of it is scraped and saved by Clearview AI. That searchable database is then sold to police departments and federal agencies, Clearview says, but additional reports indicate that the company has also given access to other clients, including billionaires, retail chains like Walmart and Macy's, the NBA, and even some high schools.

Advertisement

As Clearview AI's software and business have become more clear, controversy has followed. Here's the full rundown on Clearview AI, its software, and who the company works with:

Clearview AI went from unknown startup to controversy in mid January when the New York Times ran an exposé.

In mid-January, Clearview AI went from unknown startup to the star of its very own New York Times exposé.

"The Secretive Company That Might End Privacy as We Know It" was the title of the piece, and it revealed the stunning details of what Clearview's tech could do: "You take a picture of a person, upload it and get to see public photos of that person, along with links to where those photos appeared. The system — whose backbone is a database of more than three billion images that Clearview claims to have scraped from Facebook, YouTube, Venmo and millions of other websites — goes far beyond anything ever constructed by the United States government or Silicon Valley giants."

It was the first look the public got at a company that, until then, was operating in secrecy.

What the Times piece revealed, beyond the functionality of Clearview's tools, was stunning: The company had scraped billions of publicly available image from major social media platforms like Twitter, Facebook, and YouTube. Moreover, it put those images into a searchable database then sold that tool to American law enforcement.

"More than 600 law enforcement agencies have started using Clearview in the past year," the piece pointed out. Contracts to use the service cost as much as $50,000 for a two-year deal.

Soon after the piece ran, social media giants began sending cease-and-desist letters to Clearview AI.

In the initial report on Clearview, a slew of major tech platforms were named as targets for their scraping efforts: Facebook, Twitter, YouTube, and Venmo.

Clearview mined each of those platforms for user photos, and then added them to Clearview's database, which it sells. And that process — lifting user photos from social platforms, then selling those photos — breaches the terms of service of every platform from which the photos were taken.

"YouTube's Terms of Service explicitly forbid collecting data that can be used to identify a person," YouTube spokesperson Alex Joseph told Business Insider in an email in early February. "Clearview has publicly admitted to doing exactly that, and in response we sent them a cease and desist letter."

Similar sentiments were shared by all the major social platforms, from Facebook to Twitter to LinkedIn and Venmo.

Clearview AI CEO Hoan Ton-That defended the company in an interview on "CBS This Morning."

As major tech companies openly pushed back against Clearview's method for building an image database, Clearview's chief executive, Hoan Ton-That, went on the defensive. He argued that his company's software isn't doing anything illegal, and doesn't need to delete any of the images it has stored, because it's protected under US law.

"There is a First Amendment right to public information," he told CBS This Morning in an interview published in early February. "The way that we have built our system is to only take publicly available information and index it that way."

As for his response to the cease-and-desist letters? "Our legal counsel has reached out to them, and are handling it accordingly."

Clearview AI's lawyer, Tor Ekeland, told Business Insider in an emailed statement, "Clearview is a photo search engine that only uses publicly available data on the Internet. It operates in much the same way as Google's search engine. We are in receipt of Google and YouTube's letter and will respond accordingly."

Check out the full interview with Clearview AI CEO Hoan Ton-That right here:

Then, in late February, Clearview disclosed a stunning error: The company's entire client list was leaked in a data breach, which led to the revelation that Clearview was selling its services to lots of clients outside of law enforcement.

Due to a "flaw" in Clearview AI's security measures, the company's entire list of clients was leaked, the company confirmed in late February.

As if that weren't problematic enough, the list of clients included some particularly notable institutions that fall pretty far outside the realm of law enforcement: Macy's, Kohl's, Walmart, and the NBA, among others, Buzzfeed News first reported.

The full client list spells out just how many people have access to Clearview's tech — people in more than 2,200 law enforcement departments, government agencies, and companies across 27 countries.

This directly contradicts how Clearview describes itself: "Clearview is a new research tool used by law enforcement agencies to identify perpetrators and victims of crimes." On Clearview's website, applying for the software means clicking a "Request Access" button directly below a sign that reads, "Available now for Law Enforcement."

Then, in March, another New York Times piece on the company revealed another stunning detail: The company's founders casually gave access to the software to potential investors and friends, who immediately abused it.

When John Catsimatidis was finishing dinner in October 2018 at Cipriani in downtown Manhattan, he spotted something amiss: His daughter was also eating dinner there, on a date with an unknown man.

"I wanted to make sure he wasn't a charlatan," Catsimatidis, the billionaire owner of the Gristedes chain of supermarkets, told The New York Times.

He asked the waiter to snap a photo of the man without their knowing, then used his smartphone to instantly identify him using a secretive facial-recognition app. He then texted the man's biography to his daughter.

"My date was very surprised," his daughter, Andrea, said.

And indeed he should have been; John Catsimatidis was using Clearview AI's software — supposedly intended for law enforcement — as a way to freak out his daughter.

According to The Times, Catsimatidis was one of several prospective investors who were given access to the app; he said he had access through a friend who cofounded the company. Peter Thiel, David Scalzo, Hal Lambert, and the actor turned investor Ashton Kutcher were also listed in the report as either having access or being suspected of having access to the app.

You are subscribed to notifications!
Looks like you've blocked notifications!
Next Article