In May 2019, an agent at the Department of Homeland Security received a
 trove of unsettling images. Found by Yahoo in a Syrian user’s account, 
the photos seemed to document the sexual abuse of a young girl. One 
showed a man with his head reclined on a pillow, gazing directly at the 
camera. The man appeared to be white, with brown hair and a goatee, but 
it was hard to really make him out; the photo was grainy, the angle a 
bit oblique. The agent sent the man’s face to child-crime investigators 
around the country in the hope that someone might recognize him.
	When an investigator in New York saw the request, she ran the face 
through an unusual new facial-recognition app she had just started 
using, called Clearview AI. The team behind it had scraped the public 
web — social media, employment sites, YouTube, Venmo — to create a 
database with three billion images of people, along with links to the 
webpages from which the photos had come. This dwarfed the databases of 
other such products for law enforcement, which drew only on official 
photography like mug shots, driver’s licenses and passport pictures; 
with Clearview, it was effortless to go from a face to a Facebook 
account. 
	The app turned up an odd hit: an Instagram photo of a heavily muscled 
Asian man and a female fitness model, posing on a red carpet at a 
bodybuilding expo in Las Vegas. The suspect was neither Asian nor a 
woman. But upon closer inspection, you could see a white man in the 
background, at the edge of the photo’s frame, standing behind the 
counter of a booth for a workout-supplements company. That was the 
match. On Instagram, his face would appear about half as big as your 
fingernail. The federal agent was astounded.
	The agent contacted the supplements company and obtained the booth 
worker’s name: Andres Rafael Viola, who turned out to be an Argentine 
citizen living in Las Vegas. Another investigator found Viola’s Facebook
 account. His profile was public; browsing it, the investigator found 
photos of a room that matched one from the images, as well as pictures 
of the victim, a 7-year-old. Law-enforcement officers arrested Viola in 
June 2019. He later pleaded guilty to sexually assaulting a child and 
producing images of the abuse and was sentenced to 35 years in prison. (Viola’s lawyer did not respond to multiple requests for comment.) 
	At the time, the use of Clearview in Viola’s case was not made public; I
 learned about it recently, through court documents, interviews with 
law-enforcement officials and a promotional PowerPoint presentation that Clearview made.
 The case represented the technology’s first use on a child-exploitation
 case by Homeland Security Investigations, or H.S.I., which is the 
investigative arm of Immigrations and Customs Enforcement. (Such crimes 
fall under the agency because, pre-internet, so much abuse material was 
being sent by mail internationally.) “It was an interesting first foray 
into our Clearview experience,” said Erin Burke, chief of H.S.I.’s Child
 Exploitation Investigations Unit. “There was no way we would have found
 that guy.”
	Few outside law enforcement knew of Clearview’s 
existence back then. That was by design: The government often avoids 
tipping off would-be criminals to cutting-edge investigative techniques,
 and Clearview’s founders worried about the reaction to their product. 
Helping to catch sex abusers was clearly a worthy cause, but the 
company’s method of doing so — hoovering up the personal photos of 
millions of Americans — was unprecedented and shocking. Indeed, when the
 public found out about Clearview last year, in a New York Times article I wrote, an immense backlash ensued.
	Facebook, LinkedIn, Venmo and Google sent cease-and-desist letters to 
the company, accusing it of violating their terms of service and 
demanding, to no avail, that it stop using their photos. BuzzFeed 
published a leaked list of Clearview users,
 which included not just law enforcement but major private organizations
 including Bank of America and the N.B.A. (Each says it only tested the 
technology and was never a client.) I discovered that the company had 
made the app available to investors, potential investors and business 
partners, including a billionaire who used it to identify his daughter’s date when the couple unexpectedly walked into a restaurant where he was dining.
	Computers once performed facial recognition rather imprecisely, by 
identifying people’s facial features and measuring the distances among 
them — a crude method that did not reliably result in matches. But 
recently, the technology has improved significantly, because of advances
 in artificial intelligence. A.I. software can analyze countless photos 
of people’s faces and learn to make impressive predictions about which 
images are of the same person; the more faces it inspects, the better it
 gets. Clearview is deploying this approach using billions of photos 
from the public internet. By testing legal and ethical limits around the
 collection and use of those images, it has become the front-runner in 
the field.
	After Clearview’s activities came to light, Senator Ed Markey of 
Massachusetts wrote to the company asking that it reveal its 
law-enforcement customers and give Americans a way to delete themselves from Clearview’s database.
 Officials in Canada, Britain, Australia and the European Union 
investigated the company. There were bans on police use of facial 
recognition in parts of the United States, including Boston and 
Minneapolis, and state legislatures imposed restrictions on it, with 
Washington and Massachusetts declaring that a judge must sign off before
 the police run a search.
	In Illinois and Texas, companies already had to obtain consent from 
residents to use their “faceprint,” the unique pattern of their face, 
and after the Clearview revelations, Senators Bernie Sanders and Jeff 
Merkley proposed a version of Illinois’s law for the whole country. 
California has a privacy law giving citizens control over how their data
 is used, and some of the state’s residents invoked that provision to 
get Clearview to stop using their photos. (In March, California 
activists filed a lawsuit in state court.) Perhaps most significant, 10 
class-action complaints were filed against Clearview around the United 
States for invasion of privacy, along with lawsuits from the A.C.L.U. 
and Vermont’s attorney general. “This is a company that got way out over
 its skis in an attempt to be the first with this business model,” 
Nathan Freed Wessler, one of the A.C.L.U. lawyers who filed the 
organization’s lawsuit, in Illinois state court, told me.
	It seemed entirely possible that Clearview AI would be sued, legislated or shamed out of existence. But that didn’t happen....