Back to News & Commentary

What鈥檚 Wrong With Airport Face Recognition?

CBP officer at customs comparing IDs with woman's real face
CBP officer at customs comparing IDs with woman's real face
Jay Stanley,
Senior Policy Analyst,
老澳门开奖结果 Speech, Privacy, and Technology Project
Share This Page
August 4, 2017

U.S. Customs and Border Protection (CBP) has launched a 鈥淭raveler Verification Service鈥 (TVS) that envisions applying face recognition to all airline passengers, including U.S. citizens, boarding flights exiting the United States. This system raises very serious privacy issues.

What we know about this program comes from a DHS issued on the program, and a CBP Deputy Executive Assistant Commissioner John Wagner gave to privacy advocates in Washington this week. CBP鈥檚 plan is to install cameras at boarding gates to take photos of, and apply face recognition to, all cross-border passengers at the boarding gate to their aircraft. Currently being operated in six airports around the country (Boston Logan, New York JFK, Dulles in D.C., Hartsfield-Jackson in Atlanta, Chicago O鈥橦are, and Bush in Houston), the TVS program is part of a larger program called 鈥淏iometric Entry/Exit.鈥 That program is DHS鈥檚 attempt to comply with a that the agency use biometrics to keep track of visitors entering and exiting the United States, in order to identify individuals who overstay their visas.

The way the system works is that before departure, CBP obtains the passenger manifest for each flight, and then reaches through the government鈥檚 extensive, interconnected set of databases to assemble photographs on each passenger. Those include passport and visa photos as well as photos 鈥渃aptured by CBP during the entry inspection鈥 and 鈥渇rom other DHS encounters.鈥 The agency then compares face recognition templates (essentially, patterns) derived from those database photos to templates derived from live photographs taken by a camera at the boarding gate.

There are a number of very serious problems with this program from a privacy standpoint:

  • It utilizes the most dangerous biometric: face recognition. While Congress has directed CBP to collect biometrics from noncitizens as part of the entry/exit program, Congress did not specify which biometric the agency should use, and from a privacy perspective, face recognition is (along with iris recognition) the most dangerous biometric to use. That鈥檚 because it has greater potential for expansion and misuse: for example, you can subject thousands of people an hour to face recognition when they鈥檙e walking down the sidewalk without their knowledge, let alone permission or participation. You can鈥檛 do that with fingerprints. Face recognition databases could be plugged in to every surveillance camera in America, creating a giant infrastructure for government tracking and control. Wagner told me that the agency opted for face recognition instead of fingerprints because of the greater ease and practicality of the technology as well as the 鈥渙ptics of us taking fingerprints from people.鈥 Of course, fingerprints do have a negative association in the public mind鈥攂ut that鈥檚 because of their use in tracking and identifying accused criminals. And tracking and identifying is exactly what the photos are being used for here. If, as Wagner suggests, taking a photo seems more benign to the public, that鈥檚 only because the public鈥檚 intuitions about privacy have not caught up with what the technology can do. And fingerprints work fine in the context of international travel, as they are already used for the Global Entry frequent traveler program.
  • It normalizes face recognition as a checkpoint technology. Security technologies that are applied only at airports because of heightened government concerns about the security of air travel tend, over time, to expand outward into society. Magnetometers, for example, spread from airports to a wide variety of venues, including sports stadiums, government buildings, and even some high schools. That dynamic takes place partly because it socializes people to accept such technologies as normal and acceptable, and partly because government agencies and others push it outward in a futile quest for perfect security everywhere. Wagner said of face recognition, 鈥淚 think this is where the technology is headed.鈥 But 鈥渢he technology鈥 is not an autonomous, inevitable force; we as a society are in control, and can choose what to deploy and not to deploy. And we should not want to turn into a checkpoint society, where we are subject to ceaseless status and identity checks at every turn, constantly monitoring, evaluating, and sorting citizens into 鈥済o鈥 and 鈥渘o-go鈥 categories. The ease of implementing face recognition makes that all-too-real a threat.
  • It will inevitably be subject to mission creep. Once CBP begins collecting biometrics from every person traveling across the border, including Americans, there is a significant likelihood that that practice will expand not only to new places but also for new purposes. For the moment, CBP says it will delete the live photos captured at the gate within 14 days for citizens, and that it only uses them to verify identity by comparing them with the database photos. But customs officials have already talked about dropping that restriction, only that 鈥渇or now, we鈥檙e discarding that information.鈥 How long before CBP begins holding them for longer periods of time, and using them for new purposes? Got a group of photos of wanted bank robbers, drug dealers, or, for that matter, reckless drivers? Why not run passengers鈥 photos against those databases and maybe catch a few?
  • Face recognition has a major reliability problem. The fact is, people鈥檚 faces don鈥檛 stay the same, and people look like each other鈥攏ot only identical twins but also . Studies also show that face recognition suffers when trying to match the faces of African Americans, raising the prospect of yet another racial injustice in our society. There are also higher error rates for women and children. Even changes in expression can render the technology inaccurate, which is why Americans are not allowed to smile in their passport photos. Wagner said that tests so far have found a 4% false negative rate. That means one in 25 people will be told by the machine, 鈥渟orry, you鈥檙e not who you claim to be.鈥 Those people will then be sent to a CBP officer at the gate for visual comparison with their passport photo (American citizens) or fingerprints or other checks (visitors). But if this program scales up, the logistics of making a CBP officer available at every gate to examine that 4% of rejected passengers would be a major obstacle. On the other hand, Wagner suggested the agency might let airline workers do the check, which means zero change from how things are already being done. Overall, it seems very strange for the government to be going all-in on a technology with such a high inaccuracy rate. And Wagner鈥檚 4% number was provided without evidence; the government should make its face-matching algorithms public so that independent studies can be made of their reliability, including ethnic and other differences in that reliability.
  • It is fundamentally unnecessary and wastes taxpayer money. In the United States, all arriving passengers pass through a CBP checkpoint, but there is no infrastructure of CBP checkpoints through which departing passengers must pass, as there is in Europe and a number of other countries. That means creating a system to collect biometrics from exiting travelers is a hugely complex and expensive new enterprise. That is one reason why DHS has in the past Congress鈥檚 naive obsession with using sexy biometrics technology and argued that a full biometric exit tracking system is unnecessary, because using just biographic data (name, DOB, etc) from the information systems DHS already has in place to track exactly who is on an aircraft was enough to satisfy Congress鈥檚 goal of tracking visa overstays. As with any technology, boosters can always cite scenarios where biometrics will prevent problems, such as imposters flying under another person鈥檚 name. But (as with any technology) the proper question is how widespread and how harmful are those scenarios, and what are the costs and downsides of measures to prevent it. DHS has already spent billions on programs such as Real ID supposedly to prevent imposters, and in light of the downsides discussed above, the cost-benefit calculus here makes no sense. Since face recognition is highly unreliable, the only thing the billions spent on this program will achieve is whatever marginal improvement in detection there may be between machine and human effectiveness in matching faces鈥攁 highly uncertain benefit. Especially since for 4% of travelers the matching will probably be done the way it already is鈥攎anually by a human being.
  • This system is being built in the context of an agency with a troubling record. This is not a technology being deployed by an agency with a history of behaving well. CBP has a terrible track record of use-of-force and other incidents of abuse, and external reports have found a continuing pattern of poor oversight and training of agents as well as a 鈥culture of impunity and violence.鈥 It is an agency that lacks oversight, due process, and transparency, and鈥攖hough it is our nation鈥檚 largest police force鈥攔efuses to be held accountable to basic 21st century police best practices. You would think that the agency charged with guarding our borders would not harass people who are leaving the country, but you would . The creation of an institutional CBP presence where there has never been one before (at the gates of departing aircraft) raises the prospect that the kinds of abuses we have seen at other borders will spread to this new context, and people will be unfairly sanctioned without the kind of due process that they would normally receive, such as the right to go before a judge. For example, someone wrongly suspected of a visa overstay may be pressured into signing papers by an agent at the gate that results in a 10-year ban from returning to the United States. The system, Wagner said, will include the ability for CBP to tag someone for various reasons for official intervention at the gate.

Given all of these serious social implications, one of the biggest problems with the program is that it hasn鈥檛 been authorized by Congress. Considering the significance of applying face recognition to the entire American cross-border traveling population, we should expect CBP to subject the program to the full democratic process. To the contrary, as Harrison Rudolph from the Georgetown Center on Privacy & Technology , 鈥淐ongress has passed Biometric Exit bills at least nine times. In each, it has been clear: This is a program meant for foreign nationals. In fact, when President Trump issued an executive order in January on Biometric Exit, it was actually reissued to clarify that it didn鈥檛 apply to American citizens.鈥 Unfortunately, based on the briefing provided to advocates, CBP seems to believe that even a nationwide rollout of their pilot wouldn鈥檛 require separate Congressional authorization. But this program is clearly not what Congress contemplated, and should be subject to a vigorous public debate that allows members of Congress to weigh the privacy impact and other costs of the new proposal against its purported benefits.

It should be noted that the narrow data-privacy impact of this system is different than some other face recognition deployments. When the technology is deployed, say, on the street, it may collect several kinds of information that its operator does not already possess: (1) photographs of subjects鈥 faces; (2) their identity, if those photographs can be matched to others; and (3) the time and place where they were seen. In the case of this program focused on international travelers, however, CBP already has all of that information. They have access to the passenger manifests and already know exactly who is departing the country on each flight. And, they already have access to a photograph of each traveler. CBP does run each traveler against an opaque computerized risk assessment engine called the Automated Targeting System, which we at the 老澳门开奖结果 have been criticizing for years鈥攂ut it does that irrespective of whether a passenger鈥檚 photo is taken at the gate.

That said, as we have seen there are plenty of significant and problematic societal implications of this program. The biggest from a privacy perspective is that it represents a major step鈥攑robably the most major yet鈥攊n placing the United States on the road toward widespread use of face recognition as a technology for tracking and control, and for very little gain. Congress and CBP should end it.

Learn More 老澳门开奖结果 the Issues on This Page