New Trends May Help TikTok Collect Your Personal, Unchangeable Biometric Identifiers
People post pictures and recordings of themselves every day online. It feels harmless, and for the most part, it is. But new technologies can add a dangerous layer to casual social media posting: the possible extraction, without our knowledge or consent, of our biometric identifiers.
Biometric identifiers are measurements of our personal, unchangeable biological characteristics that uniquely identify us 鈥 like fingerprints and faceprints. Biometric identifiers can also be captured from the patterns in our eyes or the sound of our voices. Unlike other forms of personal information, biometric identifiers 肠补苍鈥檛 be changed, even once compromised. People can change their names and addresses to shield their whereabouts and identities from individuals who seek to harm them, but they 肠补苍鈥檛 change their faces. And, since biometric identifiers can be collected surreptitiously and at a distance, they offer a surveillance capability unlike any other technology in the past, making it dangerously easy to identify and track us at protests, AA meetings, counseling sessions, political rallies, religious gatherings, and more.
This danger isn鈥檛 far off into the future. The technology exists now: A can already identify you by your iris from 40 feet away. have been extracted from photos of exposed fingers, taken both up close and from a distance. Because this can happen at a distance, or from recordings posted online, it鈥檚 possible without our knowledge or consent. For many of us, hitting the record button to create content online can now pose an unintentional privacy risk that could result in the loss of control over our most sensitive information.
Last month, TikTok users started a seemingly innocuous that highlights the irises in their eyes. Users show their eyes up close, then use a high resolution filter to show the details, patterns, and colors of their irises. What started as a new filter in response to the popular TV series 鈥淓uphoria鈥 has turned into a rapidly-growing trend: More than 700,000 videos have been created in the span of around a month. However, as these videos proliferated, so did concerns about people inadvertently exposing their unique biometric identifiers to collection through automated scanning of these iris images. Such scanning is a form of biometric technology similar to face recognition and fingerprinting.
Despite TikTok users expressing concern, the company has not answered critical questions, including whether TikTok has partnered with companies selling iris recognition technology, whether they鈥檙e harvesting this sensitive data, where it鈥檚 being hosted, or what they might use this data for. The lack of transparency is a big part of the problem.
From using facial recognition to automatically for people鈥檚 faces 鈥 even those people who are not on the app 鈥 to to inform auto-captioning, many of TikTok鈥檚 key features rely on their ability to capture mass amounts of data, raising a host of data collection and privacy issues. This includes a in that TikTok of collecting subscribers鈥 biometrics and other personally identifying data, tracking the data and, in some cases, sharing it with third parties without consent. The lawsuit alleged that the collection of biometric identifiers without consent violated the Illinois right to privacy by establishing guardrails for how companies can collect and use people鈥檚 biometric identifiers.
TikTok鈥檚 parent company ByteDance, for its part, denied these accusations, although as part of a settlement they did agree to pay $92 million to individuals whose interests were represented by the class action, as well as to hire a third-party firm for three years to oversee data privacy training for employees, and to avoid collecting biometrics and location data from users without notifying them and without complying with BIPA and similar laws.
After TikTok settled, they quietly updated their in summer 2021 to add facial features, audio, and words spoken to a list of information the company can collect automatically. Perhaps most troublingly, the policy specified that the company 鈥渕ay collect biometric identifiers and biometric information鈥 from its users鈥 content, including 鈥渇aceprints and voiceprints.鈥 When reached out to TikTok, they could not confirm what product developments necessitated the addition of biometric data to its list of disclosures about the information it may automatically collect from users, but said it would ask for consent in the case such data collection practices began.
The language used in TikTok鈥檚 updated policy is vague, raising questions about how it might be applied or abused. For example, there鈥檚 a line in the policy saying: 鈥淲here required by law, we will seek any required permissions from you prior to any such collection.鈥 Companies should commit to only collecting and using people鈥檚 sensitive biometric identifiers with express prior consent. Unfortunately, intense industry lobbying over the course of years means that very few states have laws requiring consent before companies can collect people鈥檚 biometric identifiers. Only Illinois, Texas, and Washington have enacted these kinds of biometric privacy laws. Nor is there a comprehensive federal policy on how companies should go about storing, selling, or sharing this incredibly important personal information. These sparse restrictions are a nightmare for our civil liberties. Companies, like Clearview AI for example, have offered up massive biometric databases to private companies, police, federal agencies, and wealthy individuals, allowing them to secretly track and target whomever they wished using this technology.
If businesses like TikTok have access to our biometric identifiers, there is little regulation defining what they 肠补苍鈥檛 do with them, which is what could make trends like the Euphoria filter so concerning. TikTok is not the only social network operating in a largely unregulated environment where personal data drives corporate profits, and it鈥檚 also the only company sitting on a ton of biometric data. That means that we all need to be vigilant about the growing number of apps and filters that take our most personal data in exchange for creating entertaining content. Companies like TikTok have an obligation to protect their more than privacy.
As long as an app claims the right to collect and store biometric data, it is critical that users are given the resources they need to understand what they鈥檙e giving away, as well as the ability to protect that information. Particularly, companies should obtain active user consent through a transparent process before collecting biometric identifiers, and include provisions that allow users to deny consent to such data collection without any penalties. And on a bigger scale, more states must join places like Illinois in enacting legislation that ensures that companies only use our biometrics with our knowledge and consent.
The stakes are high, wide-ranging, and unpredictable: When apps have access to our most personal information, there are endless avenues for exploitation.