Privacy-enhancing technologies (PET) are technologies that embody fundamental data protection principles by minimizing personal data use, maximizing data security, and empowering individuals. PETs allow
online users to protect the
privacy
Privacy (, ) is the ability of an individual or group to seclude themselves or information about themselves, and thereby express themselves selectively.
The domain of privacy partially overlaps with security, which can include the concepts of a ...
of their
personally identifiable information (PII), which is often provided to and handled by services or applications. PETs use techniques to minimize an information system's possession of
personal data without losing functionality.
Generally speaking, PETs can be categorized as either hard or soft privacy technologies.
Goals of PETs
The objective of PETs is to protect
personal data and assure technology users of two key privacy points: their own information is kept confidential, and management of
data protection is a priority to the organizations who hold responsibility for any
PII. PETs allow users to take one or more of the following actions related to personal data that is sent to and used by
online service providers, merchants or other users (this control is known as
self-determination
Self-determination refers to a people's right to form its own political entity, and internal self-determination is the right to representative government with full suffrage.
Self-determination is a cardinal principle in modern international la ...
). PETs aim to minimize personal data collected and used by service providers and merchants, use
pseudonym
A pseudonym (; ) or alias () is a fictitious name that a person assumes for a particular purpose, which differs from their original or true meaning ( orthonym). This also differs from a new name that entirely or legally replaces an individual's o ...
s or anonymous data credentials to provide anonymity, and strive to achieve informed consent about giving personal data to online service providers and merchants. In Privacy Negotiations, consumers and service providers establish, maintain, and refine privacy policies as individualized agreements through the ongoing choice among service alternatives, therefore providing the possibility to negotiate the terms and conditions of giving personal data to online service providers and merchants (data handling/privacy policy negotiation). Within private negotiations, the transaction partners may additionally bundle the personal information collection and processing schemes with monetary or non-monetary rewards.
PETs provide the possibility to remotely audit the enforcement of these terms and conditions at the online service providers and merchants (assurance), allow users to log, archive and look up past transfers of their personal data, including what data has been transferred, when, to whom and under what conditions, and facilitate the use of their legal rights of data inspection, correction and deletion. PETs also provide the opportunity for consumers or people who want privacy-protection to hide their personal identities. The process involves masking one's personal information and replacing that information with pseudo-data or an anonymous identity.
Families of PETs
Privacy-enhancing Technologies can be distinguished based on their assumptions.
Soft privacy technologies
Soft privacy technologies are used where it can be assumed that a third-party can be trusted for the processing of data. This model is based on
compliance,
consent
Consent occurs when one person voluntarily agrees to the proposal or desires of another. It is a term of common speech, with specific definitions used in such fields as the law, medicine, research, and sexual consent. Consent as understood i ...
, control and auditing.
Example technologies are
access control
In physical security and information security, access control (AC) is the action of deciding whether a subject should be granted or denied access to an object (for example, a place or a resource). The act of ''accessing'' may mean consuming ...
,
differential privacy
Differential privacy (DP) is a mathematically rigorous framework for releasing statistical information about datasets while protecting the privacy of individual data subjects. It enables a data holder to share aggregate patterns of the group while ...
, and
tunnel encryption (SSL/TLS).
An example of soft privacy technologies is increased transparency and access. Transparency involves granting people with sufficient details about the rationale used in automated decision-making processes. Additionally, the effort to grant users access is considered soft privacy technology. Individuals are usually unaware of their right of access or they face difficulties in access, such as a lack of a clear automated process.
Hard privacy technologies
With hard privacy technologies, no single entity can violate the privacy of the user. The assumption here is that third-parties cannot be trusted. Data protection goals include
data minimization and the reduction of trust in third-parties.
Examples of such technologies include
onion routing
Onion routing is a technique for anonymous communication over a computer network. In an onion network, messages are encapsulated in layers of encryption, analogous to the layers of an onion. The encrypted data is transmitted through a series o ...
, the
secret ballot
The secret ballot, also known as the Australian ballot, is a voting method in which a voter's identity in an election or a referendum is anonymous. This forestalls attempts to influence the voter by intimidation, blackmailing, and potential vote ...
, and VPNs used for democratic elections.
Existing PETs
PETs have evolved since their first appearance in the 1980s. At intervals, review articles have been published on the state of privacy technology:
* A principal, though fundamentally theoretical, overview of terminology and principal anonymization technology is found in Pfitzmann & Hansen'
terminology of anonymity
* In 1997,
reportby Goldberg, Wagner and Brewer at the
University of California, Berkeley
The University of California, Berkeley (UC Berkeley, Berkeley, Cal, or California), is a Public university, public Land-grant university, land-grant research university in Berkeley, California, United States. Founded in 1868 and named after t ...
summarized PETs.
* In 2003, Borking, Blarkom and Olk reviewed the technologies from a data protection perspective in their Handbook of privacy enhancing technologies.
* In 2007, Fritsch published an historic, taxonomic and practica
overview of contemporary privacy-enhancing technologyfor the Internet for the research project PETWeb.
* In 2008, Fritsch and Abie documented the gap between implemented PETs and their successful deployment in
research roadmap for PETs
* In 2015, Heurix et al. published a taxonomy of privacy enhancing technologies.
* A specialization of PET research that looks into increasing the transparency of data processing is called Transparency Enhancing Technologies (TETs). A review article by Janic et al. summarizes developments in TETs. Murmann and Fischer-Hübner published
review of transparency toolsin 2017.
* In 2019, the
World Economic Forum
The World Economic Forum (WEF) is an international non-governmental organization, international advocacy non-governmental organization and think tank, based in Cologny, Canton of Geneva, Switzerland. It was founded on 24 January 1971 by German ...
published
white paperexploring PET use cases in financial technology and infrastructure.
* The Boston Women's Workforce Council published reports i
2017an
2019exploring the gender pay gap in a number of Boston-based companies. The data was compared using PETs, to ensure that sensitive employee information remained private throughout.
* In 2020, Identiq published a
ebookdiscussing PETs that are actively being used in identity validation.
* In 2021, the
European Data Protection Board, which oversees the enforcement of GDPR, and the European Union Agency for Cybersecurity publishe
technical guidancesupporting Secure Multi-Party Computation as a valid privacy-preserving safeguard, applying to both healthcare and cybersecurity use cases.
Example PETs
Examples of existing privacy enhancing technologies are:
General PET building blocks:
*
Obfuscation
Obfuscation is the obscuring of the intended meaning of communication by making the message difficult to understand, usually with confusing and ambiguous language. The obfuscation might be either unintentional or intentional (although intent ...
refers to the many practices of adding distracting or misleading data to a log or profile, which may be especially useful for frustrating precision analytics after data has already been lost or disclosed. Its effectiveness against humans is questioned, but it has greater promise against shallow algorithms. Obfuscating also hides personal information or sensitive data through computer algorithms and masking techniques. This technique can also involve adding misleading or distracting data or information so it's harder for an attacker to obtain the needed data.
*Providing users with access to their personal data: Here, a user gains control over the privacy of their data within a service because the service provider's infrastructure allows users to inspect, correct or delete all their data that is stored at the service provider.
*
Pseudonymization is a data management technique that replaces an individual's identity or personal information with an artificial identifiers known as Pseudonyms. This de-identification method enables contents and fields of information to be covered up so as to deter attacks and hackers from obtaining important information. These Pseudonyms can be either placed in groups or for individual pieces of information. Overall, they serve to discourage information stealing while also maintaining data integrity and data analysis.
PETs for Privacy-Preserving Communication:
*
Communication anonymizers hiding a user's real online identity (email address, IP address, etc.) and replacing it with a non-traceable identity (disposable / one-time email address, random IP address of hosts participating in an anonymising network, pseudonym, etc.). They can be applied to everyday applications like email, Web browsing,
P2P networking,
VoIP
Voice over Internet Protocol (VoIP), also known as IP telephony, is a set of technologies used primarily for voice communication sessions over Internet Protocol (IP) networks, such as the Internet. VoIP enables voice calls to be transmitted as ...
, Chat, instant messaging, etc.
*Shared bogus online accounts. This technology de-links an online account from a specific user's habits by allowing many users to share the account, and setting up fake personal information in the account settings. To accomplish this, one person creates an account for a website like MSN, providing bogus data for their name, address, phone number, preferences, life situation etc. They then publish their user-IDs and passwords on the internet. Everybody can now use this account comfortably. Thereby the user is sure that there is no personal data about him or her in the account profile. (Moreover, he is freed from the hassle of having to register at the site himself.)
PETs for Privacy Preserving Data Processing are PETs that facilitate data processing or the production of statistics while preserving privacy of the individuals providing raw data, or of the specific raw data elements. Some examples include:
*
Enhanced privacy ID (EPID) is a
digital signature algorithm
The Digital Signature Algorithm (DSA) is a Public-key cryptography, public-key cryptosystem and Federal Information Processing Standards, Federal Information Processing Standard for digital signatures, based on the mathematical concept of modular e ...
supporting anonymity. Unlike traditional digital signature algorithms (e.g.,
PKI), in which each entity has a unique public verification key and a unique private signature key, EPID provides a common group public verification key associated with many of unique private signature keys. EPID was created so that a device could prove to an external party what kind of device it is (and optionally what software is running on the device) without needing to also reveal exact identity, i.e., to prove you are an authentic member of a group without revealing ''which'' member. It has been in use since 2008.
*
Zero-knowledge proof
In cryptography, a zero-knowledge proof (also known as a ZK proof or ZKP) is a protocol in which one party (the prover) can convince another party (the verifier) that some given statement is true, without conveying to the verifier any information ...
is a method by which one party (the prover) can prove to another party (the verifier) that they know a value x, without conveying any information apart from the fact that they know the value x.
*
Ring signature is a type of
digital signature that can be performed by any member of a set of users that each have a pair of cryptographic keys.
*
Format-preserving encryption (FPE), refers to encrypting in such a way that the output (the
ciphertext
In cryptography, ciphertext or cyphertext is the result of encryption performed on plaintext using an algorithm, called a cipher. Ciphertext is also known as encrypted or encoded information because it contains a form of the original plaintext ...
) is in the same format as the input (the plaintext)
*
Blinding is a cryptography technique by which an agent can provide a service to a client in an encoded form without knowing either the real input or the real output.
PETs for Privacy Preserving Data Analytics are a subset of the PETs used for data processing that are specifically designed for the publishing of statistical data. Some examples include:
*
Homomorphic encryption
Homomorphic encryption is a form of encryption that allows computations to be performed on encrypted data without first having to decrypt it. The resulting computations are left in an encrypted form which, when decrypted, result in an output th ...
is a form of encryption that allows computation on ciphertexts.
*
Secure multi-party computation is a method for parties to jointly compute a function over their inputs while keeping those inputs private.
*
Non-interactive zero-knowledge proof
Across the many fields concerned with interactivity, including information science, computer science, human-computer interaction, communication, and industrial design, there is little agreement over the meaning of the term "interactivity", but ...
(NIZKs) are zero-knowledge proofs that require no interaction between the prover and verifier.
*
Differential privacy
Differential privacy (DP) is a mathematically rigorous framework for releasing statistical information about datasets while protecting the privacy of individual data subjects. It enables a data holder to share aggregate patterns of the group while ...
: An algorithm is constrained so that the results or outputs of a data analysis can't tell if a certain individuals' information is being used to analyze and form the results. This technique focuses on large databases and hides the identity of individual "inputs" who might have private data and privacy concerns,
*
Federated learning
Federated learning (also known as collaborative learning) is a machine learning technique in a setting where multiple entities (often called clients) collaboratively train a model while keeping their data Decentralized computing, decentralized, ra ...
is a machine learning technique that trains models across multiple distributed nodes. Each node houses a local, private dataset.
*
Adversarial stylometry methods may allow authors writing anonymously or pseudonymously to resist having their texts linked to their other identities due to linguistic clues.
Future PETs
Examples of privacy enhancing technologies that are being researched or developed include limited disclosure technology, anonymous credentials, negotiation and enforcement of data handling conditions, and data transaction logs.
Limited disclosure technology provides a way of protecting individuals' privacy by allowing them to share only enough personal information with service providers to complete an interaction or transaction. This technology is also designed to limit tracking and correlation of users’ interactions with these third parties. Limited disclosure uses
cryptographic
Cryptography, or cryptology (from "hidden, secret"; and ''graphein'', "to write", or '' -logia'', "study", respectively), is the practice and study of techniques for secure communication in the presence of adversarial behavior. More gen ...
techniques and allows users to retrieve data that is vetted by a provider, to transmit that data to a relying party, and have these relying parties trust the authenticity and integrity of the data.
Anonymous credentials are asserted properties or rights of the
credential
A credential is a piece of any document that details a qualification, competence, or authority issued to an individual by a third party with a relevant or ''de facto'' authority or assumed competence to do so.
Examples of credentials include aca ...
holder that don't reveal the true identity of the holder; the only information revealed is what the holder of the credential is willing to disclose. The assertion can be issued by the user himself/herself, by the provider of the online service or by a third party (another service provider, a government agency, etc.). For example:
Online car rental. The car rental agency doesn't need to know the true identity of the customer. It only needs to make sure that the customer is over 23 (as an example), that the customer has a drivers license,
health insurance
Health insurance or medical insurance (also known as medical aid in South Africa) is a type of insurance that covers the whole or a part of the risk of a person incurring medical expenses. As with other types of insurance, risk is shared among ma ...
(i.e. for accidents, etc.), and that the customer is paying. Thus there is no real need to know the customers name nor their address or any other
personal information. Anonymous credentials allow both parties to be comfortable: they allow the customer to only reveal so much data which the car rental agency needs for providing its service (
data minimization), and they allow the car rental agency to verify their requirements and get their money. When ordering a car online, the user, instead of providing the classical name, address and
credit card number, provides the following credentials, all issued to
pseudonym
A pseudonym (; ) or alias () is a fictitious name that a person assumes for a particular purpose, which differs from their original or true meaning ( orthonym). This also differs from a new name that entirely or legally replaces an individual's o ...
s (i.e. not to the real name of the customer):
* An assertion of minimal age, issued by the state, proving that the holder is older than 23 (note: the actual age is not provided)
* A
driving licence, i.e. an assertion, issued by the motor vehicle control agency, that the holder is entitled to drive cars
* A
proof of insurance, issued by the health insurance
* Digital cash
Negotiation and enforcement of data handling conditions. Before ordering a product or service online, the user and the online service provider or merchant negotiate the type of
personal data that is to be transferred to the service provider. This includes the conditions that shall apply to the handling of the personal data, such as whether or not it may be sent to third parties (profile selling) and under what conditions (e.g. only while informing the user), or at what time in the future it shall be deleted (if at all). After the transfer of personal data took place, the agreed upon data handling conditions are technically enforced by the infrastructure of the service provider, which is capable of managing and processing and data handling obligations. Moreover, this enforcement can be remotely audited by the user, for example by verifying chains of certification based on
Trusted computing modules or by verifying privacy seals/labels that were issued by third party auditing organizations (e.g. data protection agencies). Thus instead of the user having to rely on the mere promises of service providers not to abuse
personal data, users will be more confident about the service provider adhering to the negotiated data handling conditions
Lastly, the data transaction log allows users the ability to log the personal data they send to service provider(s), the time in which they do it, and under what conditions. These logs are stored and allow users to determine what data they have sent to whom, or they can establish the type of data that is in possession by a specific service provider. This leads to more
transparency, which is a pre-requisite of being in control.
See also
*
Crypto-shredding
*
Cypherpunk
*
Digital credentials
*
Digital self-determination
*
Enhanced privacy ID (EPID)
*
Identity management
*
Information privacy
Information privacy is the relationship between the collection and dissemination of data, technology, the public expectation of privacy, contextual information norms, and the legal and political issues surrounding them. It is also known as dat ...
*
Information processing
*
Information security
Information security is the practice of protecting information by mitigating information risks. It is part of information risk management. It typically involves preventing or reducing the probability of unauthorized or inappropriate access to data ...
*
Privacy
Privacy (, ) is the ability of an individual or group to seclude themselves or information about themselves, and thereby express themselves selectively.
The domain of privacy partially overlaps with security, which can include the concepts of a ...
*
Privacy by design
*
Privacy Engineering
*
Privacy-enhanced Electronic Mail
*
Privacy software
*
Privacy policy
*
Self-sovereign identity
References
*
*
Notes
External links
PETs in general:
The EU PRIME research project(2004 to 2008) aiming at studying and developing novel PETs
About PETs from the Center for Democracy and TechnologyAnnual symposium on PETsReport about PETs from the META Group, published by the Danish ministry of sciencebroken link
Anonymous credentials:
*IBM Zürich Research Lab'
idemix*
Stefan Brands
Stefan Brands is the designer of the core cryptographic protocols of Microsoft's U-Prove technology.
Business career
Following his academic research on these protocols during the nineties, they were implemented and marketed under the U-Pro ...
'
U-Prove Digital credentialbr>
'credentica'*which is now owned by Microsof
U-ProvePrivacy policy negotiation:
*The
W3C's P3P
*IBM's
EPAL
*Sören Preibusch
Implementing Privacy Negotiations in E-Commerce ''Discussion Papers of DIW Berlin'' 526, 2005
{{DEFAULTSORT:Privacy Enhancing Technologies
Privacy
Identity management
Data protection