Abstract
While rich medical, behavioral, and sociodemographic data are key to modern datadriven research, their collection and use raise legitimate privacy concerns. Anonymizing datasets through deidentification and sampling before sharing them has been the main tool used to address those concerns. We here propose a generative copulabased method that can accurately estimate the likelihood of a specific person to be correctly reidentified, even in a heavily incomplete dataset. On 210 populations, our method obtains AUC scores for predicting individual uniqueness ranging from 0.84 to 0.97, with low falsediscovery rate. Using our model, we find that 99.98% of Americans would be correctly reidentified in any dataset using 15 demographic attributes. Our results suggest that even heavily sampled anonymized datasets are unlikely to satisfy the modern standards for anonymization set forth by GDPR and seriously challenge the technical and legal adequacy of the deidentification releaseandforget model.
Introduction
In the last decade, the ability to collect and store personal data has exploded. With two thirds of the world population having access to the Internet^{1}, electronic medical records becoming the norm^{2}, and the rise of the Internet of Things, this is unlikely to stop anytime soon. Collected at scale from financial or medical services, when filling in online surveys or liking pages, this data has an incredible potential for good. It drives scientific advancements in medicine^{3}, social science^{4,5}, and AI^{6} and promises to revolutionize the way businesses and governments function^{7,8}.
However, the largescale collection and use of detailed individuallevel data raise legitimate privacy concerns. The recent backlashes against the sharing of NHS [UK National Health Service] medical data with DeepMind^{9} and the collection and subsequent sale of Facebook data to Cambridge Analytica^{10} are the latest evidences that people are concerned about the confidentiality, privacy, and ethical use of their data. In a recent survey, >72% of U.S. citizens reported being worried about sharing personal information online^{11}. In the wrong hands, sensitive data can be exploited for blackmailing, mass surveillance, social engineering, or identity theft.
Deidentification, the process of anonymizing datasets before sharing them, has been the main paradigm used in research and elsewhere to share data while preserving people’s privacy^{12,13,14}. Data protection laws worldwide consider anonymous data as not personal data anymore^{15,16} allowing it to be freely used, shared, and sold. Academic journals are, e.g., increasingly requiring authors to make anonymous data available to the research community^{17}. While standards for anonymous data vary, modern data protection laws, such as the European General Data Protection Regulation (GDPR) and the California Consumer Privacy Act (CCPA), consider that each and every person in a dataset has to be protected for the dataset to be considered anonymous^{18,19,20}. This new higher standard for anonymization is further made clear by the introduction in GDPR of pseudonymous data: data that does not contain obvious identifiers but might be reidentifiable and is therefore within the scope of the law^{16,18}.
Yet numerous supposedly anonymous datasets have recently been released and reidentified^{15,21,22,23,24,25,26,27,28,29,30,31}. In 2016, journalists reidentified politicians in an anonymized browsing history dataset of 3 million German citizens, uncovering their medical information and their sexual preferences^{23}. A few months before, the Australian Department of Health publicly released deidentified medical records for 10% of the population only for researchers to reidentify them 6 weeks later^{24}. Before that, studies had shown that deidentified hospital discharge data could be reidentified using basic demographic attributes^{25} and that diagnostic codes, year of birth, gender, and ethnicity could uniquely identify patients in genomic studies data^{26}. Finally, researchers were able to uniquely identify individuals in anonymized taxi trajectories in NYC^{27}, bike sharing trips in London^{28}, subway data in Riga^{29}, and mobile phone and credit card datasets^{30,31}.
Statistical disclosure control researchers and some companies are disputing the validity of these reidentifications: as datasets are always incomplete, journalists and researchers can never be sure they have reidentified the right person even if they found a match^{32,33,34,35}. They argue that this provides strong plausible deniability to participants and reduce the risks, making such deidentified datasets anonymous including according to GDPR^{36,37,38,39}. Deidentified datasets can be intrinsically incomplete, e.g., because the dataset only covers patients of one of the hospital networks in a country or because they have been subsampled as part of the deidentification process. For example, the U.S. Census Bureau releases only 1% of their decennial census and sampling fractions for international census range from 0.07% in India to 10% in South American countries^{40}. Companies are adopting similar approaches with, e.g., the Netflix Prize dataset including <10% of their users^{41}.
Imagine a health insurance company who decides to run a contest to predict breast cancer and publishes a deidentified dataset of 1000 people, 1% of their 100,000 insureds in California, including people’s birth date, gender, ZIP code, and breast cancer diagnosis. John Doe’s employer downloads the dataset and finds one (and only one) record matching Doe’s information: male living in Berkeley, CA (94720), born on January 2^{nd} 1968, and diagnosed with breast cancer (selfdisclosed by John Doe). This record also contains the details of his recent (failed) stage IV treatments. When contacted, the insurance company argues that matching does not equal reidentification: the record could belong to 1 of the 99,000 other people they insure or, if the employer does not know whether Doe is insured by this company or not, to anyone else of the 39.5M people living in California.
Our paper shows how the likelihood of a specific individual to have been correctly reidentified can be estimated with high accuracy even when the anonymized dataset is heavily incomplete. We propose a generative graphical model that can be accurately and efficiently trained on incomplete data. Using sociodemographic, survey, and health datasets, we show that our model exhibits a mean absolute error (MAE) of 0.018 on average in estimating population uniqueness^{42} and an MAE of 0.041 in estimating population uniqueness when the model is trained on only a 1% population sample. Once trained, our model allows us to predict whether the reidentification of an individual is correct with an average falsediscovery rate of <6.7% for a 95% threshold \(( {\widehat {\xi _x}\, > \,0.95} )\) and an error rate 39% lower than the best achievable populationlevel estimator. With population uniqueness increasing fast with the number of attributes available, our results show that the likelihood of a reidentification to be correct, even in a heavily sampled dataset, can be accurately estimated, and is often high. Our results reject the claims that, first, reidentification is not a practical risk and, second, sampling or releasing partial datasets provide plausible deniability. Moving forward, they question whether current deidentification practices satisfy the anonymization standards of modern data protection laws such as GDPR and CCPA and emphasize the need to move, from a legal and regulatory perspective, beyond the deidentification releaseandforget model.
Results
Using Gaussian copulas to model uniqueness
We consider a dataset \({\cal{D}}\), released by an organization, and containing a sample of \(n_{\cal{D}}\) individuals extracted at random from a population of n individuals, e.g., the US population. Each row x^{(i)} is an individual record, containing d nominal or ordinal attributes (e.g., demographic variables, survey responses) taking values in a discrete sample space \({\cal{X}}\). We consider the rows x^{(i)} to be independent and identically distributed, drawn from the probability distribution X with \({\Bbb P}(X = {\boldsymbol{x}})\), abbreviated p(x).
Our model quantifies, for any individual x, the likelihood ξ_{x} for this record to be unique in the complete population and therefore always successfully reidentified when matched. From ξ_{x}, we derive the likelihood κ_{x} for x to be correctly reidentified when matched, which we call correctness. If Doe’s record x^{(d)} is unique in \({\cal{D}}\), he will always be correctly reidentified (\(\kappa _{{\boldsymbol{x}}^{(d)}} = 1\) and \(\xi _{{\boldsymbol{x}}^{(d)}} = 1\)). However, if two other people share the same attribute (\({\boldsymbol{x}}^{(d)}\) not unique, \(\xi _{{\boldsymbol{x}}^{(d)}} = 0\)), Doe would still have one chance out of three to have been successfully reidentified \(\left( {\kappa _{{\boldsymbol{x}}^{(d)}} = 1/3} \right)\). We model \(\xi _{\boldsymbol{x}}\) as:
and κ_{x} as:
with proofs in “Methods”.
We model the joint distribution of X_{1}, X_{2}, … X_{d} using a latent Gaussian copula^{43}. Copulas have been used to study a wide range of dependence structures in finance^{44}, geology^{45}, and biomedicine^{46} and allow us to model the density of X by specifying separately the marginal distributions, easy to infer from limited samples, and the dependency structure. For a large sample space \({\cal{X}}\) and a small number \(n_{\cal{D}}\) of available records, Gaussian copulas provide a good approximation of the density using only d(d − 1)/2 parameters for the dependency structure and no hyperparameter.
The density of a Gaussian copula C_{Σ} is expressed as:
with a covariance matrix Σ, u ∈ [0, 1]^{d}, and Φ the cumulative distribution function (CDF) of a standard univariate normal distribution.
We estimate from \({\cal{D}}\) the marginal distributions Ψ (marginal parameters) for X_{1}, …, X_{d} and the copula distribution Σ (covariance matrix), such that p(x) is modeled by
with F_{j} the CDF of the discrete variable X_{j}. In practice, the copula distribution is a continuous distribution on the unit cube, and p(x) its discrete counterpart on \({\cal{X}}\) (see Supplementary Methods).
We select, using maximum likelihood estimation, the marginal distributions from categorical, logarithmic, and negative binomial count distributions (see Supplementary Methods). Sampling the complete set of covariance matrices to estimate the association structure of copulas is computationally expensive for large datasets. We rely instead on a fast twostep approximate inference method: we infer separately each pairwise correlation factor Σ_{ij} and then project the constructed matrix Σ on the set of symmetric positive definite matrices to accurately recover the copula covariance matrix (see “Methods”).
We collect five corpora from publicly available sources: population census (USA and MERNIS) as well as surveys from the UCI Machine Learning repository (ADULT, MIDUS, HDV). From each corpus, we create populations by selecting subsets of attributes (columns) uniformly. The resulting 210 populations cover a large range of uniqueness values (0–0.96), numbers of attributes (2–47), and records (7108–9M individuals). For readability purposes, we report in the main text the numerical results for all five corpora but will show figures only for USA. Figures for MERNIS, ADULT, MIDUS, and HDV are similar and available in Supplementary Information.
Figure 1a shows that, when trained on the entire population, our model correctly estimates population uniqueness \(\Xi _X = \mathop {\sum}\nolimits_{{\boldsymbol{x}} \in {\cal{X}}} p({\boldsymbol{x}})\left(1  p({\boldsymbol{x}})\right)^{n  1}\), i.e., the expected percentage of unique individuals in (x^{(1)}, x^{(2)}, …, x^{(n)}). The MAE between the empirical uniqueness of our population Ξ_{X} and the estimated uniqueness \(\widehat {\Xi _X}\) is 0.028 ± 0.026 [mean ± s.d.] for USA and 0.018 ± 0.019 on average across every corpus (see Table 1). Figure 1a and Supplementary Fig. 1 furthermore show that our model correctly estimates uniqueness across all values of uniqueness, with low withinpopulation s.d. (Supplementary Table 3).
Figure 1b shows that our model estimates population uniqueness very well even when the dataset is heavily sampled (see Supplementary Fig. 2, for other populations). For instance, our model achieves an MAE of 0.029 ± 0.015 when the dataset only contains 1% of the USA population and an MAE of 0.041 ± 0.053 on average across every corpus. Table 1 shows that our model reaches a similarly low MAE, usually <0.050, across corpora and sampling fractions.
Likelihood of successful reidentification
Once trained, we can use our model to estimate the likelihood of his employer having correctly reidentified John Doe, our 50yearold male from Berkeley with breast cancer. More specifically, given an individual record x, we can use the trained model to compute the likelihood \(\widehat {\xi _{\boldsymbol{x}}} = \left(1  q({\boldsymbol{x}}\,\,\Sigma ,\Psi )\right)^{n  1}\) for this record x to be unique in the population. Our model takes into account information on both marginal prevalence (e.g., breast cancer prevalence) and global attribute association (e.g., gender and breast cancer). Since the cdf. of a Gaussian copula distribution has no closeform expression, we evaluate q(xΣ, Ψ) with a numerical integration of the latent continuous joint density inside the hyperrectangle defined by the d components (x_{1}, x_{2}, …, x_{d})^{47,48}. We assume no prior knowledge on the order of outcomes inside marginals for nominal attributes and randomize their order.
Figure 2a shows that, when trained on 1% of the USA populations, our model predicts very well individual uniqueness, achieving a mean AUC (area under the receiveroperator characteristic curve (ROC)) of 0.89. For each population, to avoid overfitting, we train the model on a single 1% sample, then select 1000 records, independent from the training sample, to test the model. For reidentifications that the model predicts to be always correct (\(\widehat {\xi _{\boldsymbol{x}}}\, > \, 0.95\), estimated individual uniqueness >95%), the likelihood of them to be incorrect (falsediscovery rate) is 5.26% (see bottomright inset in Fig. 2a). ROC curves for the other populations are available in Supplementary Fig. 3 and have overall a mean AUC of 0.93 and mean falsediscovery rate of 6.67% for \(\widehat {\xi _{\boldsymbol{x}}}\, > \, 0.95\) (see Supplementary Table 1).
Finally, Fig. 2b shows that our model outperforms even the best theoretically achievable prediction using only population uniqueness, i.e., assigning the score \(\xi _{\boldsymbol{x}}^{{\mathrm{(pop)}}} = \Xi _X\) to every individual (ground truth population uniqueness, see Supplementary Methods). We use the Brier Score (BS)^{49} to measure the calibration of probabilistic predictions: \({\mathrm{BS}} = \frac{1}{n}\mathop {\sum}\nolimits_{i = 1}^n {\left(\xi _{{\boldsymbol{x}}^{(i)}}  \widehat {\xi _{{\boldsymbol{x}}^{(i)}}}\right)^2}\) with, in our case, \(\xi _{{\boldsymbol{x}}^{(i)}}\) the actual uniqueness of the record \({\boldsymbol{x}}^{(i)}\) (1 if \({\boldsymbol{x}}^{(i)}\) is unique and 0 if not) and \(\widehat {\xi _{{\boldsymbol{x}}^{(i)}}}\) the estimated likelihood. Our model obtains scores on average 39% lower than the best theoretically achievable prediction using only population uniqueness, emphasizing the importance of modeling individuals’ characteristics.
Appropriateness of the deidentification model
Using our model, we revisit the (successful) reidentification of Gov. Weld^{25}. We train our model on the 5% Public Use Microdata Sample (PUMS) files using ZIP code, date of birth, and gender and validate it using the last national estimate^{50}. We show that, as a male born on July 31, 1945 and living in Cambridge (02138), the information used by Latanya Sweeney at the time, William Weld was unique with a 58% likelihood (ξ_{x} = 0.58 and κ_{x} = 0.77), meaning that Latanya Sweeney’s reidentification had 77% chances of being correct. We show that, if his medical records had included number of children—5 for William Weld—, her reidentification would have had 99.8% chances of being correct! Figure 3a shows that the same combinations of attributes (ZIP code, date of birth, gender, and number of children) would also identify 79.4% of the population in Massachusetts with high confidence \(( {\widehat {\xi _{\boldsymbol{x}}} \,\,> \,\, 0.80} )\). We finally evaluate the impact of specific attributes on William Weld’s uniqueness. We either change the value of one of his baseline attributes (ZIP code, date of birth, or gender) or add one extra attribute, in both cases picking the attribute at random from its distribution (see Supplementary Methods). Figure 3c shows, for instance, that individuals with 3 cars or no car are harder to reidentify than those with 2 cars. Similarly, it shows that it would not take much to reidentify people living in Harwich Port, MA, a city of <2000 inhabitants.
Modern datasets contain a large number of points per individuals. For instance, the data broker Experian sold Alteryx access to a deidentified dataset containing 248 attributes per household for 120M Americans^{51}; Cambridge university researchers shared anonymous Facebook data for 3M users collected through the myPersonality app and containing, among other attributes, users’ age, gender, location, status updates, and results on a personality quiz^{52}. These datasets do not necessarily share all the characteristics of the one studied here. Yet, our analysis of the reidentification of Gov. Weld by Latanya Sweeney shows that few attributes are often enough to render the likelihood of correct reidentification very high. For instance, Fig. 3b shows that the average individual uniqueness increases fast with the number of collected demographic attributes and that 15 demographic attributes would render 99.98% of people in Massachusetts unique.
Our results, first, show that few attributes are often sufficient to reidentify with high confidence individuals in heavily incomplete datasets and, second, reject the claim that sampling or releasing partial datasets, e.g., from one hospital network or a single online service, provide plausible deniability. Finally, they show that, third, even if population uniqueness is low—an argument often used to justify that data are sufficiently deidentified to be considered anonymous^{53}—, many individuals are still at risk of being successfully reidentified by an attacker using our model.
As standards for anonymization are being redefined, incl. by national and regional data protection authorities in the EU, it is essential for them to be robust and account for new threats like the one we present in this paper. They need to take into account the individual risk of reidentification and the lack of plausible deniability—even if the dataset is incomplete—, as well as legally recognize the broad range of provable privacyenhancing systems and security measures that would allow data to be used while effectively preserving people’s privacy^{54,55}.
Discussion
In this paper, we proposed and validated a statistical model to quantify the likelihood for a reidentification attempt to be successful, even if the disclosed dataset is heavily incomplete.
Beyond the claim that the incompleteness of the dataset provides plausible deniability, our method also challenges claims that a low population uniqueness is sufficient to protect people’s privacy^{53,56}. Indeed, an attacker can, using our model, correctly reidentify an individual with high likelihood even if the population uniqueness is low (Fig. 3a). While more advanced guarantees like kanonymity^{57} would give every individual in the dataset some protection, they have been shown to be NPHard^{58}, hard to achieve in modern highdimensional datasets^{59}, and not always sufficient^{60}.
While developed to estimate the likelihood of a specific reidentification to be successful, our model can also be used to estimate population uniqueness. We show in Supplementary Note 1 that, while not its primary goal, our model performs consistently better than existing methods to estimate population uniqueness on all five corpora (Supplementary Fig. 4, P < 0.05 in 78 cases out of 80 using Wilcoxon’s signedrank test)^{61,62,63,64,65,66} and consistently better than previous attempts to estimate individual uniqueness^{67,68}. Existing approaches, indeed, exhibit unpredictably large over and underestimation errors. Finally, a recent work quantifies the correctness of individual reidentification in incomplete (10%) hospital data using complete population frequencies^{24}. Compared to this work, our approach does not require external data nor to assume this external data to be complete.
To study the stability and robustness of our estimations, we perform further experiments (Supplementary Notes 2–8).
First, we analyze the impact of marginal and association parameters on the model error and show how to use exogenous information to lower it. Table 1 and Supplementary Note 7 show that, at very small sampling fraction (below 0.1%), where the error is the largest, the error is mostly determined by the marginals, and converges after few hundred records when the exact marginals are known. The copula covariance parameters exhibit no significant bias and decrease fast when the sample size increases (Supplementary Note 8).
As our method separates marginals and association structure inference, exogenous information from larger data sources could also be used to estimate marginals with higher accuracy. For instance, count distributions for attributes such as date of birth or ZIP code could be directly estimated from national surveys. We replicate our analysis on the USA corpus using a subsampled dataset to infer the association structure along with the exact counts for marginal distributions. Incorporating exogenous information reduces, e.g., the mean MAE of uniqueness across all corpora by 48.6% (P < 0.01, Mann–Whitney) for a 0.1% sample. Exogenous information become less useful as the sampling fraction increases (Supplementary Table 2).
Second, our model assumes that \({\cal{D}}\) is either uniformly sampled from the population of interest X or, as several census bureaus are doing, released with poststratification weights to match the overall population. We believe this to be a reasonable assumption as biases in the data would greatly affect its usefulness and affect any application of the data, including our model. To overcome an existing sampling bias, the model can be (i) further trained on a random sample from the population \({\cal{D}}\) (e.g., microdata census or survey data) and then applied to a nonuniform released sample (e.g., hospital data, not uniformly sampled from the population) or (ii) trained using better, potentially unbiased, estimates for marginals or association structure coming from other sources (see above).
Third, since \({\cal{D}}\) is a sample from the population X, only the records that are unique in the sample can be unique in the population. Hence, we further evaluate the performance on our model only on records that are sample unique and show that it only marginally decrease the AUC (Supplementary Note 5). We therefore prefer to not restrict our predictions to sample unique records as (a) our models need to perform well on nonsample unique records for us to be able to estimate correctness and (b) to keep the method robust if oversampling or sampling with replacement were to have been used.
Methods
Inferring marginals distributions
Marginals can be either (i) unknown and are estimated from the marginals of the population sample \(X_{\cal{S}}\), this is the assumption used in the main text, or (ii) known with their exact distribution and cumulative density function directly available.
In the first case, we fit marginal counts to categorical (naive plugin estimator), negative binomial, and logarithmic distributions using maximum loglikelihood. We compare the obtained distributions and select the best likelihood according to its Bayesian information criterion (BIC):
where \(\widehat L\) is the maximized value of the likelihood function, \(n_{\cal{D}}\) the number of individuals in the sample \({\cal{D}}\), and k the number of parameters in the fitted marginal distribution.
Inferring the parameters of the latent copula
Each cell Σ_{ij} of the Σ covariance matrix of a multivariate copula distribution is the correlation parameter of a pairwise copula distribution. Hence, instead of inferring Σ from the set of all covariance matrices, we separately infer every cell Σ_{ij} ∈ [0, 1] from the joint sample of \({\cal{D}}_i\) and \({\cal{D}}_j\). We first measure the mutual information \(I({\cal{D}}_i;{\cal{D}}_j)\) between the two attributes and select \(\sigma = \widehat {\Sigma _{ij}}\) minimizing the Euclidean distance between the empirical mutual information and the mutual information of the inferred joint distribution.
In practice, since the cdf. of a Gaussian copula is not tractable, we use a bounded Nelder–Mead minimization algorithm. For a given (σ, (Ψ_{i}, Ψ_{j})), we sample from the distribution q(⋅σ, (Ψ_{i}, Ψ_{j})) and generate a discrete bivariate sample Y from which we measure the objective:
We then project the obtained \(\widehat \Sigma\) matrix on the set of SDP matrices by solving the following optimization problem:
Modeling the association structure using mutual information
We use the pairwise mutual information to measure the strength of association between attributes. For a dataset \({\cal{D}}\), we denote by \(I_{\cal{D}}\) the mutual information matrix where each cell \(I({\cal{D}}_i;{\cal{D}}_j)\) is the mutual information between attributes \({\cal{D}}_i\) and \({\cal{D}}_j\). When evaluating mutual information from small samples, obtained scores are often overestimating the strength of association. We apply a correction for randomness using a permutation model^{69}:
In practice, we estimate the expected mutual information between \({\cal{D}}_i\) and \({\cal{D}}_j\) with successive permutations of \({\cal{D}}_j\). We found that the adjusted mutual information provides significant improvement for small samples and large support size \({\cal{X}}\) compared to the naive estimator.
Theoretical and empirical population uniqueness
For n individuals x^{(1)}, x^{(2)}, …, x^{(n)} drawn from X, the uniqueness Ξ_{X} is the expected percentage of unique individuals. It can be estimated either (i) by computing the mean of individual uniqueness or (ii) by sampling a synthetic population of n individuals from the copula distribution. In the former case, we have
where T_{x} = [∃!i, x^{(i)} = x] equals one if there exists a single individual i such as x^{(i)} = x and zero otherwise. T_{x} follows a binomial distribution B(p(x), n). Therefore
and
This requires iterating over all combinations of attributes, whose number grows exponentially as the number of attributes increases, and quickly becomes computationally intractable. The second method is therefore often more tractable and we use it to estimate population uniqueness in the paper.
For cumulative marginal distributions F_{1}, F_{2}, …, F_{d} and copula correlation matrix Σ, the algorithm 1 (Supplementary Methods) samples n individuals from q(⋅Σ,Ψ) using the latent copula distribution. From the n generated records (y^{(1)}, y^{(2)}, …, y^{(n)}), we compute the empirical uniqueness
Individual likelihood of uniqueness and correctness
The probability distribution \(q( \cdot \,\,\Sigma ,\Psi )\) can be computed by integrating over the latent copula density. Note that the marginal distributions X_{1} to X_{d} are discrete, causing the inverses \(F_1^{  1}\) to \(F_d^{  1}\) to have plateaus. When estimating p(x), we integrate over the latent copula distribution inside the hypercube \([x_1  1,x_1] \times [x_2  1,x_2] \times \ldots \times [x_d  1,x_d]\):
with ϕ_{Σ} the density of a zeromean multivariate normal (MVN) of correlation matrix Σ. Several methods have been proposed in the literature to estimate MVN rectangle probabilities. Genz and Bretz^{47,48} proposed a randomized quasi Monte Carlo method which we use to estimate the discrete copula density.
The likelihood ξ_{x} for an individual’s record x to be unique in a population of n individuals can be derived from p_{X}(X = x):
Similarly, the likelihood \(\kappa _{\boldsymbol{x}}\) for an individual’s record x to be correctly matched in a population of n individuals can be derived from \(p_X(X = {\boldsymbol{x}})\). With \(T \equiv \mathop {\sum}\nolimits_{i = 1}^n {\left[ {{\boldsymbol{x}}^{(i)} = {\boldsymbol{x}}} \right]}  1\), the number of potential false positives in the population, we have:
Note that, since records are independent, T follows a binomial distribution B(n − 1, p(x)).
We substitute the expression for ξ_{x} in the last formula and obtain:
Data availability
The USA corpus, extracted from the 1Percent Public Use Microdata Sample (PUMS) files, is available at https://www.census.gov/main/www/pums.html. The 5% PUMS files used to estimate the correctness of Governor Weld’s reidentification are also available at the same address. The ADULT corpus, extracted from the Adult Income dataset, is available at https://archive.ics.uci.edu/ml/datasets/adult. The HDV corpus, extracted from the Histoire de vie survey, is available at https://www.insee.fr/fr/statistiques/2532244. The MIDUS corpus, extracted from the Midlife in the United States survey, is available at https://www.icpsr.umich.edu/icpsrweb/ICPSR/series/203. The MERNIS corpus is extracted from a complete population database of virtually all 48 million individuals born before early 1991 in Turkey that was made available online in April 2016 after a data leak from Turkey’s Central Civil Registration System. Our use of this data was approved by Imperial College as it provides a unique opportunity to perform uniqueness estimation on a complete census survey. Owing to the sensitivity of the data, we have only analyzed a copy of the dataset where every distinct value was replaced by a unique integer to obfuscate records, without loss of precision for uniqueness modeling. A complete description of each corpus is available in the Supplementary Information.
Code availability
All simulations were implemented in Julia and Python. The source code to reproduce the experiments is available at https://cpg.doc.ic.ac.uk/individualrisk, along with documentation, tests, and examples.
References
Poushter, J. Smartphone ownership and internet usage continues to climb in emerging economies (Pew Research Center, Washington, DC, 2016). http://www.pewglobal.org/2016/02/22/smartphoneownershipandinternetusagecontinuestoclimbinemergingeconomies/.
Yang, N. & Hing, E. National electronic health records survey. https://cdc.gov/nchs/data/ahcd/nehrs/2015_nehrs_ehr_by_specialty.pdf (2015).
Murdoch, T. B. & Detsky, A. S. The inevitable application of big data to health care. JAMA 309, 1351–1352 (2013).
Wyber, R. et al. Big data in global health: improving health in low and middleincome countries. Bull. World Health Organ. 93, 203–208 (2015).
Lazer, D. et al. Life in the network: the coming age of computational social science. Science 323, 721 (2009).
Halevy, A., Norvig, P. & Pereira, F. The unreasonable effectiveness of data. IEEE Intell. Syst. 24, 8–12 (2009).
Kitchin, R. The realtime city? Big data and smart urbanism. GeoJournal 79, 1–14 (2014).
McAfee, A., Brynjolfsson, E., Davenport, T. H., Patil, D. J. & Barton, D. Big data: the management revolution. Harv. Bus. Rev. 90, 60–68 (2012).
Hodson, H. Revealed: Google AI has access to huge haul of NHS patient data. New Scientist (29 Apr 2016).
Cadwalladr, C. & GrahamHarrison, E. Revealed: 50 million facebook profiles harvested for Cambridge Analytica in major data breach. The Guardian (17 Mar 2018).
Morey, T., Forbath, T. & Schoop, A. Customer data: designing for transparency and trust. Harv. Bus. Rev. 93, 96–105 (2015).
Polonetsky, J., Tene, O. & Finch, K. Shades of gray: seeing the full spectrum of practical data DeIdentification. Santa Clara Law Rev. 56, 593–629 (2016).
Office for Civil Rights, HHS. Standards for privacy of individually identifiable health information. Federal Register. https://ncbi.nlm.nih.gov/pubmed/12180470 (2002).
Malin, B., Benitez, K. & Masys, D. Never too old for anonymity: a statistical standard for demographic data sharing via the HIPAA privacy rule. J. Am. Med. Inform. Assoc. 18, 3–10 (2011).
Rothstein, M. A. Is deidentification sufficient to protect health privacy in research? Am. J. Bioeth. 10, 3–11 (2010).
Council of European Union. Regulation (EU) 2016/679. Off. J. Eur. Union L 119, 1–88 (2016).
Hrynaszkiewicz, I., Norton, M. L., Vickers, A. J. & Altman, D. G. Preparing raw clinical data for publication: guidance for journal editors, authors, and peer reviewers. BMJ 340, c181 (2010).
Opinion 05/2014 on anonymisation techniques. Technical Report, Article 29 Data Protection Working Party. http://ec.europa.eu/justice/article29/documentation/opinionrecommendation/files/2014/wp216_en.pdf (2014).
Rubinstein, I. Framing the discussion. https://fpf.org/wpcontent/uploads/2016/11/Rubinstein_framingpaper.pdf (2016).
Cal. Civil Code. Assembly Bill No. 375 §§ 1798.100–1798.198 (2018).
Narayanan, A. & Felten, E. W. No silver bullet: deidentification still doesn’t work. http://randomwalker.info/publications/nosilverbulletdeidentification.pdf (2014).
Ohm, P. Broken promises of privacy: responding to the surprising failure of anonymization. UCLA. Law Rev. 57, 1701 (2010).
Hern, A. ‘Anonymous’ browsing data can be easily exposed, researchers reveal. The Guardian (1 Aug 2017).
Culnane, C., Rubinstein, B. I. P. & Teague, V. Health data in an open world. Preprint at: https://arxiv.org/abs/1712.05627 (2017).
Sweeney, L. Weaving technology and policy together to maintain confidentiality. J. Law Med. Ethics 25, 98–110. 82 (1997).
Loukides, G., Denny, J. C. & Malin, B. The disclosure of diagnosis codes can breach research participants’ privacy. J. Am. Med. Inform. Assoc. 17, 322–327 (2010).
Douriez, M., Doraiswamy, H., Freire, J. & Silva, C. T. Anonymizing NYC taxi data: does it matter? In 2016 IEEE International Conference on Data Science and Advanced Analytics (DSAA), 140–148 (IEEE, Piscataway, NJ, 2016).
Siddle, J. I know where you were last summer: London’s public bike data is telling everyone where you’ve been. https://vartree.blogspot.com/2014/04/iknowwhereyouwerelastsummer.html (2014). Accessed 7 Feb 2019.
Lavrenovs, A. & Podins, K. Privacy violations in Riga open data public transport system. In 2016 IEEE 4th Workshop on Advances in Information, Electronic and Electrical Engineering (AIEEE), 1–6 (IEEE, Piscataway, NJ, 2016). https://doi.org/10.1109/AIEEE.2016.7821808.
de Montjoye, Y.A., Hidalgo, C. A., Verleysen, M. & Blondel, V. D. Unique in the crowd: the privacy bounds of human mobility. Sci. Rep. 3, 1376 (2013).
de Montjoye, Y.A., Radaelli, L., Singh, V. K. & Pentland, A. Unique in the shopping mall: on the reidentifiability of credit card metadata. Science 347, 536–539 (2015).
Matthews, G. J. & Harel, O. Data confidentiality: a review of methods for statistical disclosure limitation and methods for assessing privacy. Stat. Surv. 5, 1–29 (2011).
BarthJones, D. The ‘reidentification’ of Governor William Weld’s medical information: a critical reexamination of health data identification risks and privacy protections, then and now. https://ssrn.com/abstract=2076397 (2012).
El Emam, K. & Arbuckle, L. Deidentification: a critical debate. https://fpf.org/2014/07/24/deidentificationacriticaldebate/ (2014).
Sánchez, D., Martnez, S. & DomingoFerrer, J. Comment on “unique in the shopping mall: on the reidentifiability of credit card metadata”. Science 351, 1274 (2016).
Reiter, J. P. Estimating risks of identification disclosure in microdata. J. Am. Stat. Assoc. 100, 1103–1112 (2005).
Fienberg, S. E. & Sanil, A. P. A Bayesian approach to data disclosure: optimal intruder behavior for continuous data. J. Stat. 13, 75 (1997).
Duncan, G. & Lambert, D. The risk of disclosure for microdata. J. Bus. Econ. Stat. 7, 207–217 (1989).
Office of the Australian Information Commissioner. Deidentification and the Privacy Act. https://www.oaic.gov.au/agenciesandorganisations/guides/deidentificationandtheprivacyact (2018).
Ruggles, S., King, M. L., Levison, D., McCaa, R. & Sobek, M. IPUMSInternational. Hist. Methods 36, 60–65 (2003).
Bennett, J. & Lanning, S. The Netflix prize. In Proc. KDD Cup and Workshop, 35–38 (ACM, New York, NY, 2007). http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.115.6998.
Sweeney, L. Simple demographics often identify people uniquely. Health 671, 1–34 (2000).
Genest, C. & Mackay, J. The joy of copulas: bivariate distributions with uniform marginals. Am. Stat. 40, 280–283 (1986).
Cherubini, U., Luciano, E. & Vecchiato, W. Copula Methods in Finance (WileyBlackwell, Hoboken, NJ, 2004).
Genest, C. & Favre, A.C. Everything you always wanted to know about copula modeling but were afraid to ask. J. Hydrol. Eng. 12, 347–368 (2007).
Wang, W. & Wells, M. T. Model selection and semiparametric inference for bivariate failuretime data. J. Am. Stat. Assoc. 95, 62–72 (2000).
Genz, A. Numerical computation of multivariate normal probabilities. J. Comput. Graph. Stat. 1, 141–149 (1992).
Genz, A. & Bretz, F. Computation of Multivariate Normal and t Probabilities (Springer Science & Business Media, Berlin, 2009).
Brier, G. W. Verification of forecasts expressed in terms of probability. Mon. Weather Rev. 78, 1–3 (1950).
Golle, P. Revisiting the uniqueness of simple demographics in the US population. In 5th ACM Workshop on Privacy in Electronic Society (ACM, New York, NY, 2006). https://doi.org/10.1145/1179601.1179615.
FoxBrewster, T. 120 million american households exposed in ‘massive’ ConsumerView database leak. Forbes (2017).
Waterfield, P. & Revell, T. Huge new facebook data leak exposed intimate details of 3m users. New Scientist (2018).
El Emam, K. & Arbuckle, L. Anonymizing Health Data (O’Reilly, Newton, MA, 2013).
D’Acquisto, G. et al. Privacy by design in big data: an overview of privacy enhancing technologies in the era of big data analytics. Technical Report. European Union Agency for Network and Information Security (2015).
Cho, H., Wu, D. J. & Berger, B. Secure genomewide association analysis using multiparty computation. Nat. Biotechnol. 36, 547–551 (2018).
Cavoukian, A. & Castro, D. Big data and innovation, setting the record straight: deidentification does work. http://www2.itif.org/2014bigdatadeidentification.pdf (2014).
Sweeney, L. kanonymity: a model for protecting privacy. Int. J. Uncertain. Fuzziness Knowl. Based Syst. 10, 557–570 (2002).
Meyerson, A. & Williams, R. On the complexity of optimal kanonymity. In Proc. 23rd ACM SIGMODSIGACTSIGART Symposium on Principles of Database Systems, 223–228 (2004). https://doi.org/10.1145/1055558.1055591.
Aggarwal, C. C. On kanonymity and the curse of dimensionality. In Proceedings of the 31st International Conference on Very Large Data Bases, VLDB ‘05, 901–909 (VLDB Endowment, 2005). http://dl.acm.org/citation.cfm?id=1083592.1083696.
Li, N., Li, T. & Venkatasubramanian, S. tcloseness: privacy beyond kanonymity and ldiversity. In 2007 IEEE 23rd International Conference on Data Engineering, 106–115 (IEEE, 2007). https://doi.org/10.1109/ICDE.2007.367856.
Ewens, W. J. The sampling theory of selectively neutral alleles. Theor. Popul. Biol. 3, 87–112 (1972).
Chen, G. & KellerMcNulty, S. Estimation of identification disclosure risk in microdata. J. Stat. 14, 79 (1998).
Hoshino, N. Applying pitman’s sampling formula to microdata disclosure risk assessment. J. Stat. 17, 499 (2001).
Keller, W. J. & Pannekoek, J. Disclosure control of microdata. J. Am. Stat. Assoc. 85, 38–45 (1990).
Dankar, F. K., El Emam, K., Neisa, A. & Roffey, T. Estimating the reidentification risk of clinical data sets. BMC Med. Inform. Decis. Mak. 12, 66 (2012).
Pitman, J. Random discrete distributions invariant under sizebiased permutation. Adv. Appl. Probab. 28, 525–539 (1996).
Skinner, C. J. & Holmes, D. J. Estimating the reidentification risk per record in microdata. J. Stat. 14, 361 (1998).
Skinner, C. & Shlomo, N. Assessing identification risk in survey microdata using LogLinear models. J. Am. Stat. Assoc. 103, 989–1001 (2008).
Vinh, N. X., Epps, J. & Bailey, J. Information theoretic measures for clusterings comparison: variants, properties, normalization and correction for chance. J. Mach. Learn. Res. 11, 2837–2854 (2010).
Acknowledgements
L.R. is the recipient of a doctoral fellowship from the Belgian Fund for Scientific Research (F.R.S.FNRS). This collaboration was made possible thanks to Imperial College’s European Partners Fund and a WBI World Excellence Grant. We acknowledge support from the Information Commissioner Office for the development of the online demonstration tool.
Author information
Authors and Affiliations
Contributions
L.R. designed and performed experiments, analyzed the data, and wrote the paper; Y.A.d.M. and J.M.H. designed experiments and wrote the paper.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Peer review information: Nature Communications thanks Antoine Boutet, Vanessa Teague, and other anonymous reviewer(s) for their contribution to the peer review of this work.
Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Rocher, L., Hendrickx, J.M. & de Montjoye, YA. Estimating the success of reidentifications in incomplete datasets using generative models. Nat Commun 10, 3069 (2019). https://doi.org/10.1038/s41467019109333
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41467019109333
This article is cited by

Patientcentric synthetic data generation, no reason to risk reidentification in biomedical data analysis
npj Digital Medicine (2023)

COVIDDynamic: A largescale longitudinal study of socioemotional and behavioral change across the pandemic
Scientific Data (2023)

A pragmatic perspective on AI transparency at workplace
AI and Ethics (2023)

Clinical significance, challenges and limitations in using artificial intelligence for electrocardiographybased diagnosis
International Journal of Arrhythmia (2022)

Data privacy protection in scientific publications: process implementation at a pharmaceutical company
BMC Medical Ethics (2022)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.