Due to HDSR’s poor formatting of mathematical equations and tables, I recommend reading the paper PDF here, rather than the HDSR publication.
This article is Part II of a trio of papers. See Part I here and Part III here.
Protecting an individual’s privacy when releasing their data is inherently an exercise in relativity, regardless of how privacy is qualified or quantified. This is because we can only limit the gain in information about an individual relative to what could be derived from other sources. This framing is the essence of differential privacy (DP), through which this article examines two statistical disclosure control (SDC) methods for the United States Decennial Census: the Permutation Swapping Algorithm (PSA), which resembles the 2010 Census’s disclosure avoidance system (DAS), and the TopDown Algorithm (TDA), which was used in the 2020 DAS. To varying degrees, both methods leave unaltered certain statistics of the confidential data—their invariants—and hence neither can be readily reconciled with DP, at least as originally conceived. Nevertheless, we show how invariants can naturally be integrated into DP and use this to establish that the PSA satisfies pure DP subject to the invariants it necessarily induces, thereby proving that this traditional SDC method can, in fact, be understood from the perspective of DP. By a similar modification to zero-concentrated DP, we also provide a DP specification for the TDA. Finally, as a point of comparison, we consider a counterfactual scenario in which the PSA was adopted for the 2020 Census, resulting in a reduction in the nominal protection loss budget but at the cost of releasing many more invariants. This highlights the pervasive danger of comparing budgets without accounting for the other dimensions on which DP formulations vary (such as the invariants they permit). Therefore, while our results articulate the mathematical guarantees of SDC provided by the PSA, the TDA, and the 2020 DAS in general, care must be taken in translating these guarantees into actual privacy protection—just as is the case for any DP deployment.
Due to HDSR’s poor formatting of mathematical equations and tables, I recommend reading the paper PDF here, rather than the HDSR publication.
This article is Part II of a trio of papers. See Part I here and Part III here.
James Bailie, Ruobin Gong and Xiao-Li Meng (2026). “A Refreshment Stirred, Not Shaken: Invariant-preserving Deployments of Differential Privacy for the US Decennial Census”. Harvard Data Science Review, Special Issue 6, doi: 10.1162/99608f92.dab78690
Loading...
The trio of papers (of which this article is the second part) were originally presented together as the working paper Can Swapping be Differentially Private? A Refreshment Stirred, not Shaken.