Poster
Differentially Private Analysis for Binary Response Models: Optimality, Estimation, and Inference
Ce Zhang · Yixin Han · Yafei Wang · Xiaodong Yan · Linglong Kong · Ting Li · Bei Jiang
East Exhibition Hall A-B #E-900
In today's data-driven world, it’s essential to protect people’s sensitive information, like personal survey answers, while still allowing scientists to make useful conclusions. One common way to do this is through “randomized response,” a technique that introduces intentional noise to protect individual answers. But the challenge is: how do we still make accurate conclusions from such noisy data? Our paper presents a new method that balances privacy and accuracy more effectively. We design an improved system that works especially well when the responses are “yes or no” (binary), and we show that it performs better than older methods, especially in real-world tasks like detecting plagiarism in student surveys. We also create a way to give researchers confidence intervals — a key tool in statistics — even when the data is privatized. This work helps ensure privacy doesn't come at the cost of scientific reliability, making it valuable for fields like health, education, and social science.