Abstract
Hypothesis testing is a statistical inference framework for determining the true distribution among a set of possible distributions for a given data set. Privacy restrictions may require the curator of the data or the respondents themselves to share data with the test only after applying a randomizing privacy mechanism. This work considers mutual information (MI) as the privacy metric for measuring leakage. In addition, motivated by the Chernoff-Stein lemma, the relative entropy between pairs of distributions of the output (generated by the privacy mechanism) is chosen as the utility metric. For these metrics, the goal is to find the optimal privacy-utility tradeoff (PUT) and the corresponding optimal privacy mechanism for both binary and m -ary hypothesis testing. Focusing on the high privacy regime, Euclidean information-theoretic approximations of the binary and m -ary PUT problems are developed. The solutions for the approximation problems clarify that an MI-based privacy metric preserves the privacy of the source symbols in inverse proportion to their likelihoods.
Original language | English (US) |
---|---|
Article number | 8125176 |
Pages (from-to) | 1058-1071 |
Number of pages | 14 |
Journal | IEEE Transactions on Information Forensics and Security |
Volume | 13 |
Issue number | 4 |
DOIs | |
State | Published - Apr 2018 |
Keywords
- Hypothesis testing
- Rényi divergence
- euclidean information theory
- mutual information
- privacy mechanism
- privacy-guaranteed data publishing
- relative entropy
ASJC Scopus subject areas
- Safety, Risk, Reliability and Quality
- Computer Networks and Communications