Thursday, March 14th, 2019

9:00 am9:10 am
Speaker: Lalitha Sankar (Arizona State University)

 . 

9:10 am10:10 am
Speaker: Aaron Wagner (Cornell University)

How much information is "leaked" in a side channel?  Despite decades of work on these channels, including the development of many sophisticated mitigation mechanisms for specific side channels, the fundamental question of how to measure the key quantity of interest---leakage---has received surprisingly little attention. Many metrics have been used in the literature, but these metrics either lack a cogent operational justification or mislabel systems
that are obviously insecure as secure.

We propose a new metric called "maximal leakage," defined as the logarithm of the multiplicative increase, upon observing the public data, of the probability of correctly guessing a randomized function of the private information, maximized over all such randomized functions.
We provide an operational justification for this definition, show how it can be computed in practice, and discuss how it relates to existing metrics, including mutual information, local differential privacy, and a certain under-appreciated metric in the computer science literature. We also present some structural results for optimal mechanisms under this metric. Among other findings, we show that mutual information underestimates leakage while local differential privacy overestimates it.

This is joint work with Ibrahim Issa, Sudeep Kamath, Ben Wu, and Ed Suh.

10:25 am11:25 am
Speaker: Lalitha Sankar (Arizona State University)

In many data-sharing or learning applications, ensuring unnecessary inference or inappropriate use of sensitive data is essential while simultaneously guaranteeing usefulness of the data. It is now well accepted that randomizing mechanisms are needed to ensure privacy or fairness. In this talk, we will discuss a recently introduced class of leakage measures that allow quantifying the information a learning adversary can infer from a post-randomized dataset. In particular, we will focus on maximal alpha leakage as a new class of adversarially motivated tunable leakage measures that is based on guessing an arbitrary function of a dataset conditioned on the released dataset. The choice of alpha determines the specific adversarial action ranging from refining a belief for alpha = 1 to guessing the best posterior for alpha = ∞. Relationship of this measure to mutual information, maximal leakage, maximal information, Renyi DP, and local DP will be discussed, in particular from the viewpoint of adversarial actions. The tutorial-style talk will also include discussion of adversarial knowledge of side information as well as the consequences of using this measure to design privacy mechanisms.

This is joint work with Jiachun Liao, Oliver Kosut, and Flavio Calmon.

1:30 pm2:15 pm
Speaker: Flavio du Pin Calmon (Harvard University)

We present a short overview of a few information-theoretic methods for understanding and ensuring fairness in machine learning algorithms. First, we discuss how perturbation of measure approaches (e.g., influence functions) can be used to interpret and correct for bias in a given machine learning model. We then overview how tools from rate-distortion theory may be useful for designing data pre-processing mechanisms for ensuring fairness. Finally, we conclude with future research directions that may be of interest to both data scientists and information theorists.

2:30 pm3:15 pm
Speaker: Peter Kairouz (Google AI)

We present Generative Adversarial Privacy and Fairness (GAPF), a data-driven framework for learning private and fair representations of large-scale datasets. GAPF leverages recent advances in generative adversarial networks (GANs) to allow a data holder to learn ``universal'' data representations that decouple a set of sensitive attributes from the rest of the dataset. Under GAPF, finding the optimal privacy/fairness mechanism is formulated as a constrained minimax game between a private/fair encoder and an adversary. We show that for appropriately chosen adversarial loss functions, GAPF provides privacy guarantees against information-theoretic adversaries and enforces demographic parity. We also evaluate the performance of GAPF on the GENKI and CelebA face datasets and the Human Activity Recognition (HAR) dataset.

Based on joint work with Chong Huang (ASU), Xiao Chen and Ram Rajagopal (Stanford), and Lalitha Sankar (ASU).

Friday, March 15th, 2019

9:00 am9:40 am
Speaker: Kamalika Choudhari (UCSD)

The vast majority of computer science literature in privacy can be broadly divided into two categories -- inferential, where we are trying to bound the inferences an adversary can make based on auxiliary information, and differential, where the idea is to ensure that participation of an entity or an individual does not change the outcome significantly.

In this talk, I will present two new case-studies, one in each framework. The first looks at a form of inferential privacy that allows more fine-grained control in a local setting than the individual level. The second looks at privacy against adversaries who have bounded learning capacity, and has ties to the theory of generative adversarial networks.

9:40 am10:20 am
Speaker: Flavio du Pin Calmon (Harvard University)

In this talk, we overview recent efforts in understanding privacy-utility trade-offs (PUTs) from an information-theoretic perspective. Under certain metrics for privacy and utility (e.g., mutual information and other $f$-divergences), PUTs can be understood specific instantiations of a broader class of formulations called ``bottleneck problems,'' which include the information bottleneck and the privacy funnel. These formulations are closely related to data processing inequalities, and reveal interesting facets of the trade-offs that emerge when designing privacy-assuring mechanisms.

10:35 am11:05 am
Speaker: Peter Kairouz (Google AI)

We consider a setting where a designer would like to assess the efficacy of a privacy/fairness scheme by evaluating its performance against a finite-capacity adversary interested in learning a sensitive attribute from released data. Here, a finite-capacity adversary is a learning agent with limited statistical knowledge (finite number of data samples) and limited expressiveness capabilities limited to those expressed by a neural network. We provide probabilistic bounds on the discrepancy between the risk performance of such a finite capacity adversary relative to an infinite capacity adversary for the squared and log-losses, where an infinite-capacity adversary is one with full statistical knowledge and expressiveness capabilities. Our bounds quantify both the generalization error resulting from limited samples and the function approximation limits resulting from finite expressiveness. We illustrate our results for both scalar and multi-dimensional Gaussian mixture models.

Based on joint work with Mario Diaz (ASU/CIMAT), Chong Huang (ASU), and Lalitha Sankar (ASU).

1:00 pm2:00 pm

 . 

1:30 pm2:00 pm
Speaker: Aaron Wagner (Cornell University)

How much information is "leaked" in a side channel?  Despite decades of work on these channels, including the development of many sophisticated mitigation mechanisms for specific side channels, the fundamental question of how to measure the key quantity of interest---leakage---has received surprisingly little attention. Many metrics have been used in the literature, but these metrics either lack a cogent operational justification or mislabel systems
that are obviously insecure as secure.

We propose a new metric called "maximal leakage," defined as the logarithm of the multiplicative increase, upon observing the public data, of the probability of correctly guessing a randomized function of the private information, maximized over all such randomized functions.
We provide an operational justification for this definition, show how it can be computed in practice, and discuss how it relates to existing metrics, including mutual information, local differential privacy, and a certain under-appreciated metric in the computer science literature. We also present some structural results for optimal mechanisms under this metric. Among other findings, we show that mutual information underestimates leakage while local differential privacy overestimates it.

This is joint work with Ibrahim Issa, Sudeep Kamath, Ben Wu, and Ed Suh.

2:00 pm2:35 pm
Speaker: Lalitha Sankar (Arizona State University)

Building upon my previous talk on information leakage measures and their value in providing privacy guarantees against learning adversaries, in this talk, I will discuss the robustness of privacy guarantees that can be made via alpha leakage measures when designing mechanisms from a finite number of samples. The talk will also highlight recent work by researchers in the information theory community on noise adding mechanisms. Finally, we will focus on the significance of the adversarial model in understanding both mechanism design and the guarantees provided.

Mechanism design is based on joint work with Hao Wang, Mario Diaz, and Flavio Calmon. Additive noise mechanisms describes the work of Prakash Narayan and Arvind Nageswaran. Adversarial models and loss functions is based on work with Tyler Sypherd, Mario Diaz, and Peter Kairouz.