Thursday, November 16th, 2023

Sam Hopkins, Massachusetts Institute of Technology

Synopsis

While statistical learning theory gives reliable foundations for estimation and inference from ideal data sets, real world data sets are quite far from the ideal. With sources of imperfections such as noisy observations, miscoded values, missing values, and distribution shift between collection and implementation time, it has now become critical to develop tools which can give robust insights from such imperfect data.

The workshop aims to bring together researchers from Statistics, Computer Science, and Operations Research to discuss novel perspectives on and approaches towards imperfect data.

Sivaraman Balakrishnan, Carnegie Mellon University

Logistics

  • Date: Thursday, November 16, 2023
  • In-person Location: Northwestern University: Mudd Library 3rd floor, 2233 Tech Drive, Evanston
  • Register: Click here to register

Tamara Broderick, Massachusetts Institute of Technology

Schedule (more details are forthcoming)

9:15 – 9:50 Breakfast
9:50 – 10:00 Opening remarks
10:00 – 10:45 Sam Hopkins (MIT)
10:45 – 11:30 Sivaraman Balakrishnan (CMU)
11:30 – 11:45 Coffee break
11:45 – 12:30 Tamara Broderick (MIT)

12:30 – 1:45 Lunch
 
1:45- 2:30 Lightning talks
  • Nian Si (University of Chicago)
  • Subhodh Kotekal (University of Chicago)
  • Yuzhang Shang (IIT)
2:30-3:00 Coffee break
3:00-3:45 Neil Gong (Duke)
3:45-4:30 Victor Veitch (University of Chicago)
4:30-5:15 Free chat

5:15-   Dinner

 

Neil Gong, Duke University

Titles and Abstracts

Speaker: Sivaraman Balakrishnan

Title: Minimax Theory for Causal Inference

Abstract: One of the most salient examples of attempting to make rigorous inferences from imperfect data is in attempting to infer causal quantities from observational data. The causal analysis of (observational) data plays a central role in essentially every scientific field. Many recent developments in causal inference, and functional estimation problems more generally, have been motivated by the fact that classical one-step de-biasing methods, or their more recent sample-split double machine-learning avatars, can outperform plugin estimators under surprisingly weak conditions. However, from a theoretical perspective our understanding of how to construct these estimators for non-standard functionals, how to assess their optimality, and how to improve them, is still far from complete.

I will present two vignettes within this theme. The first part develops minimax theory for estimating the conditional average treatment effect (CATE). Many methods for estimating CATEs have been proposed, but there remain important theoretical gaps in understanding if and when such methods are optimal. We close some of these gaps, by providing sharp minimax-rates for estimating the CATE when the nuisance functions are Holder smooth — highlighting important differences between the estimation of the CATE, and its more well-studied global counterpart (the average treatment effect).

In the second part, I will focus more broadly on functional estimation problems, and develop some minimax lower bounds for “structure-agnostic” functional estimation, to understand the strengths and limitations of the double machine learning perspective on functional estimation.
 
This talk will be based on joint work with Edward Kennedy and Larry Wasserman.

____

Speaker: Tamara Broderick

Title: An Automatic Finite-Sample Robustness Check: Can Dropping a Little Data Change Conclusions?

Abstract: Practitioners will often analyze a data sample with the goal of applying any conclusions to a new population. For instance, if economists conclude microcredit is effective at alleviating poverty based on observed data, policymakers might decide to distribute microcredit in other locations or future years. Typically, the original data is not a perfect random sample from the population where policy is applied — but researchers might feel comfortable generalizing anyway so long as deviations from random sampling are small, and the corresponding impact on conclusions is small as well. Conversely, researchers might worry if a very small proportion of the data sample was instrumental to the original conclusion. So we propose a method to assess the sensitivity of statistical conclusions to the removal of a very small fraction of the data set. Manually checking all small data subsets is computationally infeasible, so we propose an approximation based on the classical influence function. Our method is automatically computable for common estimators. We provide finite-sample error bounds on approximation performance and a low-cost exact lower bound on sensitivity. We find that sensitivity is driven by a signal-to-noise ratio in the inference problem, does not disappear asymptotically, and is not decided by misspecification. Empirically we find that many data analyses are robust, but the conclusions of several influential economics papers can be changed by removing (much) less than 1% of the data.

____

Speaker: Neil Gong

Title: Secure Content Moderation for Generative AI

Abstract: Generative AI–such as GPT-4 and DALL-E 3–raises many ethical and legal concerns such as the generation of harmful content, scaling disinformation and misinformation campaigns, as well as disrupting education and learning. Content moderation for generative AI aims to address these ethical and legal concerns via 1) preventing a generative AI model from synthesizing harmful content, and 2) detecting AI-generated content. Prevention is often implemented using safety filters, while detection is implemented by watermark. Both prevention and watermark-based detection have been recently widely deployed by industry. In this talk, we will discuss the security of existing prevention and watermark-based detection methods in adversarial settings.

____

Speaker: Sam Hopkins

Title: The Full Landscape of Robust Mean Testing: Sharp Separations between Oblivious and Adaptive Contamination

Abstract: We consider the question of Gaussian mean testing, a fundamental task in high-dimensional distribution testing and signal processing, subject to adversarial corruptions of the samples. We focus on the relative power of different adversaries, and show that, in contrast to the common wisdom in robust statistics, there exists a strict separation between adaptive adversaries (strong contamination) and oblivious ones (weak contamination) for this task. We design both new testing algorithms and new lower bounds to show that robust testing in the presence of an oblivious adversary requires strictly fewer samples than in the presence of an adaptive one. Joint work with Clement Canonne, Jerry Li, Allen Liu, and Shyam Narayanan, appeared in FOCS 2023.

____

Speaker: Subhodh Kotekal

Title: Optimal null estimation in the two-groups model

Abstract: The advent of large-scale inference has spurred reexamination of conventional statistical thinking. In a series of highly original articles, Efron showed in some examples that the ensemble of the null distributed test statistics grossly deviated from the theoretical null distribution, and Efron persuasively illustrated the danger in assuming the theoretical null’s veracity for downstream inference. Though intimidating in other contexts, the large-scale setting is to the statistician’s benefit here. There is now potential to estimate, rather than assume, the null distribution. We adopt Efron’s suggestion and consider rate-optimal estimation of the null in the two-groups model without imposing any assumptions on the nonnull effects. The minimax upper bound is obtained by considering estimators based on the empirical characteristic function and the classical kernel mode estimator. Faster rates than those in Huber’s contamination model are achievable by exploiting the Gaussian character of the data.

____

Speaker: Nian Si (collaborative work with Shengbo Wang (Stanford), Jose Blanchet (Stanford), and Zhengyuan Zhou (NYU Stern))

Title: On the Foundation of Distributionally Robust Reinforcement Learning

Abstract: Motivated by learning a good policy that is robust to environment shifts between training time (i.e., in a simulator) and test time (i.e., in a real environment), we contribute to the theoretical foundation of distributionally robust reinforcement learning (DRRL).  We accomplish this through the introduction of a comprehensive modeling framework centered around distributionally robust Markov decision processes (DRMDPs). This framework obliges the decision maker to select an optimal policy under the worst-case distributional shift provisioned by an adversary. By unifying and extending existing formulations in the literature, we provide a rigorous construction of a DRMDP, which encompasses several modeling attributes pertaining to both the decision maker and the adversary. These attributes include the granularity at which the decision maker and/or the adversary can adapt, where we  examine three common scenarios including history-dependent, Markov, and Markov time-homogeneous decision maker/adversary.  Furthermore, we explore the flexibility of shifts that the adversary can induce, where we further examine the SA-rectangular and S-rectangular adversaries.

____

Speaker: Victor Veitch

Title: Linear Structure of High-Level Concepts in Text-Controlled Generative Models

Abstract: Text controlled generative models (such as large language models or text-to-image diffusion models) operate by embedding natural language into a vector representation, then using this representation to sample from the model’s output space. This talk concerns how high-level semantics are encoded in the algebraic structure of representations. In particular, we look at the idea that such representations are “linear”—what this means, why such structure emerges, and how it can be used for precision understanding and control of generative models.

Victor Veitch, University of Chicago

Organizers

Join Our Newsletter