|New Strategies for Inference on High-Dimensional Data
|2:00 pm - 3:00 pm
|Lady Shaw Building LT1
|Latest Seminars and Events
|Professor Alexander GIESSING
How can we conduct inference on high-dimensional data when classical asymptotic theory fails because noise accumulates faster than at root-n rate and function classes are non-Donsker? In this talk, I discuss two complementary approaches to address this problem: First, I introduce a new framework for inference on regression functions in the presence of high-dimensional nuisance parameters. At its core is a convex program that aims to minimize the mean squared prediction error of the regression function by trading oﬀ bias and variance optimally. The resulting regression function is asymptotically normal and semiparametric eﬃcient “in a weak sense”. I will focus on a speciﬁc implementation of this framework for estimating the conditional mean when outcomes are missing at random. Second, I discuss recent advances on the high-dimensional bootstrap. I propose a new bootstrap procedure that takes the correlation structure of the data into account and, under assumption on the correlation structure, allows consistent bootstrapping of the sampling distribution of supremum-type statistics. These theoretical results generalize the existing Gaussian multiplier bootstrap and broaden its applicability for inference on high-dimensional data. They also provide new intuition for why the high-dimensional bootstrap works or doesn't.