Method for heavy-tailed concentration. So far it has been applied only to scalar heavy-tailed mean estimation, but one could imagine analyzing it in a multivariate setting as well.
Since the empirical mean is ruined by outliers, a natural idea is to remove some fraction of the points and then compute the empirical mean on the remaining points. These are known as trimmed mean estimators. It’s easy to define: Given a random sample and an integer , define
where are the order-statistics.
Oliveira and Orenstein prove that if you remove of the largest and smallest points for , then the trimmed mean is sub-Gaussian. See here for a discussion.
Oliveira et al (2025) perform a robust analysis of the trimmed mean, examining the tails of the estimator and showing they obey a CLT under certain conditions. They also show the estimator is minimax-optimal up to constants under the adversarial contamination model.
A related idea is proposed by Lee and Valiant who don’t remove entire points but instead weighted fractions of each observations:
At the highest level: in order to return a -robust estimate of the mean, our estimator “throws out the most extreme points in the samples”, and returns the mean of what remains. More specifically, outliers are thrown out in a weighted manner, where we throw out a fraction of each data point, with the fraction proportional to the square of its distance from a median-of-means initial guess for the mean, where the fraction is capped at 1, and the proportionality constant is chosen so that the total weight thrown out equals exactly .
Their estimator achieves a bound of , where is the variance of the observations.