Designing Statistical Estimators That Balance Sample Size, Risk, and Computational Cost
This paper proposes a tradeoff between computational time, sample complexity, and statistical accuracy that applies to statistical estimators based on convex optimization. When we have a large amount of data, we can exploit excess samples to decrease statistical risk, to decrease computational cost, or to trade off between the two. We propose to achieve this tradeoff by varying the amount of smoothing applied to the optimization problem. This work uses regularized linear regression as a case study to argue for the existence of this tradeoff both theoretically and experimentally. We also apply our method to describe a tradeoff in an image interpolation problem.
time-data (2).pdf
openaccess
5.37 MB
Adobe PDF
bcb26854a9ac9d73e521e9c2abc137b9