Abstract:
We consider a Gaussian sequence space model $X_\lambda=f_\lambda+\xi_\lambda$; where $\xi$ has a diagonal covariance matrix $\sum=diag(\sigma^2_\lambda)$. This heterogeneous model may appear in frameworks where the variance is fluctuating, for example in inverse problems or fractional Brownian motion. We consider mainly the situation where the parameter is sparse. Our goal is to estimate the unknown parameter by a model selection approach. The heterogenous case is much more involved than the direct model. Indeed, there is no more symmetry inside the stochastic process that one needs to control, since each empirical coefficient has its own variance. The problem and the penalty do not only depend on the number of coefficients that one selects, but also on their position. This appears also in the minimax bounds where the worst coefficients will go to the larger variances. However, with a careful and explicit choice of the penalty we are able to select the correct coefficients and get a sharp non-asymptotic control of the risk of our procedure. Results are also obtained for full model selection and a family of thresholds. We obtain a minimax upper bound, the estimator almost attains the lower bound (up to a constant 2). Moreover, the procedure is fully adaptive, we obtain an explicit penalty, valid in the mathematical proofs and in simulations. [Joint work with Markus Reiss]