Année
2025
Auteurs
ALQUIER Pierre, KENGNE William
Abstract
In a groundbreaking work, [58] proved the minimax optimality of deep neural networks with ReLU activation for least-squares regression estimation over a large class of functions defined by composition. In this paper, we extend these results in many directions. First, we remove the i.i.d. assumption on the observations, to allow some time dependence. The observations are assumed to be a Markov chain with a non-null pseudo-spectral gap. Then, we study a more general class of machine learning problems, which includes least-squares and logistic regression as special cases. Leveraging on PAC-Bayes oracle inequalities and a version of Bernstein inequality due to [53], we derive upper bounds on the estimation risk for a generalized Bayesian estimator. In the case of least-squares regression, this bound matches (up to a logarithmic factor) the lower bound in [58]. We establish a similar lower bound for classification with the logistic loss, and prove that the proposed DNN estimator is optimal in the minimax sense.
ALQUIER, P. et KENGNE, W. (2025). Minimax optimality of deep neural networks on dependent data via PAC-Bayes bounds. The Electronic Journal of Statistics, 19, pp. 5895-5924.