Documents
-
- Download
- Title Pages_Contents
- open access
-
- Download
- Chapter 2
- open access
- Full text at publishers site
-
- Download
- Chapter 4
- open access
- Full text at publishers site
-
- Download
- Bibliography
- open access
-
- Download
- Summary in English
- open access
-
- Download
- Summary in Dutch
- open access
-
- Download
- Propositions
- open access
In Collections
This item can be found in the following collections:
Risk bounds for deep learning
In this thesis, deep learning is studied from a statistical perspective. Convergence rates for the worst case risk bounds of neural network estimators are obtained in the classification, density estimation and linear regression model. Special attention is given to the role of input dimension since in practice, neural networks have shown promising results for high dimensional input settings.
First, the estimation of conditional class probabilities under the cross-entropy loss is studied. A challenge with this loss is that it becomes unbounded near zero. To deal with this, the loss is truncated. Convergence rates are obtained for a neural network estimator under this truncated loss.
The second problem considered is density estimation. A two step procedure is proposed. The first step transforms the density estimation problem into a regression problem by constructing response variables using a kernel density estimator on half of the data. In the second step, a...
Show moreIn this thesis, deep learning is studied from a statistical perspective. Convergence rates for the worst case risk bounds of neural network estimators are obtained in the classification, density estimation and linear regression model. Special attention is given to the role of input dimension since in practice, neural networks have shown promising results for high dimensional input settings.
First, the estimation of conditional class probabilities under the cross-entropy loss is studied. A challenge with this loss is that it becomes unbounded near zero. To deal with this, the loss is truncated. Convergence rates are obtained for a neural network estimator under this truncated loss.
The second problem considered is density estimation. A two step procedure is proposed. The first step transforms the density estimation problem into a regression problem by constructing response variables using a kernel density estimator on half of the data. In the second step, a neural network is fitted to this constructed data. Convergence rates for this method are obtained using existing approximation results for compositional functions.
Finally, forward gradient descent is studied. This is a biologically motivated alternative for gradient descent. Convergence rates are derived for this method in the linear regression model with random design.
- All authors
- Bos, J.M.
- Supervisor
- Grünwald, P.D.; Schmidt-Hieber, A.J.
- Committee
- Derks, G.L.A.; Goeman, J.J.; Rohde, A.; Kohler, M.; Castro, R.M.
- Qualification
- Doctor (dr.)
- Awarding Institution
- Mathematical Institute (MI), Faculty of Science, Leiden University
- Date
- 2024-06-19
- ISBN (print)
- 9789464699784
Funding
- Sponsorship
- NWO/STAR, NWO