Data di Pubblicazione:
2003
Abstract:
The general problem of reconstructing an unknown function from a
finite collection of samples is considered, in case the position of
each input vector in the training set is not fixed beforehand, but is
part of the learning process. In particular, the consistency of the
Empirical Risk Minimization (ERM) principle is analyzed, when the
points in the input space are generated by employing a purely
deterministic algorithm (deterministic learning).
When the output generation is not subject to noise, classical
number-theoretic results, involving discrepancy and variation, allow
to establish a sufficient condition for the consistency of the ERM
principle. In addition, the adoption of low-discrepancy sequences
permits to achieve a learning rate of O(1/L), being L the size of
the training set. An extension to the noisy case is discussed.
Tipologia CRIS:
01.01 Articolo in rivista
Keywords:
Neural networks; Learning; Optimization
Elenco autori:
Cervellera, Cristiano; Muselli, Marco
Link alla scheda completa: