Abstract for holden_tr113

Cambridge University Engineering Department Technical Report CUED/F-INFENG/TR113

QUANTIFYING GENERALIZATION IN LINEARLY WEIGHTED NEURAL NETWORKS

Sean Holden and Martin Anthony

February 1993

The Vapnik-Chervonenkis dimension has proven to be of great use in the theoretical study of generalization in artificial neural networks. The `probably approximately correct' learning framework is described and the importance of the VC dimension is illustrated. We then investigate the VC dimension of certain types of linearly weighted neural networks. First, we obtain bounds on the VC dimensions of radial basis function networks with basis functions of several types. Secondly, we calculate the VC dimension of polynomial discriminant functions defined over both real and binary-valued inputs.


(ftp:) holden_tr113.ps.Z (http:) holden_tr113.ps.Z
PDF (automatically generated from original PostScript document - may be badly aliased on screen):
  (ftp:) holden_tr113.pdf | (http:) holden_tr113.pdf

If you have difficulty viewing files that end '.gz', which are gzip compressed, then you may be able to find tools to uncompress them at the gzip web site.

If you have difficulty viewing files that are in PostScript, (ending '.ps' or '.ps.gz'), then you may be able to find tools to view them at the gsview web site.

We have attempted to provide automatically generated PDF copies of documents for which only PostScript versions have previously been available. These are clearly marked in the database - due to the nature of the automatic conversion process, they are likely to be badly aliased when viewed at default resolution on screen by acroread.