Abstract for freitas_icslp98

To appear in Proceedings of the 5th International Conference on Spoken Language Processing, Sydney, Australia, December 1998.

GLOBAL OPTIMISATION OF NEURAL NETWORK MODELS VIA SEQUENTIAL SAMPLE-IMPORTANCE RESAMPLING

JFG de Freitas, SE Johnson, M Niranjan and AH Gee

December 1998

We propose a novel strategy for training neural networks using sequential Monte Carlo algorithms. This global optimisation strategy allows us to learn the probability distribution of the network weights in a sequential framework. It is well suited to applications involving on-line, nonlinear or non-stationary signal processing. We show how the new algorithms can outperform extended Kalman filter (EKF) training.


| (ftp:) freitas_icslp98.ps.gz | (http:) freitas_icslp98.ps.gz | (ftp:) freitas_icslp98.pdf | (http:) freitas_icslp98.pdf | (http:) freitas_icslp98.html/ |

If you have difficulty viewing files that end '.gz', which are gzip compressed, then you may be able to find tools to uncompress them at the gzip web site.

If you have difficulty viewing files that are in PostScript, (ending '.ps' or '.ps.gz'), then you may be able to find tools to view them at the gsview web site.

We have attempted to provide automatically generated PDF copies of documents for which only PostScript versions have previously been available. These are clearly marked in the database - due to the nature of the automatic conversion process, they are likely to be badly aliased when viewed at default resolution on screen by acroread.