Paper
22 March 1996 Empirical estimation of generalization ability of neural networks
Author Affiliations +
Abstract
This work concentrates on a novel method for empirical estimation of generalization ability of neural networks. Given a set of training (and testing) data, one can choose a network architecture (number of layers, number of neurons in each layer etc.), an initialization method, and a learning algorithm to obtain a network. One measure of performance of the trained network is how closely its actual output approximates the desired output for an input that it has never seen before. Current methods provide a `number' that indicates the estimation of generalization ability of the network. However, this number provides no further information to understand the contributing factors when generalization stability is not very good. The proposed method uses a number of parameters to define generalization ability. A set of values of these parameters provides an estimate of generalization ability. In addition, a value of each parameter indicate the contribution of such factors as network architecture, initialization method, and training data set etc. Furthermore, a method has been developed to verify the validity of estimated values of the parameters.
© (1996) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Dilip Sarkar "Empirical estimation of generalization ability of neural networks", Proc. SPIE 2760, Applications and Science of Artificial Neural Networks II, (22 March 1996); https://doi.org/10.1117/12.235979
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Network architectures

Neural networks

Data modeling

Neurons

Algorithm development

Metals

Signal processing

Back to Top