Empirical estimation of generalization ability of neural networks

Research output: Contribution to journalArticle

Abstract

This work concentrates on a novel method for empirical estimation of generalization ability of neural networks. Given a set of training (and testing) data, one can choose a network architecture (number of layers, number of neurons in each layer etc.), an initialization method, and a learning algorithm to obtain a network. One measure of the performance of a trained network is how closely its actual output approximates the desired output for an input that it has never seen before. Current methods provide a "number" that indicates the estimation of the generalization ability of the network. However, this number provides no further information to understand the contributing factors when the generalization ability is not very good. The method proposed uses a number of parameters to define the generalization ability. A set of the values of these parameters provide an estimate of the generalization ability. In addition, the value of each parameter indicates the contribution of such factors as network architecture, initialization method, training data set, etc. Furthermore, a method has been developed to verify the validity of the estimated values of the parameters.

Original languageEnglish (US)
Pages (from-to)3-15
Number of pages13
JournalNeural Network World
Volume11
Issue number1
StatePublished - Jan 1 2001

Fingerprint

Aptitude
Network architecture
Neural networks
Learning algorithms
Neurons
Testing
Learning

Keywords

  • Artificial neural networks
  • Cross-validation
  • Error back-propagation learning
  • Feed-forward networks
  • Generalization ability
  • Voting networks

ASJC Scopus subject areas

  • Software
  • Neuroscience(all)
  • Hardware and Architecture
  • Artificial Intelligence

Cite this

Empirical estimation of generalization ability of neural networks. / Sarkar, Dilip.

In: Neural Network World, Vol. 11, No. 1, 01.01.2001, p. 3-15.

Research output: Contribution to journalArticle

@article{aefa0281f69649e8a45d336701e03146,
title = "Empirical estimation of generalization ability of neural networks",
abstract = "This work concentrates on a novel method for empirical estimation of generalization ability of neural networks. Given a set of training (and testing) data, one can choose a network architecture (number of layers, number of neurons in each layer etc.), an initialization method, and a learning algorithm to obtain a network. One measure of the performance of a trained network is how closely its actual output approximates the desired output for an input that it has never seen before. Current methods provide a {"}number{"} that indicates the estimation of the generalization ability of the network. However, this number provides no further information to understand the contributing factors when the generalization ability is not very good. The method proposed uses a number of parameters to define the generalization ability. A set of the values of these parameters provide an estimate of the generalization ability. In addition, the value of each parameter indicates the contribution of such factors as network architecture, initialization method, training data set, etc. Furthermore, a method has been developed to verify the validity of the estimated values of the parameters.",
keywords = "Artificial neural networks, Cross-validation, Error back-propagation learning, Feed-forward networks, Generalization ability, Voting networks",
author = "Dilip Sarkar",
year = "2001",
month = "1",
day = "1",
language = "English (US)",
volume = "11",
pages = "3--15",
journal = "Neural Network World",
issn = "1210-0552",
publisher = "Institute of Computer Science",
number = "1",

}

TY - JOUR

T1 - Empirical estimation of generalization ability of neural networks

AU - Sarkar, Dilip

PY - 2001/1/1

Y1 - 2001/1/1

N2 - This work concentrates on a novel method for empirical estimation of generalization ability of neural networks. Given a set of training (and testing) data, one can choose a network architecture (number of layers, number of neurons in each layer etc.), an initialization method, and a learning algorithm to obtain a network. One measure of the performance of a trained network is how closely its actual output approximates the desired output for an input that it has never seen before. Current methods provide a "number" that indicates the estimation of the generalization ability of the network. However, this number provides no further information to understand the contributing factors when the generalization ability is not very good. The method proposed uses a number of parameters to define the generalization ability. A set of the values of these parameters provide an estimate of the generalization ability. In addition, the value of each parameter indicates the contribution of such factors as network architecture, initialization method, training data set, etc. Furthermore, a method has been developed to verify the validity of the estimated values of the parameters.

AB - This work concentrates on a novel method for empirical estimation of generalization ability of neural networks. Given a set of training (and testing) data, one can choose a network architecture (number of layers, number of neurons in each layer etc.), an initialization method, and a learning algorithm to obtain a network. One measure of the performance of a trained network is how closely its actual output approximates the desired output for an input that it has never seen before. Current methods provide a "number" that indicates the estimation of the generalization ability of the network. However, this number provides no further information to understand the contributing factors when the generalization ability is not very good. The method proposed uses a number of parameters to define the generalization ability. A set of the values of these parameters provide an estimate of the generalization ability. In addition, the value of each parameter indicates the contribution of such factors as network architecture, initialization method, training data set, etc. Furthermore, a method has been developed to verify the validity of the estimated values of the parameters.

KW - Artificial neural networks

KW - Cross-validation

KW - Error back-propagation learning

KW - Feed-forward networks

KW - Generalization ability

KW - Voting networks

UR - http://www.scopus.com/inward/record.url?scp=0035047277&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=0035047277&partnerID=8YFLogxK

M3 - Article

AN - SCOPUS:0035047277

VL - 11

SP - 3

EP - 15

JO - Neural Network World

JF - Neural Network World

SN - 1210-0552

IS - 1

ER -