Journal Title
Title of Journal:
|
|
Publisher
Springer, New York, NY
|
|
|
|
Authors: David J C MacKay
Publish Date: 1996
Volume: , Issue: , Pages: 211-254
Abstract
Bayesian probability theory provides a unifying framework for data modeling In this framework the overall aims are to find models that are well matched to the data and to use these models to make optimal predictions Neural network learning is interpreted as an inference of the most probable parameters for the model given the training data The search in model space ie the space of architectures noise models preprocessings regularizers and weight decay constants also then can be treated as an inference problem in which we infer the relative probability of alternative models given the data This provides powerful and practical methods for controlling comparing and using adaptive network models This chapter describes numerical techniques based on Gaussian approximations for implementation of these methods
Keywords:
.
|
Other Papers In This Journal:
|