Elsevier

Neural Networks

Volume 10, Issue 2, March 1997, Pages 231-241
Neural Networks

A Bayesian Approach to Model Selection in Hierarchical Mixtures-of-Experts Architectures

https://doi.org/10.1016/S0893-6080(96)00050-0Get rights and content
Under a Creative Commons license
open archive

Abstract

There does not exist a statistical model that shows good performance on all tasks. Consequently, the model selection problem is unavoidable; investigators must decide which model is best at summarizing the data for each task of interest. This article presents an approach to the model selection problem in hierarchical mixtures-of-experts architectures. These architectures combine aspects of generalized linear models with those of finite mixture models in order to perform tasks via a recursive “divide-and-conquer” strategy. Markov chain Monte Carlo methodology is used to estimate the distribution of the architectures' parameters. One part of our approach to model selection attempts to estimate the worth of each component of an architecture so that relatively unused components can be pruned from the architecture's structure. A second part of this approach uses a Bayesian hypothesis testing procedure in order to differentiate inputs that carry useful information from nuisance inputs. Simulation results suggest that the approach presented here adheres to the dictum of Occam's razor; simple architectures that are adequate for summarizing the data are favored over more complex structures. © 1997 Elsevier Science Ltd. All Rights Reserved.

Keywords

Modular architecture
Hierarchical architecture
Model selection
Bayesian analysis
Gibbs sampling

Cited by (0)