Skip to main content
Log in

Efficient GA Based Techniques for Classification

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

A common approach to evaluating competing models in a classification context is via accuracy on a test set or on cross-validation sets. However, this can be computationally costly when using genetic algorithms with large datasets and the benefits of performing a wide search are compromised by the fact that estimates of the generalization abilities of competing models are subject to noise. This paper shows that clear advantages can be gained by using samples of the test set when evaluating competing models. Further, that applying statistical tests in combination with Occam's razor produces parsimonious models, matches the level of evaluation to the state of the search and retains the speed advantages of test set sampling.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. S. Salzberg, “A critique of current research and methods,” Department of Computer Science, John Hopkins University, Technical Report JHU-95/06, 1995.

  2. A. Miller, Subset Selection in Regression, Chapman and Hall, 1990.

  3. Statlog data and documentation at ftp.ncc.up.pt/pub/statlog.

  4. R. Kohavi and D. Sommerfield, “Feature subset selection using the wrapper model: Overfitting and dynamic search space toplogogy,” in First Int. Conf. on Knowledge Discovery and Data Mining, 1995, pp. 192-197.

  5. J. Fitzpatrick and T. Grefenstette, “Genetic algorithms in noisy environments,” Machine Learning, vol. 3,no. 2/3, pp. 101-120, 1985.

    Google Scholar 

  6. S. Rana and D. Whitley et al., “Searching in the presence of noise,” Parallel Problem Solving from Nature, vol. 4, pp. 198-207, 1996.

    Google Scholar 

  7. J. Holland, Adaptation in Natural and Artificial Systems, University of Michigan Press, 1975.

  8. F. Brill and D. Brown et al., “Fast genetic selection of features for neural network classifiers,” IEEE Trans. Neural Networks, vol. 3,no. 2, pp. 324-328, 1992.

    Google Scholar 

  9. R. Smith and E. Dike et al., “Inheritance in genetic algorithms,” in Proceedings of the ACM 1995 Symposium on Applied Computing, ACM Press, 1994.

  10. O. Maron and A. Moore, “Hoeffding races: Accelerating model selection search for classification and function approximation,” Advances in Neural Information Processing Systems 6, Morgan Kaufmann, 1994.

  11. A. Moore and M. Lee, “Efficient algorithms for minimizing cross validation error,” in Proc. Eleventh Int. Conf. on Machine Learning, Morgan Kaufmann, 1994.

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

About this article

Cite this article

Sharpe, P.K., Glover, R.P. Efficient GA Based Techniques for Classification. Applied Intelligence 11, 277–284 (1999). https://doi.org/10.1023/A:1008386925927

Download citation

  • Issue Date:

  • DOI: https://doi.org/10.1023/A:1008386925927

Navigation