Skip to main content

A Novel GA-Taguchi-Based Feature Selection Method

  • Conference paper
Intelligent Data Engineering and Automated Learning – IDEAL 2008 (IDEAL 2008)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 5326))

Abstract

This work presents a novel GA-Taguchi-based feature selection method. Genetic algorithms are utilized with randomness for “global search” of the entire search space of the intractable search problem. Various genetic operations, including crossover, mutation, selection and replacement are performed to assist the search procedure in escaping from sub-optimal solutions. In each iteration in the proposed nature-inspired method, the Taguchi methods are employed for “local search” of the entire search space and thus can help explore better feature subsets for next iteration. The two-level orthogonal array is utilized for a well-organized and balanced comparison of two levels for features—a feature is or is not selected for pattern classification—and interactions among features. The signal-to-noise ratio (SNR) is then used to determine the robustness of the features. As a result, feature subset evaluation efforts can be significantly reduced and a superior feature subset with high classification performance can be obtained. Experiments are performed on different application domains to demonstrate the performance of the proposed nature-inspired method. The proposed hybrid GA-Taguchi-based approach, with wrapper nature, yields superior performance and improves classification accuracy in pattern classification.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Blake, C.L., Merz, C.J.: UCI Repository of machine learning databases. University of California, Department of Information and a Computer Science, Irvine (1998), http://www.ics.uci.edu/~mlearn/MLRepository.html

    Google Scholar 

  2. Blum, A., Langley, P.: Selection of Relevant Features and Examples in Machine Learning. Artificial Intelligence 97, 245–272 (1997)

    Article  MathSciNet  MATH  Google Scholar 

  3. Cawley, G.C., Talbot, N.L.C.: Efficient Leave-one-out Cross-validation of Kernel Fisher Discriminant Classifiers. Pattern Recognition 36, 2585–2592 (2003)

    Article  MATH  Google Scholar 

  4. Cover, T.M., Hart, P.E.: Nearest Neighbor Pattern Classification. IEEE Trans. on Information Theory 13, 21–27 (1967)

    Article  MATH  Google Scholar 

  5. Dasarathy, B.V.: Nearest Neighbor (NN) Norms: NN Pattern Classification Techniques. IEEE Computer Society Press, Los Alamitos (1990)

    Google Scholar 

  6. Dash, M., Liu, H.: Feature Selection for Classification. Intelligent Data Analysis 2, 232–256 (1997)

    Google Scholar 

  7. Doak, J.: An Evaluation of Feature Selection Methods and Their Application to Computer Security. Technical Report, Univ. of California at Davis, Dept. Computer Science (1992)

    Google Scholar 

  8. Duda, R.O., Hart, P.E.: Pattern Classification and Scene Analysis. John Wiley & Sons, Chichester (1973)

    MATH  Google Scholar 

  9. Goldberg, D.E.: Genetic Algorithms in Search, Optimization and Machine Learning. Addison Wesley, Reading (1989)

    MATH  Google Scholar 

  10. Hall, M.A.: Correlation-based Feature Subset Selection for Machine Learning. PhD Dissertation, University of Waikato (1998)

    Google Scholar 

  11. Holland, J.H.: Adaptation in natural and artificial systems. University of Michigan Press, Ann Arbor (1975)

    Google Scholar 

  12. Inza, I., Larrañaga, P., Sierra, B.: Feature Subset Selection by Bayesian Networks: a Comparison with Genetic and Sequential Algorithms. International Journal of Approximate Reasoning 27, 143–164 (2001)

    Article  MATH  Google Scholar 

  13. John, G.H., Kohavi, R., Pfleger, K.: Irrelevant Feature and the Subset Selection Problem. In: Proc. 11th Int’l Conf. Machine Learning, pp. 121–129 (1994)

    Google Scholar 

  14. Kohavi, R., John, G.H.: Wrappers for Feature Subset Selection. Artificial Intelligence 97, 273–324 (1997)

    Article  MATH  Google Scholar 

  15. Liu, H., Motoda, H.: Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic, Boston (1998)

    Book  MATH  Google Scholar 

  16. Liu, H., Setiono, R.: A Probabilistic Approach to Feature Selection - A Filter Solution. In: Proc. of 13th International Conference on Machine Learning, pp. 319–327 (1996)

    Google Scholar 

  17. Liu, H., Yu, L.: Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Trans. Knowl. Data Eng. 17, 491–502 (2005)

    Article  Google Scholar 

  18. Mitchell, M.: An Introduction to Genetic Algorithms. MIT Press, Cambridge (1992)

    MATH  Google Scholar 

  19. Stone, M.: Cross-validatory Choice and Assessment of Statistical Predictions. Journal of the Royal Statistical Society B 36, 111–147 (1974)

    MathSciNet  MATH  Google Scholar 

  20. Taguchi, G., Chowdhury, S., Taguchi, S.: Robust Engineering. McGraw-Hill, New York (2000)

    MATH  Google Scholar 

  21. Wu, Y., Wu, A., Taguchi, G.: Taguchi Methods for Robust Design. ASME, New York (2000)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2008 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Yang, CH., Huang, CC., Wu, KC., Chang, HY. (2008). A Novel GA-Taguchi-Based Feature Selection Method. In: Fyfe, C., Kim, D., Lee, SY., Yin, H. (eds) Intelligent Data Engineering and Automated Learning – IDEAL 2008. IDEAL 2008. Lecture Notes in Computer Science, vol 5326. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-88906-9_15

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-88906-9_15

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-88905-2

  • Online ISBN: 978-3-540-88906-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics