skip to main content
research-article

Self-Adaptive Particle Swarm Optimization for Large-Scale Feature Selection in Classification

Authors Info & Claims
Published:24 September 2019Publication History
Skip Abstract Section

Abstract

Many evolutionary computation (EC) methods have been used to solve feature selection problems and they perform well on most small-scale feature selection problems. However, as the dimensionality of feature selection problems increases, the solution space increases exponentially. Meanwhile, there are more irrelevant features than relevant features in datasets, which leads to many local optima in the huge solution space. Therefore, the existing EC methods still suffer from the problem of stagnation in local optima on large-scale feature selection problems. Furthermore, large-scale feature selection problems with different datasets may have different properties. Thus, it may be of low performance to solve different large-scale feature selection problems with an existing EC method that has only one candidate solution generation strategy (CSGS). In addition, it is time-consuming to find a suitable EC method and corresponding suitable parameter values for a given large-scale feature selection problem if we want to solve it effectively and efficiently. In this article, we propose a self-adaptive particle swarm optimization (SaPSO) algorithm for feature selection, particularly for large-scale feature selection. First, an encoding scheme for the feature selection problem is employed in the SaPSO. Second, three important issues related to self-adaptive algorithms are investigated. After that, the SaPSO algorithm with a typical self-adaptive mechanism is proposed. The experimental results on 12 datasets show that the solution size obtained by the SaPSO algorithm is smaller than its EC counterparts on all datasets. The SaPSO algorithm performs better than its non-EC and EC counterparts in terms of classification accuracy not only on most training sets but also on most test sets. Furthermore, as the dimensionality of the feature selection problem increases, the advantages of SaPSO become more prominent. This highlights that the SaPSO algorithm is suitable for solving feature selection problems, particularly large-scale feature selection problems.

Skip Supplemental Material Section

Supplemental Material

References

  1. J. Aguaron, M. T. Escobar, and J. M. Moreno-Jimenez. 2016. The precise consistency consensus matrix in a local AHP-group decision making context. Annals of Operations Research 245, 12 (2016), 245--259.Google ScholarGoogle ScholarCross RefCross Ref
  2. A. Al-Ani, A. Alsukker, and R. Khushaba. 2013. Feature subset selection using differential evolution and a wheel based search strategy. Swarm and Evolutionary Computation 9 (2013), 15--26.Google ScholarGoogle ScholarCross RefCross Ref
  3. K. Bache and M. Lichman. 2016. UCI Machine Learning Repository. http://archive.ics.uci.edu/ml/index.php.Google ScholarGoogle Scholar
  4. K. K. Bharti and P. K. Singh. 2016. Opposition chaotic fitness mutation based adaptive inertia weight BPSO for feature selection in text clustering. Applied Soft Computing 43 (2016), 20--34.Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. X. J. Chang, F. P. Nie, Y. Yang, C. Q. Zhang, and H. Huang. 2016. Convex sparse PCA for unsupervised feature learning. ACM Transactions on Knowledge Discovery from Data 11, 1 (2016), 16.Google ScholarGoogle Scholar
  6. M. Dorigo and L. M. Gambardella. 1997. Ant colony system: A cooperative learning approach to the traveling salesman problem. IEEE Transactions on Evolutionary Computation 1, 1 (1997), 53--66.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. D. B. Fogel. 1994. An introduction to simulated evolutionary optimization. IEEE Transactions on Neural Networks 5, 1 (1994), 3--14.Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. A. S. Ghareb, B. A. Abu, and A. R. Hamdan. 2016. Hybrid feature selection based on enhanced genetic algorithm for text categorization. Expert Systems with Applications 49 (2016), 31--47.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. I. A. Gheyas and L. S. Smith. 2010. Feature subset selection in large dimensionality domains. Pattern Recognition 43, 1 (2010), 5--13.Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. K. R. Harrison, A. P. Engelbrecht, and B. M. Ombuki-Berman. 2018. Self-adaptive particle swarm optimization: A review and analysis of convergence. Swarm Intelligence 12 (2018), 187--226.Google ScholarGoogle ScholarCross RefCross Ref
  11. J. H. Holland. 1975. Adaptation in Natural and Artificial Systems. University of Michigan Press.Google ScholarGoogle Scholar
  12. U. Kamath, J. Compton, R. I. Dogan, K. D. Jong, and A. Shehu. 2012. An evolutionary algorithm approach for feature generation from sequence data and its application to DNA splice site prediction. IEEE-ACM Transactions on Computational Biology and Bioinformatics 9, 5 (2012), 1387--1398.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. J. Kennedy and R. Eberhart. 1995. Particle swarm optimization. In IEEE International Conference on Neural Networks. 1942--1948.Google ScholarGoogle Scholar
  14. J. R. Koza. 1990. Genetic Programming: A Paradigm for Genetically Breeding Populations of Computer Programs to Solve Problems, Vol. 34. Stanford University.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. C. H. Li, S. X. Yang, and T. T. Nguyen. 2012. A self-learning particle swarm optimizer for global optimization problems. IEEE Transactions on Systems Man and Cybernetics Part B-Cybernetics 42, 3 (2012), 627--646.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. J. Li and D. C. Tao. 2012. On preserving original variables in Bayesian PCA with application to image analysis. IEEE Transactions on Image Processing 21, 12 (2012), 4830--4843.Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. J. J. Liang, A. K. Qin, P. N. Suganthan, and S. Baskar. 2006. Comprehensive learning particle swarm optimizer for global optimization of multimodal functions. IEEE Transactions on Evolutionary Computation 10, 3 (2006), 281--295.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. T. Marill and D. Green. 1963. On the effectiveness of receptors in recognition systems. IEEE Transactions on Information Theory 9, 1 (1963), 11--17.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. V. E. Neagoe and E. C. Neghina. 2016. Feature selection with ant colony optimization and its applications for pattern recognition in space imagery. In International Conference on Communications. 101--104.Google ScholarGoogle Scholar
  20. C. Pornsing, M. S. Sodhi, and B. F. Lamond. 2016. Novel self-adaptive particle swarm optimization methods. Soft Computing 20, 9 (2016), 3579--3593.Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. P. Pudil, J. Novovičová, and J. Kittler. 1994. Floating search methods in feature selection. Pattern Recognition Letters 15, 11 (1994), 1119--1125.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. A. K. Qin, V. L. Huang, and P. N. Suganthan. 2009. Differential evolution algorithm with strategy adaptation for global numerical optimization. IEEE Transactions on Evolutionary Computation 13, 2 (2009), 398--417.Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. T. L. Saaty. 1990. How to make a decision: The analytic hierarchy process. European Journal of Operational Research 48, 1 (1990), 9--26.Google ScholarGoogle ScholarCross RefCross Ref
  24. S. D. Stearns. 1976. On selecting features for pattern classifiers. In 3rd International Joint Conference on Pattern Recognition. 71--75.Google ScholarGoogle Scholar
  25. R. Storn and K. Price. 1997. Differential evolution -- A simple and efficient heuristic for global optimization over continuous spaces. Journal of Global Optimization 11, 4 (1997), 341--359.Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. A. Stuhlsatz, J. Lippel, and T. Zielke. 2012. Feature extraction with deep neural networks by a generalized discriminant analysis. IEEE Transactions on Neural Networks and Learning Systems 23, 4 (2012), 596--608.Google ScholarGoogle ScholarCross RefCross Ref
  27. T. Sudo, K. Goto, Y. Nojima, and H. Ishibuchi. 2015. Effects of ensemble action selection with different usage of player’s memory resource on the evolution of cooperative strategies for iterated prisoner’s dilemma game. In IEEE Congress on Evolutionary Computation. 1505--1512.Google ScholarGoogle Scholar
  28. J. L. Tang and H. Liu. 2014. Feature selection for social media data. ACM Transactions on Knowledge Discovery from Data 8, 4 (2014), 1--27.Google ScholarGoogle Scholar
  29. H. Wang, H. Sun, C. H. Li, S. Rahnamayan, and J. S. Pan. 2013. Diversity enhanced particle swarm optimization with neighborhood search. Information Sciences 223 (2013), 119--135.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Y. Wang. 2011. Chaotic self-adaptive particle swarm optimization algorithm for dynamic economic dispatch problem with valve-point effects. Expert Systems with Applications 38, 11 (2011), 14231--14237.Google ScholarGoogle Scholar
  31. Y. Wang, B. Li, T. Weise, J. Y. Wang, B. Yuan, and Q. J. Tian. 2011. Self-adaptive learning based particle swarm optimization. Information Sciences 181, 20 (2011), 4515--4538.Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. A. W. Whitney. 1971. A direct method of nonparametric measurement selection. IEEE Transactions on Computers 100, 9 (1971), 1100--1103.Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. Y. Wu, S. C. H. Hoi, T. Mei, and N. H. Yu. 2017. Large-scale online feature selection for ultra-high dimensional sparse data. ACM Transactions on Knowledge Discovery from Data 11, 4 (2017), 1--22.Google ScholarGoogle Scholar
  34. R. Xu, G. C. Anagnostopoulos, and D. C. Wunsch. 2007. Multiclass cancer classification using semisupervised ellipsoid ARTMAP and particle swarm optimization with gene expression data. IEEE-ACM Transactions on Computational Biology and Bioinformatics 4, 1 (2007), 65--77.Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. B. Xue, W. N. Browne, M. J. Zhang, and X. Yao. 2016. A survey on evolutionary computation approaches to feature selection. IEEE Transactions on Evolutionary Computation 20, 4 (2016), 606--626.Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. B. Xue, M. J. Zhang, and W. N. Browne. 2013. Particle swarm optimization for feature selection in classification: A multi-objective approach. IEEE Transactions on Cybernetics 43, 6 (2013), 1656--1671.Google ScholarGoogle ScholarCross RefCross Ref
  37. B. Xue, M. J. Zhang, and W. N. Browne. 2014a. Particle swarm optimisation for feature selection in classification: Novel initialisation and updating mechanisms. Applied Soft Computing 18 (2014), 261--276.Google ScholarGoogle ScholarCross RefCross Ref
  38. Y. Xue, J. Jiang, B. Zhao, and T. H. Ma. 2017. A self-adaptive artificial bee colony algorithm based on global best for global optimization. Soft Computing 22, 9 (2017), 2935--2952.Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. Y. Xue, S. M. Zhong, Y. Zhuang, and B. Xu. 2014b. An ensemble algorithm with self-adaptive learning techniques for high-dimensional numerical optimization. Applied Mathematics and Computation 231 (2014), 329--346.Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. H. Q. Yang, M. R. Lyu, and I. King. 2013. Efficient online learning for multitask feature selection. ACM Transactions on Knowledge Discovery from Data 7, 2 (2013), 1--27.Google ScholarGoogle Scholar
  41. J. H. Yang and V. Honavar. 1998. Feature subset selection using a genetic algorithm. IEEE Intelligent Systems and their Applications 13, 2 (1998), 44--49.Google ScholarGoogle Scholar
  42. X. S. Yang. 2008. Nature-Inspired Metaheuristic Algorithms. LuniverPress.Google ScholarGoogle Scholar
  43. K. Yu, X. D. Wu, W. Ding, and J. Pei. 2016. Scalable and accurate online feature selection for big data. ACM Transactions on Knowledge Discovery from Data 11, 2 (2016), 39.Google ScholarGoogle Scholar
  44. Y. Zhang, D. W. Gong, and J. Cheng. 2017a. Multi-objective particle swarm optimization approach for cost-based feature selection in classification. IEEE-ACM Transactions on Computational Biology and Bioinformatics 14, 1 (2017), 64--75.Google ScholarGoogle ScholarDigital LibraryDigital Library
  45. Y. Zhang, D. W. Gong, Y. Hu, and W. Q. Zhang. 2015. Feature selection algorithm based on bare bones particle swarm optimization. Neurocomputing 148 (2015), 150--157.Google ScholarGoogle ScholarCross RefCross Ref
  46. Y. Zhang, X. F. Song, and D. W. Gong. 2017b. A return-cost-based binary firefly algorithm for feature selection. Information Sciences 418 (2017), 561--574.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Self-Adaptive Particle Swarm Optimization for Large-Scale Feature Selection in Classification

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image ACM Transactions on Knowledge Discovery from Data
        ACM Transactions on Knowledge Discovery from Data  Volume 13, Issue 5
        October 2019
        258 pages
        ISSN:1556-4681
        EISSN:1556-472X
        DOI:10.1145/3364623
        Issue’s Table of Contents

        Copyright © 2019 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 24 September 2019
        • Accepted: 1 June 2019
        • Revised: 1 April 2019
        • Received: 1 March 2018
        Published in tkdd Volume 13, Issue 5

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
        • Research
        • Refereed

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format .

      View HTML Format