An improved bees algorithm local search mechanism for numerical dataset

Bees Algorithm (BA), a heuristic optimization procedure, represents one of the fundamental search techniques is based on the food foraging activities of bees. This algorithm performs a kind of exploitative neighbourhoods search combined with random explorative search. However, the main issue of BA i...

Full description

Saved in:
Bibliographic Details
Main Author: Al-Dawoodi, Aras Ghazi Mohammed
Format: Thesis
Language:eng
eng
Published: 2015
Subjects:
Online Access:https://etd.uum.edu.my/5622/1/s813731_01.pdf
https://etd.uum.edu.my/5622/2/s813731_02.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
id my-uum-etd.5622
record_format uketd_dc
institution Universiti Utara Malaysia
collection UUM ETD
language eng
eng
advisor Mahmuddin, Massudi
topic QA75 Electronic computers
Computer science
spellingShingle QA75 Electronic computers
Computer science
Al-Dawoodi, Aras Ghazi Mohammed
An improved bees algorithm local search mechanism for numerical dataset
description Bees Algorithm (BA), a heuristic optimization procedure, represents one of the fundamental search techniques is based on the food foraging activities of bees. This algorithm performs a kind of exploitative neighbourhoods search combined with random explorative search. However, the main issue of BA is that it requires long computational time as well as numerous computational processes to obtain a good solution, especially in more complicated issues. This approach does not guarantee any optimum solutions for the problem mainly because of lack of accuracy. To solve this issue, the local search in the BA is investigated by Simple swap, 2-Opt and 3-Opt were proposed as Massudi methods for Bees Algorithm Feature Selection (BAFS). In this study, the proposed extension methods is 4-Opt as search neighbourhood is presented. This proposal was implemented and comprehensively compares and analyse their performances with respect to accuracy and time. Furthermore, in this study the feature selection algorithm is implemented and tested using most popular dataset from Machine Learning Repository (UCI). The obtained results from experimental work confirmed that the proposed extension of the search neighbourhood including 4-Opt approach has provided better accuracy with suitable time than the Massudi methods.
format Thesis
qualification_name masters
qualification_level Master's degree
author Al-Dawoodi, Aras Ghazi Mohammed
author_facet Al-Dawoodi, Aras Ghazi Mohammed
author_sort Al-Dawoodi, Aras Ghazi Mohammed
title An improved bees algorithm local search mechanism for numerical dataset
title_short An improved bees algorithm local search mechanism for numerical dataset
title_full An improved bees algorithm local search mechanism for numerical dataset
title_fullStr An improved bees algorithm local search mechanism for numerical dataset
title_full_unstemmed An improved bees algorithm local search mechanism for numerical dataset
title_sort improved bees algorithm local search mechanism for numerical dataset
granting_institution Universiti Utara Malaysia
granting_department Awang Had Salleh Graduate School of Arts & Sciences
publishDate 2015
url https://etd.uum.edu.my/5622/1/s813731_01.pdf
https://etd.uum.edu.my/5622/2/s813731_02.pdf
_version_ 1747827958976872448
spelling my-uum-etd.56222021-03-18T03:08:52Z An improved bees algorithm local search mechanism for numerical dataset 2015 Al-Dawoodi, Aras Ghazi Mohammed Mahmuddin, Massudi Awang Had Salleh Graduate School of Arts & Sciences Awang Had Salleh Graduate School of Arts and Sciences QA75 Electronic computers. Computer science Bees Algorithm (BA), a heuristic optimization procedure, represents one of the fundamental search techniques is based on the food foraging activities of bees. This algorithm performs a kind of exploitative neighbourhoods search combined with random explorative search. However, the main issue of BA is that it requires long computational time as well as numerous computational processes to obtain a good solution, especially in more complicated issues. This approach does not guarantee any optimum solutions for the problem mainly because of lack of accuracy. To solve this issue, the local search in the BA is investigated by Simple swap, 2-Opt and 3-Opt were proposed as Massudi methods for Bees Algorithm Feature Selection (BAFS). In this study, the proposed extension methods is 4-Opt as search neighbourhood is presented. This proposal was implemented and comprehensively compares and analyse their performances with respect to accuracy and time. Furthermore, in this study the feature selection algorithm is implemented and tested using most popular dataset from Machine Learning Repository (UCI). The obtained results from experimental work confirmed that the proposed extension of the search neighbourhood including 4-Opt approach has provided better accuracy with suitable time than the Massudi methods. 2015 Thesis https://etd.uum.edu.my/5622/ https://etd.uum.edu.my/5622/1/s813731_01.pdf text eng public https://etd.uum.edu.my/5622/2/s813731_02.pdf text eng public masters masters Universiti Utara Malaysia [1] B. Yuce, D. T. Pham, M. S. Packianather, and E. Mastrocinque, “An enhancement to the Bees Algorithm with slope angle computation and Hill Climbing Algorithm and its applications on scheduling and continuous-type optimisation problem,” Prod. Manuf. Res., vol. 3, no. 1, pp. 3–19, 2015. [2] M. Mahmuddin, “Optimisation Using Bees Algorithm on Unlabelled Data Problem,” Cardiff University, United Kingdom, 2009. [3] K. I. Rufai, R. C. Muniyandi, and Z. A. Othman, “Improving Bee Algorithm Based Feature Selection in Intrusion Detection System Using Membrane Computing,” J. Networks, vol. 9, no. 3, pp. 523–529, 2014. [4] N. M. Hammash, “Performance Comparison of Parallel Bees Algorithm on Rosenbrock Function,” Universiti Utara Malaysia, 2012. [5] D. T. Pham, A. Ghanbarzadeh, E. Koc, S. Otri, S. Rahim, and M. Zaidi, “The bees algorithm-a novel tool for complex optimisation problems,” in Proceedings of the 2nd Virtual International Conference on Intelligent Production Machines and Systems (IPROMS 2006), 2006, pp. 454–459. [6] A. Colorni, M. Dorigo, and V. Maniezzo, “Distributed optimization by ant colonies,” in Proceedings of the first European conference on artificial life, 1991, vol. 142, pp. 134–142. [7] X.-S. Yang, “A new metaheuristic bat-inspired algorithm,” in Nature inspired cooperative strategies for optimization (NICSO 2010), Springer, 2010, pp. 65–74. [8] J. Kennedy, “Particle swarm optimization,” in Encyclopedia of Machine Learning, Springer, 2010, pp. 760–766. [9] X.-S. Yang, Nature-inspired metaheuristic algorithms. Luniver press, 2010. [10] X.-S. Yang and S. Deb, “Cuckoo search via Lévy flights,” in Nature & Biologically Inspired Computing, 2009. NaBIC 2009. World Congress on, 2009, pp. 210–214. [11] I. R. M. Association, Bioinformatics: Concepts, Methodologies, Tools, and Applications. IGI Global, 2013. [12] A. Alazzam and H. W. Lewis III, “A New Optimization Algorithm For Combinatorial Problems,” Int. J. Adv. Res. Artif. Intell., vol. 2, no. 5, 2013. [13] J. F. Gantz, D. Reinsel, C. Chute, W. Schlichting, J. Mcarthur, S. Minton, I. Xheneti, A. Toncheva, and A. Manfrediz, “The Expanding Digital Universe: A Forecast of Worldwide Information Growth through 2010, IDC Whitepaper,” MA IDC, 2007. [14] J. W. Herrmann, “A genetic algorithm for minimax optimization problems,” in Evolutionary Computation, 1999. CEC 99. Proceedings of the 1999 Congress on, 1999, vol. 2, pp. 21–30. [15] redcedartech, “How to Select the Right Optimization Method for Your Problem.” [Online]. Available: http://www.redcedartech.com/pdfs/Select_Optimization_Method.pdf. [Accessed: 10-Oct-2015]. [16] M. R. Hossain, A. M. T. Oo, and A. B. M. Ali, “The Effectiveness of Feature Selection Method in Solar Power Prediction,” J. Renew. Energy, p. 9, 2013. [17] J. F. Gantz, “The Digital Universe of Opportunities: Rich Data and the Increasing Value of the Internet of Things,” EMC Corporation, 2015. [Online]. Available: Retrieved From. [18] D. T. Pham, M. Mahmuddin, S. Otri, and H. Al-Jabbouli, “Application of the bees algorithm to the selection features for manufacturing data,” 3rd Int. Virtual Conf. Intell. Prod. Mach. Syst., 2007. [19] I. Cheng, “Hybrid Methods for Feature Selection,” Western Kentucky University, 2013. [20] K. I. Rufai, R. C. Muniyandi, and Z. a. Othman, “Improving Bee Algorithm Based Feature Selection in Intrusion Detection System Using Membrane Computing,” J. Networks, vol. 9, no. 3, pp. 523–529, Mar. 2014. [21] M. S. IP.Manimaran, “Feature Subset Selection Algorithm for Large Volumes of Data Based on Clustering,” Int. J. Adv. Res. Comput. Sci. Technol., vol. 1, no. 2, 2014. [22] M. S. Packianather and B. Kapoor, “A wrapper-based feature selection approach using Bees Algorithm for a wood defect classification system,” in System of Systems Engineering Conference (SoSE), 2015 10th, 2015, pp. 498–503. [23] M. Mahmuddin and Y. Yusof, “A Hybrid Simplex Search and Bio-Inspired Algorithm for Faster Convergence,” in International Conference on Machine Learning and Computing, IACSIT Press, Singapore, 2009. [24] I. Mandli and M. Panchal, “Selection of Most Relevant Features from High Dimensional Data using IG-GA Hybrid Approach,” vol. 3, no. 2, pp. 827–830, 2014. [25] S. Alelyani, “On Feature Selection Stability: A Data Perspective,” Citeseer, Arizona State University, USA, 2013. [26] P. Gupta, S. Jain, and A. Jain, “A Review Of Fast Clustering-Based Feature Subset Selection Algorithm,” Int. J. Sci. Technol. Res., vol. 3, no. 11, pp. 20–26, 2014. [27] M. N. S. Sherin Mary Varghese1, “Efficient Feature Subset Selection Techniques for High Dimensional Data,” Int. J. Innov. Res. Comput. Commun. Eng., vol. 2, no. 3, 2014. [28] Q. Song, J. Ni, and G. Wang, “A fast clustering-based feature subset selection algorithm for high-dimensional data,” Knowl. Data Eng. IEEE Trans., vol. 25, no. 1, pp. 1–14, 2013. [29] S. Dreyer, “Evolutionary Feature Selection,” Norwegian University of Science and Technology, Norwegian, 2013. [30] L. Özbakir, A. Baykasoğlu, and P. Tapkan, “Bees algorithm for generalized assignment problem,” Appl. Math. Comput., vol. 215, no. 11, pp. 3782–3795, 2010. [31] S. Kashef and H. Nezamabadi-pour, “An advanced ACO algorithm for feature subset selection,” Neurocomputing, 2014. [32] Y. A. Alsultanny, “Database Preprocessing and Comparison between Data Mining Methods,” Int. J. New Comput. Archit. their Appl., vol. 1, no. 1, pp. 61–73, 2011. [33] W. Y. Tang, “Feature selection algorithms for very high dimensional data and mixed data,” School of Electrical and Electronic Engineering, singapore, 2008. [34] S. Abraham and I. Simi, “An Efficient CMST-Fuzzy Based Feature Selection Algorithm with Feature Interaction for High-Dimensional Data,” Int. J. Sci. Res. Educ., vol. 2, no. 07, 2014. [35] R. Tiwari and M. P. Singh, “Correlation-based attribute selection using genetic algorithm,” Int. J. Comput. Appl., vol. 4, no. 8, 2010. [36] V. Kumar and S. Minz, “Feature Selection,” SmartCR, vol. 4, no. 3, pp. 211–229, 2014. [37] K. Sigmund, C. Hauert, and M. A. Nowak, “Reward and punishment,” Proc. Natl. Acad. Sci., vol. 98, no. 19, pp. 10757–10762, 2001. [38] L. P. Kaelbling, M. L. Littman, and A. W. Moore, “Reinforcement learning: A survey,” J. Artif. Intell. Res., pp. 237–285, 1996. [39] A. George, “Analysis and Implementation of Lambda Opt Heuristic Framework for Combinatorial Optimization,” MSc in Artificial Intelligence, 2011. [40] D. T. Pham, E. Koc, J. Y. Lee, and J. Phrueksanant, “Using the bees algorithm to schedule jobs for a machine,” in Proceedings Eighth International Conference on Laser Metrology, CMM and Machine Tool Performance, LAMDAMAP, Euspen, UK, Cardiff, 2007, pp. 430–439. [41] D. T. Pham, E. Koç, A. Ghanbarzadeh, and S. Otri, “Optimisation of the weights of multi-layered perceptrons using the bees algorithm,” in Proceedings of 5th international symposium on intelligent manufacturing systems, 2006, pp. 38–46. [42] K. P. Bennett and E. Parrado-Hernández, “The interplay of optimization and machine learning research,” J. Mach. Learn. Res., vol. 7, pp. 1265–1281, 2006. [43] T. Eitrich and B. Lang, “Efficient optimization of support vector machine learning parameters for unbalanced datasets,” J. Comput. Appl. Math., vol. 196, no. 2, pp. 425–436, 2006. [44] D. T. Pham, S. Otri, A. Ghanbarzadeh, and E. Koc, “Application of the bees algorithm to the training of learning vector quantisation networks for control chart pattern recognition,” in Information and Communication Technologies, 2006. ICTTA’06. 2nd, 2006, vol. 1, pp. 1624–1629. [45] D. T. Pham, A. Ghanbarzadeh, E. Koc, and S. Otri, “Application of the bees algorithm to the training of radial basis function networks for control chart pattern recognition,” in Proceedings of 5th CIRP international seminar on intelligent computation in manufacturing engineering (CIRP ICME’06), Ischia, Italy, 2006, pp. 711–716. [46] X.-S. Yang, “Metaheuristic Optimization,” Scholarpedia, 08-Aug-2011. [Online]. Available: http://www.scholarpedia.org/article/ Metaheuristic_Optimization. [Accessed: 07-Apr-2015]. [47] A. Kaveh and A. Nasrollahi, “Charged system search and particle swarm optimization hybridized for optimal design of engineering structures,” Sci. Iran., vol. 21, no. 2, pp. 295–305, 2014. [48] T. Bartz, Beielstein, J. Branke, J. Mehnen, and O. Mersmann, “Evolutionary Algorithms,” Wiley Interdiscip. Rev. Data Min. Knowl. Discov., vol. 4, no. 3, pp. 178–195, 2014. [49] X.-S. Yang, Nature-Inspired Optimization Algorithms, 2014, Firs. Elsevier, 2014. [50] O. Olofintoye, J. Adeyemo, and F. Otieno, “Evolutionary Algorithms and Water Resources Optimization,” in EVOLVE-A Bridge between Probability, Set Oriented Numerics, and Evolutionary Computation II, Springer, 2013, pp. 491–506. [51] G. Fogel, D. Fogel, and L. Fogel, “Evolutionary programming,” Scholarpedia, 10-Apr-2011. [Online]. Available: http://www.scholarpedia.org/article/ Evolutionary_programming. [Accessed: 07-Apr- 2015]. [52] G. K. Brar and A. K. Virk, “Deployment of nodes for Maximum Coverage in Heterogeneous Wireless Sensor Network Using Genetic Algorithm,” Int. J., vol. 2, no. 6, 2014. [53] S. Lloyd, “The universe as quantum computer,” arXiv Prepr. arXiv1312.4455, 2013. [54] F. Glover, M. Laguna, and R. Martí, “Fundamentals of scatter search and path relinking,” Control Cybern., vol. 39, no. 3, pp. 653–684, 2000. [55] T. Chengyi, “Gravitational Search Algorithm Based on Simulated Annealing,” Int. Assoc. Information, Cult. Hum. Ind. Technol., vol. 9, no. 2, 2014. [56] S. C. Varma, K. S. L. Murthy, and K. SriChandan, “Gaussian particle swarm optimization for combined economic emission dispatch,” in Energy Efficient Technologies for Sustainability (ICEETS), 2013 International Conference on, 2013, pp. 136–140. [57] P. Wang, T. Weise, and R. Chiong, “Novel evolutionary algorithms for supervised classification problems: an experimental study,” Evol. Intell., vol. 4, no. 1, pp. 3–16, 2011. [58] B. Xing and W.-J. Gao, Innovative Computational Intelligence: A Rough Guide to 134 Clever Algorithms. Springer, 2014. [59] K.-L. Du and M. N. S. Swamy, Neural networks in a softcomputing framework. Springer, 124 2006. [60] V. Kumar, “Feature Selection: A literature Review,” Smart Comput. Rev., vol. 4, p. 3, 2014. [61] M. Shokouhifar and S. Sabet, “A hybrid approach for effective feature selection using neural networks and artificial bee colony optimization,” in 3rd international conference on machine vision (ICMV 2010), 2010, pp. 502–506. [62] E. M. Karabulut, S. A. Özel, and T. İbrikçi, “A comparative study on the effect of feature selection on classification accuracy,” Procedia Technol., vol. 1, pp. 323–327, 2012. [63] M. Kabir, M. Shahjahan, and K. Murase, Ant Colony Optimization Toward Feature Selection, 1st ed. InTech, 2013. [64] P.-E. Jouve and N. Nicoloyannis, “A filter feature selection method for clustering,” in Foundations of Intelligent Systems, Springer, 2005, pp. 583–593. [65] S. B. Kotsiantis, D. Kanellopoulos, and P. E. Pintelas, “Data Preprocessing for Supervised Leaning,” vol. 1, no. 2, pp. 111–117, 2006. [66] L. Yu and H. Liu, “Feature Selection for High-Dimensional Data : A Fast Correlation- Based Filter Solution,” 2003. [67] M. Kudo and J. Sklansky, “Comparison of algorithms that select features for pattern classifiers,” Pattern Recognit., vol. 33, no. 1, pp. 25–41, 2000. [68] M. Bressan and J. Vitria, “On the selection and classification of independent features,” Pattern Anal. Mach. Intell. IEEE Trans., vol. 25, no. 10, pp. 1312–1317, 2003. [69] S. J. Raudys and A. K. Jain, “Small sample size effects in statistical pattern recognition: Recommendations for practitioners,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 13, no. 3, pp. 252–264, 1991. [70] M. A. Hall and G. Holmes, “Benchmarking attribute selection techniques for discrete class data mining,” Knowl. Data Eng. IEEE Trans., vol. 15, no. 6, pp. 1437–1447, 2003. [71] L.-Y. Chuang, H.-W. Chang, C.-J. Tu, and C.-H. Yang, “Improved binary PSO for feature selection using gene expression data,” Comput. Biol. Chem., vol. 32, no. 1, pp. 29–38, 2008. [72] A. L. Blum and R. L. Rivest, “Training a 3-node neural network is NP-complete,” in Machine learning: From theory to applications, Springer, 1993, pp. 9–28. [73] H. Zhang and G. Sun, “Feature selection using tabu search method,” Pattern Recognit., vol. 35, no. 3, pp. 701–711, 2002. [74] R. Kohavi and G. H. John, “Wrappers for feature subset selection,” Artif. Intell., vol. 97, no. 1, pp. 273–324, 1997. [75] C. Sima and E. R. Dougherty, “What should be expected from feature selection in smallsample settings,” Bioinformatics, vol. 22, no. 19, pp. 2430–2436, 2006. [76] M. S. Packianather and P. R. Drake, “Neural networks for classifying images of wood veneer. Part 2,” Int. J. Adv. Manuf. Technol., vol. 16, no. 6, pp. 424–433, 2000. [77] J. Fan and R. Li, “Statistical challenges with high dimensionality: Feature selection in knowledge discovery,” arXiv Prepr. math/0602133, 2006. [78] M. Boulle, “Khiops: A statistical discretization method of continuous attributes,” Mach. Learn., vol. 55, no. 1, pp. 53–69, 2004. [79] W. O. Saxton, Tj. Pitt, and M. Horner, “Digital image processing: the Semper system,” Ultramicroscopy, vol. 4, no. 3, pp. 343–353, 1979. [80] I. Guyon and A. Elisseeff, “An introduction to variable and feature selection,” J. Mach. Learn. Res., vol. 3, pp. 1157–1182, 2003. [81] S. R. Hasani, Z. A. Othman, S. Mostafa, and M. Kahaki, “HYBRID FEATURE SELECTION ALGORITHM FOR INTRUSION DETECTION SYSTEM,” vol. 10, no. 6, pp. 1015–1025, 2014. [82] L.-X. Zhang, J.-X. Wang, Y.-N. Zhao, and Z.-H. Yang, “A novel hybrid feature selection algorithm: using ReliefF estimation for GA-Wrapper search,” in Machine Learning and Cybernetics, 2003 International Conference on, 2003, vol. 1, pp. 380–384. [83] J. Huang, Y. Cai, and X. Xu, “A hybrid genetic algorithm for feature selection wrapper based on mutual information,” Pattern Recognit. Lett., vol. 28, no. 13, pp. 1825–1844, 2007. [84] O. L. Garnes, “Feature Selection for Text Categorisation,” Institutt for datateknikk og informasjonsvitenskap, 2009. [85] S. NAQVI, “A Hybrid Filter-Wrapper Approach for FeatureSelection,” Örebro University, 2011. [86] Y. Saeys, I. Inza, and P. Larrañaga, “A review of feature selection techniques in bioinformatics,” bioinformatics, vol. 23, no. 19, pp. 2507–2517, 2007. [87] P. Shunmugapriya, S. Kanmani, R. Supraja, and K. Saranya, “Feature selection optimization through enhanced Artificial Bee Colony algorithm,” in Recent Trends in Information Technology (ICRTIT), 2013 International Conference on, 2013, pp. 56–61. [88] R. Forsati, A. Moayedikia, A. Keikha, and M. Shamsfard, “A novel approach for feature selection based on the bee colony optimization,” Int. J. Comput. Appl., vol. 43, no. 8, pp. 13–16, 2012. [89] M. S. Srivastava, M. N. Joshi, and M. Gaur, “A Review Paper on Feature Selection Methodologies and Their Applications,” IJCSNS, vol. 14, no. 5, p. 78, 2014. [90] S. Sakthivel, S. A. Pandiyan, S. Marikani, and S. K. Selvi, “Application of Big Bang Big Crunch Algorithm for Optimal Power Flow Problems,” Int J Eng Sci, vol. 2, no. 4, pp. 41–47, 2013. [91] H. M. Genc and A. K. Hocaoglu, “Bearing-only target tracking based on big bang–big crunch algorithm,” in Computing in the Global Information Technology, 2008. ICCGI’08. The Third International Multi-Conference on, 2008, pp. 229–233. [92] Y.-J. Zheng, S.-Y. Chen, Y. Lin, and W.-L. Wang, “Bio-inspired optimization of sustainable energy systems: a review,” Math. Probl. Eng., vol. 2013, 2013. [93] D. Rai and K. Tyagi, “Bio-inspired optimization techniques: a critical comparative study,” ACM SIGSOFT Softw. Eng. Notes, vol. 38, no. 4, pp. 1–7, 2013. [94] S. Binitha and S. S. Sathya, “A survey of bio inspired optimization algorithms,” Int. J. Soft Comput. Eng., vol. 2, no. 2, pp. 137–151, 2012. [95] A. Imanguliyev, “Enhancements for the Bees Algorithm.” Cardiff University, 2013. [96] J. Brownlee, Clever algorithms: nature-inspired programming recipes, 1st ed. Australia: Jason Brownlee, 2012. [97] E. Gerhardt and H. M. Gomes, “Artificial bee colony (ABC) algorithm for engineering optimization problems,” EngOpt, 2012. [98] D. Karaboga and B. Akay, “A comparative study of artificial bee colony algorithm,” Appl. Math. Comput., vol. 214, no. 1, pp. 108–132, 2009. [99] A. Baykasoglu, L. Ozbakir, and P. Tapkan, “Artificial bee colony algorithm and its application to generalized assignment problem,” Swarm Intell. Focus Ant Part. swarm Optim., pp. 113–144, 2007. [100] M. V. S. Prasad, C. H. S. Kumar, and T. Maneesha, “Feature Selection Using An Effective Dimensionality Reduction Technique,” vol. 3, no. 5, pp. 480 – 485, 2014. [101] Y.-C. Wu, W.-P. Lee, and C.-W. Chien, “Modified the performance of differential evolution algorithm with dual evolution strategy,” in 2009 International Conference on Machine Learning and Computing, IPCSIT, 2011, vol. 3, pp. 57–63. [102] M. Mahdiani, H. M. Judi, and N. S. Ashaari, “Selection of Shape Features in Control Chart Pattern Classification Using Bees Algorithm.,” Int. J. Inf. Process. Manag., vol. 5, no. 3, 2014. [103] M. Beekman and F. L. W. Ratnieks, “Long range foraging by the honey bee, Apis mellifera L.,” Funct. Ecol., vol. 14, no. 4, pp. 490–496, 2000. [104] D. T. Pham and A. Ghanbarzadeh, “Multi-objective optimisation using the bees algorithm,” in 3rd International Virtual Conference on Intelligent Production Machines and Systems (IPROMS 2007): Whittles, Dunbeath, Scotland, 2007, vol. 242, pp. 111–116. [105] D. T. Pham, A. Ghanbarzadeh, E. Koc, S. Otri, S. Rahim, and M. Zaidi, “The bees algorithm. Technical note,” Manuf. Eng. Centre, Cardiff Univ. UK, pp. 1–57, 2005. [106] D. T. Pham, S. Otri, A. Afify, M. Mahmuddin, and H. Al-Jabbouli, “Data clustering using the bees algorithm,” Liverpool, UK., 2007. [107] D. T. Pham and M. Castellani, “The bees algorithm: modelling foraging behaviour to solve continuous optimization problems,” Proc. Inst. Mech. Eng. Part C J. Mech. Eng. Sci., vol. 223, no. 12, pp. 2919–2938, 2009. [108] J. Phrueksanant, “Machine scheduling using the Bees Algorithm.” Cardiff University, 2013. [109] H. A. Le Thi, H. M. Le, and T. P. Dinh, “Feature selection in machine learning: an exact penalty approach using a Difference of Convex function Algorithm,” Mach. Learn., pp. 1–24, 2014. [110] M. S. K. K. Lenin, B. Ravindranath Reddy, “Dwindling of real power loss by using Improved Bees Algorithm,” Int. J. Recent Res. Electr. Electron. Eng., vol. 1, no. 1, pp. (34–42), 2014. [111] E. Mastrocinque, B. Yuce, A. Lambiase, and M. S. Packianather, “A multi-objective optimisation for supply chain network using the Bees Algorithm,” Int. J. Eng. Bus. Manag., vol. 5, pp. 1–11, 2013. [112] H. F. Wedde, M. Farooq, T. Pannenbaecker, B. Vogel, C. Mueller, J. Meth, R. Jeruschkat, M. Duhm, L. Bensmann, and G. Kathagen, “T. B” ning. BeeHive–An Energy-Aware Scheduling and Routing Framework,” Technical report-pg439, LSIII, School of Computer Science, University of Dortmund, 2004. [113] R. Martí and G. Reinelt, “Heuristic Methods,” in The Linear Ordering Problem, Springer, 2011, pp. 17–40. [114] M. Gruber, “Exact and heuristic approaches for solving the bounded diameter minimum spanning tree problem,” Vienna, Austria, 2009. [115] C. Blum, J. Puchinger, G. R. Raidl, and A. Roli, “Hybrid metaheuristics in combinatorial optimization: A survey,” Appl. Soft Comput., vol. 11, no. 6, pp. 4135–4151, 2011. [116] C. Blum and A. Roli, “Metaheuristics in combinatorial optimization: Overview and conceptual comparison,” ACM Comput. Surv., vol. 35, no. 3, pp. 268–308, 2003. [117] J. Fosin, D. Davidović, and T. Carić, “A gpu implementation of local search operators for symmetric travelling salesman problem,” PROMET-Traffic&Transportation, vol. 25, no. 3, pp. 225–234, 2013. [118] G. A. Croes, “A method for solving traveling-salesman problems,” Oper. Res., vol. 6, no. 6, pp. 791–812, 1958. [119] M. M. Flood, “The traveling-salesman problem,” Oper. Res., vol. 4, no. 1, pp. 61–75, 1956. [120] D. S. Johnson and L. A. McGeoch, “The traveling salesman problem: A case study in local optimization,” Local search Comb. Optim., vol. 1, pp. 215–310, 1997. [121] F. Bock, “An algorithm for solving’ traveling salesman’and related network optimization problems, presented at the 14th National Meeting of the Operations Research Society of America, St,” Louis, Missouri, 1958. 129 [122] S. Lin, “Computer solutions of the traveling salesman problem,” Bell Syst. Tech. Journal, vol. 44, no. 10, pp. 2245–2269, 1965. [123] C. Walshaw, “A Multilevel Approach to the Travelling Salesman Problem,” INFORMS, 2002. [124] M. Kordos, “Search-based algorithms for multilayer perceptrons.” The Silesian University of Technology, 2005. [125] E. Romero and J. M. Sopena, “Performing feature selection with multilayer perceptrons,” Neural Networks, IEEE Trans., vol. 19, no. 3, pp. 431–441, 2008. [126] G. Kumar and K. Kumar, “A Multi-objective Genetic Algorithm Based Approach for Effective Intrusion Detection Using Neural Networks,” in Intelligent Methods for Cyber Warfare, Springer, 2015, pp. 173–200. [127] S. Ravuri and A. Stolcke, “Neural Network Models for Lexical Addressee Detection,” in Fifteenth Annual Conference of the International Speech Communication Association, 2014. [128] Z. N. S. Vanini, K. Khorasani, and N. Meskin, “Fault detection and isolation of a dual spool gas turbine engine using dynamic neural networks and multiple model approach,” Inf. Sci. (Ny)., vol. 259, pp. 234–251, 2014. [129] E. Gasca, J. S. Sánchez, and R. Alonso, “Eliminating redundancy and irrelevance using a new MLP-based feature selection method,” Pattern Recognit., vol. 39, no. 2, pp. 313–315, 2006. [130] A. Verikas and M. Bacauskiene, “Feature selection with neural networks,” Pattern Recognit. Lett., vol. 23, no. 11, pp. 1323–1335, 2002. [131] G. Castellano and A. M. Fanelli, “Variable selection using neural-network models,” Neurocomputing, vol. 31, no. 1, pp. 1–13, 2000. [132] M. M. Kabir, M. M. Islam, and K. Murase, “A new wrapper feature selection approach using neural network,” Neurocomputing, vol. 73, no. 16, pp. 3273–3283, 2010. [133] R. Chakraborty and N. R. Pal, “Feature selection using a neural framework with controlled redundancy,” 2014. [134] C. Chang, P. A. Verhaegen, and J. R. Duflou, “A Comparison of Classifiers for Intelligent Machine Usage Prediction,” in Intelligent Environments (IE), 2014 International Conference on, 2014, pp. 198–201. [135] Z. Yan, Z. Wang, and H. Xie, “The application of mutual information-based feature selection and fuzzy LS-SVM-based classifier in motion classification,” Comput. Methods Programs Biomed., vol. 90, no. 3, pp. 275–284, 2008. [136] G. Kim, Y. Kim, H. Lim, and H. Kim, “An MLP-based feature subset selection for HIV- 1 protease cleavage site analysis,” Artif. Intell. Med., vol. 48, no. 2, pp. 83–89, 2010. [137] R. Setiono and H. Liu, “Neural-network feature selector,” Neural Networks, IEEE Trans., vol. 8, no. 3, pp. 654–662, 1997. [138] C.-N. Hsu, H.-J. Huang, and S. Dietrich, “The ANNIGMA-wrapper approach to fast feature selection for neural nets,” Syst. Man, Cybern. Part B Cybern. IEEE Trans., vol. 32, no. 2, pp. 207–212, 2002. [139] V. Sindhwani, S. Rakshit, D. Deodhare, D. Erdogmus, J. C. Principe, and P. Niyogi, “Feature selection in MLPs and SVMs based on maximum output information,” Neural Networks, IEEE Trans., vol. 15, no. 4, pp. 937–948, 2004. [140] D. M. Santoro, E. R. Hruschska Jr, and M. do Carmo Nicoletti, “Selecting feature subsets for inducing classifiers using a committee of heterogeneous methods,” in Systems, Man and Cybernetics, 2005 IEEE International Conference on, 2005, vol. 1, pp. 375–380. [141] A. L. Blum and P. Langley, “Selection of relevant features and examples in machine learning,” Artif. Intell., vol. 97, no. 1, pp. 245–271, 1997. [142] S. Piramuthu, “Evaluating feature selection methods for learning in data mining applications,” Eur. J. Oper. Res., vol. 156, no. 2, pp. 483–494, 2004. [143] P. M. Narendra and K. Fukunaga, “A branch and bound algorithm for feature subset selection,” Comput. IEEE Trans., vol. 100, no. 9, pp. 917–922, 1977. [144] A. Jain and D. Zongker, “Feature selection: Evaluation, application, and small sample performance,” Pattern Anal. Mach. Intell. IEEE Trans., vol. 19, no. 2, pp. 153–158, 1997. [145] H. Vafaie and K. De Jong, “Genetic algorithms as a tool for feature selection in machine learning,” in Tools with Artificial Intelligence, 1992. TAI’92, Proceedings., Fourth International Conference on, 1992, pp. 200–203. [146] J. Feng, Y. Yang, H. Wang, and X.-M. Wang, “Feature selection based on genetic algorithms and support vector machines for handwritten similar Chinese characters recognition,” in Machine Learning and Cybernetics, 2004. Proceedings of 2004 International Conference on, 2004, vol. 6, pp. 3600–3605. [147] H. Handels, T. Roß, J. Kreusch, H. H. Wolff, and S. J. Poeppl, “Feature selection for optimized skin tumor recognition using genetic algorithms,” Artif. Intell. Med., vol. 16, no. 3, pp. 283–297, 1999. [148] D. J. Skarzynski, M. M. Bah, K. M. Deptula, I. Woclawek-Potocka, A. Korzekwa, M. Shibaya, W. Pilawski, and K. Okuda, “Roles of tumor necrosis factor-α of the estrous cycle in cattle: an in vivo study,” Biol. Reprod., vol. 69, no. 6, pp. 1907–1913, 2003. [149] P. L. Lanzi, “Fast feature selection with genetic algorithms: a filter approach,” in Evolutionary Computation, 1997., IEEE International Conference on, 1997, pp. 537– 540. [150] S. F. Da Silva, M. X. Ribeiro, J. D. E. S. B. Neto, C. Traina-Jr, and A. J. M. Traina, “Improving the ranking quality of medical image retrieval using a genetic feature selection method,” Decis. Support Syst., vol. 51, no. 4, pp. 810–820, 2011. [151] W. F. Punch III, E. D. Goodman, M. Pei, L. Chia-Shun, P. D. Hovland, and R. J. Enbody, “Further Research on Feature Selection and Classification Using Genetic Algorithms.,” in ICGA, 1993, pp. 557–564. [152] J. Yang and V. Honavar, “Feature subset selection using a genetic algorithm,” in Feature extraction, construction and selection, Springer, 1998, pp. 117–136. [153] P. Zhang, B. Verma, and K. Kumar, “Neural vs. statistical classifier in conjunction with genetic algorithm based feature selection,” Pattern Recognit. Lett., vol. 26, no. 7, pp. 909–919, 2005. [154] D. P. Muni, N. R. Pal, and J. Das, “Genetic programming for simultaneous feature selection and classifier design,” Syst. Man, Cybern. Part B Cybern. IEEE Trans., vol. 36, no. 1, pp. 106–117, 2006. [155] J. Yu, J. Yu, A. A. Almal, S. M. Dhanasekaran, D. Ghosh, W. P. Worzel, and A. M. Chinnaiyan, “Feature selection and molecular classification of cancer using genetic programming,” Neoplasia, vol. 9, no. 4, pp. 292–IN3, 2007. [156] R. J. Nandi, A. K. Nandi, R. M. Rangayyan, and D. Scutt, “Classification of breast masses in mammograms using genetic programming and feature selection,” Med. Biol. Eng. Comput., vol. 44, no. 8, pp. 683–694, 2006. [157] A. Seal, S. Ganguly, D. Bhattacharjee, M. Nasipuri, and C. Gonzalo-Martin, “Feature Selection using Particle Swarm Optimization for Thermal Face Recognition,” in Applied Computation and Security Systems, Springer, 2015, pp. 25–35. [158] H. B. Nguyen, B. Xue, I. Liu, and M. Zhang, “Filter based backward elimination in wrapper based PSO for feature selection in classification,” in Evolutionary Computation (CEC), 2014 IEEE Congress on, 2014, pp. 3111–3118. [159] Y. Zhang, S. Wang, P. Phillips, and G. Ji, “Binary PSO with mutation operator for feature selection using decision tree applied to spam detection,” Knowledge-Based Syst., vol. 64, pp. 22–31, 2014. [160] C.-J. Tu, L.-Y. Chuang, J.-Y. Chang, and C.-H. Yang, “Feature selection using PSOSVM,” IAENG Int. J. Comput. Sci., vol. 33, no. 1, pp. 111–116, 2007. [161] X. Wang, J. Yang, X. Teng, W. Xia, and R. Jensen, “Feature selection based on rough sets and particle swarm optimization,” Pattern Recognit. Lett., vol. 28, no. 4, pp. 459–471, 2007. [162] H. A. Firpi and E. D. Goodman, “Swarmed Feature Selection.,” in AIPR, 2004, pp. 112–118. [163] M. Karnan, K. Thangavel, R. Sivakuar, and K. Geetha, “Ant colony optimization for feature selection and classification of microcalcifications in digital mammograms,” in Advanced Computing and Communications, 2006. ADCOM 2006. International Conference on, 2006, pp. 298–303. [164] R. Bello, A. Puris, A. Nowe, Y. Martínez, and M. M. Garcia, “Two step ant colony system to solve the feature selection problem,” in Progress in Pattern Recognition, Image Analysis and Applications, Springer, 2006, pp. 588–596. [165] Y. H. Gai and G. Yu, “A Hybrid Feature Selection Based on Ant Colony Optimization and Probabilistic Neural Networks for Bearing Fault Diagnostics,” Appl. Mech. Mater., vol. 10, pp. 573–577, 2008. [166] W. S. McCulloch and W. Pitts, “A logical calculus of the ideas immanent in nervous activity,” Bull. Math. Biol., vol. 52, no. 1–2, pp. 99–115, 1990. [167] M. S. Uzer, N. Yilmaz, and O. Inan, “Feature selection method based on artificial bee colony algorithm and support vector machines for medical datasets classification,” Sci. World J., vol. 2013, 2013. [168] “UCI Machine Learning Repository: Data Sets,” UCI, 2015. [Online]. Available: https://archive.ics.uci.edu/ml/datasets.html. [Accessed: 10-Oct-2014]. [169] Y.-S. Hwang, “Wrapper-based Feature Selection Using Support Vector Machine,” Life Sci. J., vol. 11, no. 7, 2014. [170] “UCI Machine Learning Repository: Heart Disease Data Set,” UCI, 2015. [Online]. Available: https://archive.ics.uci.edu/ml/ datasets/Heart+Disease. [Accessed: 10-Oct-2014]. [171] “UCI Machine Learning Repository: Pima Indians Diabetes Data Set.” [Online]. Available: https://archive.ics.uci.edu/ml/ datasets/Pima+Indians+Diabetes. [Accessed: 24-Nov-2015]. [172] “UCI Machine Learning Repository: SPECT Heart Data Set,” University of California, Irvine, 2015. [Online]. Available: https://archive.ics.uci.edu/ml/datasets/SPECT+ Heart.[Accessed: 19-Apr-2015]. [173] “UCI Machine Learning Repository: Hepatitis Data Set.” [Online]. Available: https://archive.ics.uci.edu/ml/dataset/Hepatitis . [Accessed: 24-Nov-2015]. [174] “UCI Machine Learning Repository: Soybean (Small) Data Set,” University of California, Irvine, 2015. [Online]. Available: https://archive.ics.uci.edu/ml/datasets/Soybean +%28Small%29. [Accessed: 19-Apr-2015]. [175] “UCI Machine Learning Repository: Lung Cancer Data Set,” University of California, Irvine, 2015. [Online]. Available: https://archive.ics.uci.edu/ml/datasets/ Lung+Cancer. [Accessed: 19-Apr-2015]. [176] R. S. Michalski and R. L. Chilausky, “Learning by being told and learning from examples: An experimental comparison of the two methods of knowledge acquisition in the context of developing an expert system for soybean disease diagnosis,” Int. J. Policy Anal. Inf. Syst., vol. 4, no. 2, pp. 125–161, 1980. [177] L. A. Kurgan, K. J. Cios, R. Tadeusiewicz, M. Ogiela, and L. S. Goodenday, “Knowledge discovery approach to automated cardiac SPECT diagnosis,” Artif. Intell. Med., vol. 23, no. 2, pp. 149–169, 2001. [178] Z.-Q. Hong and J.-Y. Yang, “Optimal discriminant plane for a small number of samples and design method of classifier on the plane,” pattern Recognit., vol. 24, no. 4, pp. 317–324, 1991. [179] A. Blazinskas and A. Misevicius, “COMBINING 2-OPT, 3-OPT AND 4-OPT WITH KSWAP- KICK PERTURBATIONS FOR THE TRAVELING SALESMAN PROBLEM,” Kaunas Univ. Technol. Dep. Multimed. Eng. Studentu St, pp. 50–401, 2009. [180] D. Bonachea, E. Ingerman, J. Levy, and S. McPeak, “An Improved Adaptive Multi-Start Approach to Finding Near-Optimal Solutions to the Euclidean TSP.,” in GECCO, 2000, pp. 143–150. [181] E. Koc, “Bees Algorithm: theory, improvements and applications.” Cardiff University, 2010. [182] S. Lin and B. W. Kernighan, “An Effective Heuristic Algorithm for the Traveling-Salesman Problem,” Oper. Res., vol. 21, no. 2, pp. 498–516, Apr. 1973. [183] F. Glover, “Finding a best traveling salesman 4-opt move in the same time as a best 2-opt move,” J. Heuristics, vol. 2, no. 2, pp. 169–179, 1996.