Skip to main content
Log in

A Multimodal Optimization Algorithm Inspired by the States of Matter

  • Published:
Neural Processing Letters Aims and scope Submit manuscript

Abstract

The main objective of multi-modal optimization is to find multiple global and local optima for a problem in only one execution. Detecting multiple solutions to a multi-modal optimization formulation is especially useful in engineering, since the best solution could not represent the best realizable due to various practical restrictions. The States of Matter Search (SMS) is a recently proposed stochastic optimization technique. Although SMS is highly effective in locating single global optimum, it fails in providing multiple solutions within a single execution. To overcome this inconvenience, a new multimodal optimization algorithm called the Multi-modal States of Matter Search (MSMS) in introduced. Under MSMS, the original SMS is enhanced with new multimodal characteristics by means of: (1) the definition of a memory mechanism to efficiently register promising local optima according to their fitness values and the distance to other probable high quality solutions; (2) the modification of the original SMS optimization strategy to accelerate the detection of new local minima; and (3) the inclusion of a depuration procedure at the end of each state to eliminate duplicated memory elements. The performance of the proposed approach is compared to several state-of-the-art multimodal optimization algorithms considering a benchmark suite of fourteen multimodal problems. The results confirm that the proposed method achieves the best balance over its counterparts regarding accuracy and computational cost.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

References

  1. Panos P, Edwin R, Tuy H (2000) Recent developments and trends in global optimization. J Comput Appl Math 124:209–228

    Article  MathSciNet  MATH  Google Scholar 

  2. Floudas C, Akrotirianakis I, Caratzoulas S, Meyer C, Kallrath J (2005) Global optimization in the 21st century: advances and challenges. Comput Chem Eng 29(6):1185–1202

    Article  Google Scholar 

  3. Ying J, Ke-Cun Z, Shao-Jian Q (2007) A deterministic global optimization algorithm. Appl Math Comput 185(1):382–387

    MathSciNet  MATH  Google Scholar 

  4. Georgieva A, Jordanov I (2009) Global optimization based on novel heuristics, low-discrepancy sequences and genetic algorithms. Eur J Oper Res 196:413–422

    Article  MATH  Google Scholar 

  5. Lera D, Sergeyev Y (2010) Lipschitz and Hölder global optimization using space-filling curves. Appl Numer Math 60(1–2):115–129

    Article  MathSciNet  MATH  Google Scholar 

  6. Fogel LJ, Owens AJ, Walsh MJ (1966) Artificial intelligence through simulated evolution. Wiley, Chichester

    MATH  Google Scholar 

  7. Schwefel HP (2002) Evolution strategies: a comprehensive introduction. J Nat Comput 1(1):3–52

    Article  MathSciNet  MATH  Google Scholar 

  8. Koza JR (1990) Genetic programming: a paradigm for genetically breeding populations of computer programs to solve problems, report no. STAN-CS-90-1314. Stanford University, CA

  9. Holland JH (1975) Adaptation in natural and artificial systems. University of Michigan Press, Ann Arbor

    Google Scholar 

  10. Goldberg DE (1989) Genetic algorithms in search, optimization and machine learning. Addison Wesley, Boston

    MATH  Google Scholar 

  11. De Castro LN, Von Zuben FJ (1999) Artificial immune systems: part I—basic theory and applications. Technical report, TR-DCA 01/99

  12. Storn R, Price K (1995) Differential evolution-a simple and efficient adaptive scheme for global optimisation over continuous spaces, technical report TR-95-012, ICSI, Berkeley, CA

  13. Kirkpatrick S, Gelatt C, Vecchi M (1983) Optimization by simulated annealing. Science 220(4598):671–680

    Article  MathSciNet  MATH  Google Scholar 

  14. İlker B, Birbil S, Shu-Cherng F (2003) An electromagnetism-like mechanism for global optimization. J Glob Optim 25:263–282

    Article  MathSciNet  MATH  Google Scholar 

  15. Rashedia E, Nezamabadi-pour H, Saryazdi S (2011) Filter modeling using gravitational search algorithm. Eng Appl Artif Intell 24(1):117–122

    Article  Google Scholar 

  16. Kennedy J, Eberhart R (1995) Particle swarm optimization. In: Proceedings of the 1995 IEEE international conference on neural networks, vol 4, pp 1942–1948

  17. Dorigo M, Maniezzo V, Colorni A (1991) Positive feedback as a search strategy. Technical report no. 91-016. Politecnico di Milano

  18. Das S, Maity S, Qu BY, Suganthan PN (2011) Real-parameter evolutionary multimodal optimization—a survey of the state-of-the-art. Swarm Evol Comput 1(2):71–88

    Article  Google Scholar 

  19. Wong K-C, Chun-Ho W, Mok RKP, Peng C (2012) Evolutionary multimodal optimization using the principle of locality. Inf Sci 194:138–170

    Article  Google Scholar 

  20. Tan KC, Chiam SC, Mamun AA, Goh CK (2009) Balancing exploration and exploitation with adaptive variation for evolutionary multi-objective optimization. Eur J Oper Res 197:701–713

    Article  MATH  Google Scholar 

  21. Basak A, Das S, Chen-Tan K (2013) Multimodal optimization using a biobjective differential evolution algorithm enhanced with mean distance-based selection. IEEE Trans Evol Comput 17(5):666–685

    Article  Google Scholar 

  22. De Jong KA (1975) An analysis of the behavior of a class of genetic adaptive systems. Ph.D. dissertation. University of Michigan, Ann Arbor

  23. Goldberg DE, Richardson J (1987) Genetic algorithms with sharing for multimodal function optimization. In: Proceedings of 2nd international conference genetic algorithms, pp 41–49

  24. Petrowski AA (1996) Clearing procedure as a niching method for genetic algorithms. In: Proceedings of the 1996 IEEE international conference on evolutionary computation. IEEE Press, New York, pp 798–803

  25. Li J-P, Balazs ME, Parks GT, Clarkson PJ (2002) A species conserving genetic algorithm for multimodal function optimization. Evol Comput 10(3):207–234

    Article  Google Scholar 

  26. Mengshoel OJ, Galán SF, De Dios A (2014) Adaptive generalized crowding for genetic algorithms. Inf Sci 258:140–159

    Article  MathSciNet  Google Scholar 

  27. Miller BL, Shaw MJ (1996) Genetic algorithms with dynamic niche sharing for multimodal function optimization. In: Proceedings of the 3rd IEEE conference on evolutionary computation, pp 786–791

  28. Thomsen R (2004) Multimodal optimization using crowding-based differential evolution. In: Evolutionary computation, CEC2004. Congress

  29. Chen C-H, Liu T-K, Chou J-H (2014) A novel crowding genetic algorithm and its applications to manufacturing robots. IEEE Trans Ind Inf 10(3):1705–1716

    Article  Google Scholar 

  30. Yazdani S, Nezamabadi-pour H, Kamyab S (2014) A gravitational search algorithm for multimodal optimization. Swarm Evol Comput 14:1–14

    Article  Google Scholar 

  31. Chang W-D (2015) A modified particle swarm optimization with multiple subpopulations for multimodal function optimization problems. Appl Soft Comput 33:170–182

    Article  Google Scholar 

  32. Liang JJ, Qu BY, Mao XB, Niu B, Wang DY (2014) Differential evolution based on fitness Euclidean-distance ratio for multimodal optimization. Neurocomputing 137:252–260

    Article  Google Scholar 

  33. Biswas S, Das S, Kundu S, Patra GR (2014) Utilizing time-linkage property in DOPs: an information sharing based articial bee colony algorithm for tracking multiple optima in uncertain environments. Soft Comput 18:1199–1212

    Article  Google Scholar 

  34. Sacco WF, Henderson N, Rios-Coelho AC (2014) Topographical clearing differential evolution: a new method to solve multimodal. Prog Nucl Energy 71:269–278

    Article  Google Scholar 

  35. Lianga Y, Kwong-Sak L (2011) Genetic Algorithm with adaptive elitist-population strategies for multimodal function optimization. Appl Soft Comput 11:2017–2034

    Article  Google Scholar 

  36. Gao W, Yen GG, liu S (2014) Cluster-based differential evolution with self-adaptive strategy for multimodal optimization. IEEE Trans Cybern 44(8):1314–1327

    Article  Google Scholar 

  37. Qu BY, Suganthan PN, Das S (2013) A distance-based locally informed particle swarm model for multimodal optimization. IEEE Trans Evol Comput 17(3):387–402

    Article  Google Scholar 

  38. Dong W, Zhou M (2014) Gaussian classier-based evolutionary strategy for multimodal optimization. IEEE Trans Neural Netw Learn Syst 25(6):1200–1216

    Article  Google Scholar 

  39. Hui S, Suganthan PN (2016) Ensemble and arithmetic recombination-based speciation differential evolution for multimodal optimization. IEEE Trans Cybern 46(1):64–74

    Article  Google Scholar 

  40. Li L, Tang K (2015) History-based topological speciation for multimodal optimization. IEEE Trans Evol Comput 19(1):136–150

    Article  MathSciNet  Google Scholar 

  41. Chen G, Low CP, Yang Z (2009) Preserving and exploiting genetic diversity in evolutionary programming algorithms. IEEE Trans Evol Comput 13(3):661–673

    Article  Google Scholar 

  42. De Castro LN, Zuben FJ (2002) Learning and optimization using the clonal selection principle. IEEE Trans Evol Comput 6:239–251

    Article  Google Scholar 

  43. De Castro LN, Timmis J (2002) An artificial immune network for multimodal function optimization. In: Proceedings of the 2002 IEEE international conference on evolutionary computation. IEEE Press, New York, pp 699–704

  44. Xu Q, Lei W, Si J (2010) Predication based immune network for multimodal function optimization. Eng Appl Artif Intell 23:495–504

    Article  Google Scholar 

  45. Cuevas E, González M (2013) An optimization algorithm for multimodal functions inspired by collective animal behavior. Soft Comput 17(3):489–502

    Article  Google Scholar 

  46. Merrikh-Bayat F (2015) The runner-root algorithm: a metaheuristic for solving unimodal and multimodal optimization problems inspired by runners and roots of plants in nature. Appl Soft Comput 33:292–303

    Article  Google Scholar 

  47. Lacroix B, Molina D, Herrera F (2016) Region-based memetic algorithm with archive for multimodal optimisation. Inf Sci 367–368:719–746

    Article  Google Scholar 

  48. Roya S, Minhazul S, Das S, Ghosha S, Vasilakos AV (2013) A simulated weed colony system with subregional differential evolution for multimodal optimization. Eng Optim 45(4):459–481

    Article  MathSciNet  Google Scholar 

  49. Yahyaiea F, Filizadeh S (2011) A surrogate-model based multi-modal optimization algorithm. Eng Optim 43(7):779–799

    Article  Google Scholar 

  50. Cuevas E, Echavarría A, Ramírez-Ortegón MA (2014) An optimization algorithm inspired by the States of Matter that improves the balance between exploration and exploitation. Appl Intell 40(2):256–272

    Article  Google Scholar 

  51. Cuevas E, Echavarría A, Zaldívar D, Pérez-Cisneros M (2013) A novel evolutionary algorithm inspired by the states of matter for template matching. Expert Syst Appl 40(16):6359–6373

    Article  Google Scholar 

  52. Mohamed A-AA, El-Gaafary AAM, Mohamed YS, Hemeida AM (2016) Multi-objective states of matter search algorithm for TCSC-based smart controller design. Electr Power Syst Res 140:874–885

    Article  Google Scholar 

  53. Bailey RA (2004) Association schemes: designed experiments, algebra and combinatory. Cambridge University Press, Cambridge

    Book  MATH  Google Scholar 

  54. Barr RS, Kelly JP, Rescende MG, Stewart WR (1995) Designing and reporting on computational experiments with heuristic methods. J Heuristics 1:9–32

    Article  MATH  Google Scholar 

  55. Bartz-Beielstein T (2006) Experimental research in evolutionary computation—the new experimentalism. Natural computing series. Springer, Berlin

    MATH  Google Scholar 

  56. Batista E, França E, Borges M (2015) Improving the performance of metaheuristics: an approach combining response surface methodology and racing algorithms. Int J Eng Math 2015, Article ID 167031. https://doi.org/10.1155/2015/167031

  57. Batista E, França E (2017) Improving the fine-tuning of metaheuristics: an approach combining design of experiments and racing algorithms. J Optim 2017, Article ID 8042436. https://doi.org/10.1155/2017/8042436

  58. Calvet L, Juan A, Serrat C, Ries J (2016) A statistical learning based approach for parameter fine-tuning of metaheuristics. SORT Stat Oper Res Trans 40(1):201–224

    MathSciNet  MATH  Google Scholar 

  59. Eiben AE, Smit SK (2011) Parameter tuning for configuring and analyzing evolutionary algorithms. Swarm Evol Comput 1:19–31

    Article  Google Scholar 

  60. Eiben AE, Smit SK (2012) Evolutionary algorithm parameters and methods to tune them. In: Monfroy E, Hamadi Y, Saubion F (eds) Autonomous search. Springer, New York, pp 15–36

    Google Scholar 

  61. Giorgos K, Mark H, Eiben AE (2015) Parameter control in evolutionary algorithms: trends and challenges. IEEE Trans Evol Comput 19(2):167–187

    Article  Google Scholar 

  62. Kok KY, Rajendran P (2016) Differential-evolution control parameter optimization for unmanned aerial vehicle path planning. PLoS ONE 11(3):1–10

    Google Scholar 

  63. Ugolotti R, Cagnoni S (2014) Analysis of evolutionary algorithms using multi-objective parameter tuning, GECCO ’14. In: Proceedings of the annual conference on genetic and evolutionary computation, pp 1343–1350

  64. Kramer O, Gloger B, Gobels A (2007) An experimental analysis of evolution strategies and particle swarm optimisers using design of experiments. GECCO 07:674–681

    Google Scholar 

  65. Kramer O (2010) Evolutionary self-adaptation: a survey of operators and strategy parameters. Evol Intell 3(2):51–65

    Article  MATH  Google Scholar 

  66. Boari E, Pappa GL, Marques J, Goncalves MA, Meira W (2010) Tuning genetic programming parameters with factorial designs In: 2010 IEEE congress on evolutionary computation (CEC), pp 1–8

  67. Czarn A, MacNish C, Vijayan K, Turlach B, Gupta R (2004) Statistical exploratory analysis of genetic algorithms. IEEE Trans Evol Comput 8(4):405–421

    Article  Google Scholar 

  68. Petrovski A, Brownlee A, McCall J (2005) Statistical optimisation and tuning of GA factors. In: IEEE congress on evolutionary computation, vol 1, pp 758–764

  69. Beielstein T, Parsopoulos KE, Vrahatis MN (2002) Tuning PSO parameters through sensitivity analysis. Technical report, Reihe Computational Intelligence CI 124/02, Department of Computer Science, University of Dortmund

  70. Stodola P, Mazal J, Podhorec M (2015) Parameter tuning for the ant colony optimization algorithm used in ISR systems. Int J Appl Math Inform 9:123–126

    Google Scholar 

  71. Jackson W, Özcan E, John, R (2017) Tuning a simulated annealing metaheuristic for cross-domain search. In: IEEE congress on evolutionary computation 2017, 5–9 Donostia-San Sebastian, Spain

  72. Petrovski A, Wilson A, McCall J (1998) Statistical analysis of genetic algorithms and inference about optimal factors, School Computational Mathematical Science, Faculty of Science and Technology, The Robert Gordon University, Aberdeen, UK, technical report 2, SCMS technical report 1998/2

  73. Glover F (1989) Tabu search, part 1. ORSA J Comput 1(3):190–206

    Article  MATH  Google Scholar 

  74. Glover F (1990) Tabu search, part 2. ORSA J Comput 1(3):4–32

    Article  MATH  Google Scholar 

  75. Qu BY, Liang JJ, Wang ZY, Chen Q, Suganthan PN (2016) Novel benchmark functions for continuous multimodal optimization with comparative results. Swarm Evol Comput 26:23–34

    Article  Google Scholar 

  76. Garcia S, Molina D, Lozano M, Herrera F (2009) A study on the use of non-parametric tests for analyzing the evolutionary algorithms’ behaviour: a case study on the CEC’2005 special session on real parameter optimization. J Heuristics 15(6):617–644

    Article  MATH  Google Scholar 

  77. Wilcoxon F (1945) Individual comparisons by ranking methods. Biometrics 1:80–83

    Article  MathSciNet  Google Scholar 

  78. Xiaodong L, Engelbrecht A, Epitropakis MG (2013) Benchmark functions for CEC’2013 special session and competition on Niching methods for multimodal function optimization. 2013 IEEE congress on evolutionary computation (CEC), pp 1–10

  79. Clerc M, Kennedy J (2002) The particle swarm-explosion stability, and convergence in a multidimensional complex space. IEEE Trans Evol Comput 6(1):58–73

    Article  Google Scholar 

  80. Bernardino HS, Barbosa HJC, Fonseca LG (2011) Surrogate-assisted clonal selection algorithms for expensive optimization problems. Evol Intell 4(2):81–97

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Erik Cuevas.

Appendix A: List of Benchmark Functions

Appendix A: List of Benchmark Functions

See Tables 10 and 11.

Table 10 Low dimensional test functions used in the experimental study
Table 11 Composite test functions used in the experimental study

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Cuevas, E., Reyna-Orta, A. & Díaz-Cortes, MA. A Multimodal Optimization Algorithm Inspired by the States of Matter. Neural Process Lett 48, 517–556 (2018). https://doi.org/10.1007/s11063-017-9750-z

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11063-017-9750-z

Keywords

Navigation