TY - CHAP
T1 - Ant lion optimizer
T2 - Theory, literature review, and application in multi-layer perceptron neural networks
AU - Heidari, Ali Asghar
AU - Faris, Hossam
AU - Mirjalili, Seyedali
AU - Aljarah, Ibrahim
AU - Mafarja, Majdi
PY - 2020/1/1
Y1 - 2020/1/1
N2 - This chapter proposes an efficient hybrid training technique (ALOMLP) based on the Ant Lion Optimizer (ALO) to be utilized in dealing with Multi-Layer Perceptrons (MLPs) neural networks. ALO is a well-regarded swarm-based meta-heuristic inspired by the intelligent hunting tricks of antlions in nature. In this chapter, the theoretical backgrounds of ALO are explained in details first. Then, a comprehensive literature review is provided based on recent well-established works from 2015 to 2018. In addition, a convenient encoding scheme is presented and the objective formula is defined, mathematically. The proposed training model based on ALO algorithm is substantiated on sixteen standard datasets. The efficiency of ALO is compared with differential evolution (DE), genetic algorithm (GA), particle swarm optimization (PSO), and population-based incremental learning (PBIL) in terms of best, worst, average, and median accuracies. Furthermore, the convergence propensities are monitored and analyzed for all competitors. The experiments show that the ALOMLP outperforms GA, PBIL, DE, and PSO in classifying the majority of datasets and provides improved accuracy results and convergence rates.
AB - This chapter proposes an efficient hybrid training technique (ALOMLP) based on the Ant Lion Optimizer (ALO) to be utilized in dealing with Multi-Layer Perceptrons (MLPs) neural networks. ALO is a well-regarded swarm-based meta-heuristic inspired by the intelligent hunting tricks of antlions in nature. In this chapter, the theoretical backgrounds of ALO are explained in details first. Then, a comprehensive literature review is provided based on recent well-established works from 2015 to 2018. In addition, a convenient encoding scheme is presented and the objective formula is defined, mathematically. The proposed training model based on ALO algorithm is substantiated on sixteen standard datasets. The efficiency of ALO is compared with differential evolution (DE), genetic algorithm (GA), particle swarm optimization (PSO), and population-based incremental learning (PBIL) in terms of best, worst, average, and median accuracies. Furthermore, the convergence propensities are monitored and analyzed for all competitors. The experiments show that the ALOMLP outperforms GA, PBIL, DE, and PSO in classifying the majority of datasets and provides improved accuracy results and convergence rates.
UR - http://www.scopus.com/inward/record.url?scp=85061337722&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-12127-3_3
DO - 10.1007/978-3-030-12127-3_3
M3 - Chapter
AN - SCOPUS:85061337722
SN - 978-3-030-12129-7
T3 - Studies in Computational Intelligence
SP - 23
EP - 46
BT - Studies in Computational Intelligence
PB - Springer Verlag
CY - Switzerland
ER -