Abstract

The Heap-Based Optimization (HBO), a novel meta-heuristic algorithm, constructs a hierarchical heap resembling a company structure to facilitate interaction among search agents. This enables the search agent to learn various levels of information from the population to obtain improved solutions, but it has drawbacks such as limited search capability and slow convergence speed when tackling complex optimization problems. To address these challenges, we propose an enhanced heap optimizer based on reinforcement learning(RLHBO), which demonstrates robust stability, rapid convergence, and high solution accuracy. Firstly, fixed operators are replaced with a reinforcement learning dynamic control search strategy in RLHBO to regulate the search range of agents. This prevents agents from being restricted in their search when encountering local or global optima, thus enhancing their search capability. Additionally, RLHBO introduces a dimensional self-learning strategy and a convergence strategy based on search direction similarity to address issues such as invalid and discrete solutions, accelerating convergence. Furthermore, an initialization strategy based on quasi-oppositional learning is employed to increase the diversity of the algorithm-generated population, surpassing the limitations of random initialization and further boosting convergence speed. Extensive testing on various complex functions from CEC2017 and CEC2022 demonstrates that RLHBO, as presented in this paper, outperforms HBO and numerous other advanced MAs in terms of search capability and convergence speed.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call