运筹与管理 ›› 2024, Vol. 33 ›› Issue (6): 14-21.DOI: 10.12005/orms.2024.0175

• 理论分析与方法探讨 • 上一篇    下一篇

求解大规模优化问题的改进海洋捕食者算法

张文宇1,2, 袁永斌1, 高雪1, 张炳晨1   

  1. 1.西安邮电大学 经济与管理学院,陕西 西安 710061;
    2.中国航天系统科学与工程研究院,北京 100854
  • 收稿日期:2021-12-13 出版日期:2024-06-25 发布日期:2024-08-14
  • 通讯作者: 袁永斌(1998-),男,硕士研究生,研究方向:智能决策。
  • 作者简介:张文宇(1973-),女,山西万荣人,博士,教授,博士生导师,研究方向:智能决策,数据挖掘。
  • 基金资助:
    陕西省教育科研项目(08JK431)

Improved Marine Predators Algorithm for Large-scale Optimization Problems

ZHANG Wenyu1,2, YUAN Yongbin1, GAO Xue1, ZHANG Bingchen1   

  1. 1. College of Economics and Management, Xi’an University of Posts & Telecommunications, Xi’an 710061, China;
    2. China Research Institute of Aerospace Systems Science and Engineering, Beijing 100854, China
  • Received:2021-12-13 Online:2024-06-25 Published:2024-08-14

摘要: 针对海洋捕食者算法(Marine Predators Algorithm,MPA)在求解大规模问题时存在解精度低和易陷入局部最优的缺点,提出了一种改进的海洋捕食者算法(Improved Marine Predators Algorithm,IMPA)。采用Lloyd’s算法初始化种群使种群个体分布均匀;基于Q-learning算法自适应选择更新策略来平衡算法的探索和开发能力;通过反向学习机制提高种群多样性避免陷入局部最优。选取13个大规模(100维,500维和1000维)标准测试函数进行仿真实验,结果表明,IMPA算法在求解精度和收敛速度上优于其它对比算法。

关键词: 海洋捕食者算法, Lloyd’s算法, Q-learning机制, 反向学习

Abstract: Large-scale optimization problems have been around across various domains in real life. However, these problems exhibit high-dimensional variables and complex interdependencies among variables, rendering traditional optimization algorithms often ineffective and inefficient. Given that swarm intelligence optimization algorithms possess strong global search capabilities, inherent potential for parallelism, and distributive characteristics, they are more suitable for addressing large-scale optimization problems. Nevertheless, such algorithms typically suffer from difficulties with balancing exploration and exploitation stages and are prone to local optima. Hence, it is imperative to investigate improvement strategies for swarm intelligence optimization algorithms. The Marine Predators Algorithm (MPA) is a novel swarm intelligence optimization search algorithm inspired by the foraging behavior of marine predators in nature. It features simplicity in principle, ease of implementation, and minimal parameter settings. However, similar to other swarm intelligence algorithms, the MPA also has its drawbacks, necessitating corresponding improvements to enhance its effectiveness.
To address the shortcomings of the MPA in achieving low solution accuracy and susceptibility to local optima when solving large-scale optimization problems, this study first utilizes the Lloyd’s algorithm to initialize the prey population. This ensures an even distribution of individuals throughout the solution space, thereby enhancing the population’s global search capability. Subsequently, the three position update strategies of the MPA algorithm are employed as actions, with the number of offspring individuals surpassing their parents as the state, and the reduction in the optimization objective value as the reward. By employing the Q-learning algorithm, the optimal position update strategy is determined for each iteration, thereby balancing the exploration and exploitation processes of the algorithm and preventing it from being trapped in local optima. Additionally, a reverse operation is applied to each optimized individual after each iteration, obtaining the reverse solution of the current population. This expands the search space of the IMPA algorithm, enhances the diversity of the population, and effectively prevents the algorithm from falling into local optima.
Finally, this paper conducts a comparative analysis of the IMPA with several existing improved versions of the MPA across 13 high-dimensional test functions. The analysis includes algorithm complexity analysis, convergence speed, and optimization performance analysis, as well as a statistical characterization based on Wilcoxon tests. The results indicate that the IMPA algorithm outperforms other comparative algorithms in terms of solution accuracy and convergence speed on high-dimensional problems. It demonstrates superior convergence capability and solution accuracy in large-scale optimization problems.
Currently, the discussion of the performance of the IMPA algorithm is based solely on experiments conducted on high-dimensional test functions without testing the algorithm’s performance on actual large-scale optimization problems. The next step in research primarily involves applying the IMPA algorithm to specific large-scale constrained optimization problems and practical engineering issues.

Key words: Marine Predators Algorithm, Lloyd’s algorithm, Q-learning mechanism, opposition-based learning

中图分类号: