4.7 Article

The gradient evolution algorithm: A new metaheuristic

期刊

INFORMATION SCIENCES
卷 316, 期 -, 页码 246-265

出版社

ELSEVIER SCIENCE INC
DOI: 10.1016/j.ins.2015.04.031

关键词

Gradient-based method; Metaheuristic method; Optimization

向作者/读者索取更多资源

This study presents a new metaheuristic method that is derived from the gradient-based search method. In an exact optimization method, the gradient is used to find extreme points, as well as the optimal point. This study modifies a gradient method, and creates a metaheuristic method that uses a gradient theorem as its basic updating rule. This new method, named gradient evolution, explores the search space using a set of vectors and includes three major operators: vector updating, jumping and refreshing. Vector updating is the main updating rule in gradient evolution. The search direction is determined using the Newton-Raphson method. Vector jumping and refreshing enable this method to avoid local optima. In order to evaluate the performance of the gradient evolution method, three different experiments are conducted, using fifteen test functions. The first experiment determines the influence of parameter settings on the result. It also determines the best parameter setting. There follows a comparison between the basic and improved metaheuristic methods. The experimental results show that gradient evolution performs better than, or as well as, other methods, such as particle swarm optimization, differential evolution, an artificial bee colony and continuous genetic algorithm, for most of the benchmark problems tested. (C) 2015 Elsevier Inc. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据