4.6 Article

Supplementary-architecture weight-optimization neural networks

期刊

NEURAL COMPUTING & APPLICATIONS
卷 34, 期 13, 页码 11177-11197

出版社

SPRINGER LONDON LTD
DOI: 10.1007/s00521-022-07035-5

关键词

Artificial neural networks; Weight update equations; Supplementary architectures; Neuro-evolution; Grammatical evolution

资金

  1. National Research Foundation of South Africa [46712]

向作者/读者索取更多资源

The research in improving artificial neural networks has made significant progress, focusing on the architecture and weight update equations. However, the unexplored area of extending the definition of neural networks to include multiple architectures shows promising potential. Customizing weight update equations based on multiple architectures has been proven effective in enhancing learning ability.
Research efforts in the improvement of artificial neural networks have provided significant enhancements in learning ability, either through manual improvement by researchers or through automated design by other artificial intelligence techniques, and largely focusing on the architecture of the neural networks or the weight update equations used to optimize these architectures. However, a promising unexplored area involves extending the traditional definition of neural networks to allow a single neural network model to consist of multiple architectures, where one is a primary architecture and the others supplementary architectures. In order to use the information from all these architectures to possibly improve learning, weight update equations are customized per set-of-weights, and can each use the error of either the primary architecture or a supplementary architecture to update the values of that set-of-weights, with some necessary constraints to ensure valid updates. This concept was implemented and investigated. Grammatical evolution was used to make the complex architecture choices for each weight update equation, which succeeded in finding optimal choice combinations for classification and regression benchmark datasets, the KDD Cup 1999 intrusion detection dataset, and the UCLA graduate admission dataset. These optimal combinations were compared to traditional single-architecture neural networks, which they reliably outperformed at high confidence levels across all datasets. These optimal combinations were analysed using data mining tools, and this identified clear patterns, with the theoretical explanation provided as to how these patterns may be linked to optimality. The optimal combinations were shown to be competitive with state-of-the-art techniques on the same datasets.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据