4.6 Article

Modeling collinear data using double-layer GA-based selective ensemble kernel partial least squares algorithm

Journal

NEUROCOMPUTING
Volume 219, Issue -, Pages 248-262

Publisher

ELSEVIER
DOI: 10.1016/j.neucom.2016.09.019

Keywords

Collinear and nonlinear data modeling; Latent variable (LV); Selective ensemble learning; Double-layer genetic algorithm (DLGA) optimization; Kernel partial least squares (KPLS)

Funding

  1. post doctoral National Natural Science Foundation of China [2013M532118, 2015T81082, 2015M581355]
  2. National Natural Science Foundation of China [61573364, 61273177, 61305029, 61503066, 61573249]
  3. State Key Laboratory of Synthetical Automation for Process Industries [PAL-N201504]
  4. China National 863 Projects [2015AA043802]
  5. Priority Academic Program Development of Jiangsu Higher Education Institutions (PAPD)
  6. Jiangsu Collaborative Innovation Center on Atmospheric Environment and Equipment Technology (CICAEET) fund
  7. Liaoning Province Doctor Startup Fund [201501151]

Ask authors/readers for more resources

Collinear and nonlinear characteristics of modeling data have to be addressed for constructing effective soft measuring models. Latent variables (LVs)-based modeling approaches, such as kernel partial least squares (KPLS), can overcome these disadvantages in certain degree. Selective ensemble (SEN) modeling can improve generalization performance of learning models further. Nevertheless, how to select SEN model's learning parameters is an important open issue. In this paper, a novel SENKPLS modeling method based on double-layer genetic algorithm (DLGA) optimization is proposed. At first, one mechanism, titled outside layer adaptive GA (AGA) optimization encoding and decoding principle, is employed to produce initial learning parameter values for KPLS-based candidate-sub-models. Then, ensemble sub-models are selected and combined based on inside layer GA optimization toolbox (GAOT) and adaptive weighting fusion (AWF) algorithm. Thus, SEN models of all AGA populations are obtained. Finally, outside layer AGA optimization operations, i.e., selection, crossover and mutation processes, are repeated until the pre-set stopping criterion is satisfied. Simulation results validate the effectiveness of the proposed method as far as the synthetic data, low dimensional and high dimensional benchmark data.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available