Improved Meta-ELM with error feedback incremental ELM as hidden nodes

Weidong Zou, Fenxi Yao*, Baihai Zhang, Zixiao Guan

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

15 Citations (Scopus)

Abstract

Liao et al. (Neurocomputing 128:81–87, 2014) proposed a meta-learning approach to extreme learning machine (Meta-ELM), which can obtain good generalization performance by training multiple ELMs. However, one of its open problems is overfitting when minimizing training error. In this paper, we propose an improved meta-learning model of ELM (improved Meta-ELM) to handle the problem. The improved Meta-ELM architecture is composed of some base ELMs which are error feedback incremental extreme learning machine (EFI-ELM) and the top ELM. The improved Meta-ELM includes two stages. First, each base ELM with EFI-ELM is trained on a subset of training data. Then, the top ELM learns with the base ELMs as hidden nodes. Simulation results on some artificial and benchmark datasets show that the proposed improved Meta-ELM model is more feasible and effective than Meta-ELM.

Original languageEnglish
Pages (from-to)3363-3370
Number of pages8
JournalNeural Computing and Applications
Volume30
Issue number11
DOIs
Publication statusPublished - 1 Dec 2018

Keywords

  • EFI-ELM
  • Heterogeneous
  • Meta-ELM
  • Overfitting

Fingerprint

Dive into the research topics of 'Improved Meta-ELM with error feedback incremental ELM as hidden nodes'. Together they form a unique fingerprint.

Cite this