4.7 Article

Hyperparameter self-tuning for data streams

Journal

INFORMATION FUSION
Volume 76, Issue -, Pages 75-86

Publisher

ELSEVIER
DOI: 10.1016/j.inffus.2021.04.011

Keywords

Data Streams; Optimisation; Hyperparameters

Funding

  1. National Funds through the FCT - Fundacao para a Ciencia e a Tecnologia, Portugal (Portuguese Foundation for Science and Technology) [UIDB/50014/2020]
  2. European Commission [820437]

Ask authors/readers for more resources

This article introduces SSPT, an extension of the Self Parameter Tuning (SPT) optimization algorithm for data streams, which dynamically adjusts sample sizes during a single pass over data and converges to optimal settings with relatively few hyperparameter configurations. The algorithm automatically readjusts hyperparameters when concept drift occurs, and experiments show that it outperforms previous human expert efforts in hyperparameter tuning.
The number of Internet of Things devices generating data streams is expected to grow exponentially with the support of emergent technologies such as 5G networks. Therefore, the online processing of these data streams requires the design and development of suitable machine learning algorithms, able to learn online, as data is generated. Like their batch-learning counterparts, stream-based learning algorithms require careful hyperparameter settings. However, this problem is exacerbated in online learning settings, especially with the occurrence of concept drifts, which frequently require the reconfiguration of hyperparameters. In this article, we present SSPT, an extension of the Self Parameter Tuning (SPT) optimisation algorithm for data streams. We apply the Nelder-Mead algorithm to dynamically-sized samples, converging to optimal settings in a single pass over data while using a relatively small number of hyperparameter configurations. In addition, our proposal automatically readjusts hyperparameters when concept drift occurs. To assess the effectiveness of SSPT, the algorithm is evaluated with three different machine learning problems: recommendation, regression, and classification. Experiments with well-known data sets show that the proposed algorithm can outperform previous hyperparameter tuning efforts by human experts. Results also show that SSPT converges significantly faster and presents at least similar accuracy when compared with the previous double-pass version of the SPT algorithm.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available