4.7 Article

HAWKS: Evolving Challenging Benchmark Sets for Cluster Analysis

Journal

IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION
Volume 26, Issue 6, Pages 1206-1220

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TEVC.2021.3137369

Keywords

Benchmarking; clustering; data generator; evolutionary computation; synthetic data

Funding

  1. EPSRC Manchester Centre for Doctoral Training in Computer Science [EP/I028099/1]

Ask authors/readers for more resources

Comprehensive benchmarking of clustering algorithms is challenging due to the lack of a unique mathematical definition for this unsupervised learning approach and dependencies between generating models or clustering criteria and validation indices. The role of synthetic datasets in evaluating clustering algorithms is emphasized, and the need to construct benchmarks that cover diverse properties impacting performance is highlighted.
Comprehensive benchmarking of clustering algorithms is rendered difficult by two key factors: 1) the elusiveness of a unique mathematical definition of this unsupervised learning approach and 2) dependencies between the generating models or clustering criteria adopted by some clustering algorithms and indices for internal cluster validation. Consequently, there is no consensus regarding the best practice for rigorous benchmarking, and whether this is possible at all outside the context of a given application. Here, we argue that synthetic datasets must continue to play an important role in the evaluation of clustering algorithms, but that this necessitates constructing benchmarks that appropriately cover the diverse set of properties that impact clustering algorithm performance. Through our framework, HAWKS, we demonstrate the important role evolutionary algorithms play to support flexible generation of such benchmarks, allowing simple modification and extension. We illustrate two possible uses of our framework: 1) the evolution of benchmark data consistent with a set of hand-derived properties and 2) the generation of datasets that tease out performance differences between a given pair of algorithms. Our work has implications for the design of clustering benchmarks that sufficiently challenge a broad range of algorithms, and for furthering insight into the strengths and weaknesses of specific approaches.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available