Loading [MathJax]/extensions/tex2jax.js

Algorithm Selection as Superset Learning: Constructing Algorithm Selectors from Imprecise Performance Data

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Authors

  • Jonas Hanselle
  • Alexander Tornede
  • Marcel Wever
  • Eyke Hüllermeier

External Research Organisations

  • Paderborn University
  • Heinz Nixdorf Institute

Details

Original languageEnglish
Title of host publicationAdvances in Knowledge Discovery and Data Mining - 25th Pacific-Asia Conference, PAKDD 2021, Proceedings
EditorsKamal Karlapalem, Hong Cheng, Naren Ramakrishnan, R. K. Agrawal, P. Krishna Reddy, Jaideep Srivastava, Tanmoy Chakraborty
PublisherSpringer Science and Business Media Deutschland GmbH
Pages152-163
Number of pages12
ISBN (print)9783030757618
Publication statusPublished - 2021
Externally publishedYes
Event25th Pacific-Asia Conference on Knowledge Discovery and Data Mining, PAKDD 2021 - Virtual, Online
Duration: 11 May 202114 May 2021

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume12712 LNAI
ISSN (Print)0302-9743
ISSN (electronic)1611-3349

Abstract

Algorithm selection refers to the task of automatically selecting the most suitable algorithm for solving an instance of a computational problem from a set of candidate algorithms. Here, suitability is typically measured in terms of the algorithms’ runtimes. To allow the selection of algorithms on new problem instances, machine learning models are trained on previously observed performance data and then used to predict the algorithms’ performances. Due to the computational effort, the execution of such algorithms is often prematurely terminated, which leads to right-censored observations representing a lower bound on the actual runtime. While simply neglecting these censored samples leads to overly optimistic models, imputing them with precise though hypothetical values, such as the commonly used penalized average runtime, is a rather arbitrary and biased approach. In this paper, we propose a simple regression method based on so-called superset learning, in which right-censored runtime data are explicitly incorporated in terms of interval-valued observations, offering an intuitive and efficient approach to handling censored data. Benchmarking on publicly available algorithm performance data, we demonstrate that it outperforms the aforementioned naïve ways of dealing with censored samples and is competitive to established methods for censored regression in the field of algorithm selection.

Keywords

    Algorithm selection, Censored data, Superset learning

ASJC Scopus subject areas

Cite this

Algorithm Selection as Superset Learning: Constructing Algorithm Selectors from Imprecise Performance Data. / Hanselle, Jonas; Tornede, Alexander; Wever, Marcel et al.
Advances in Knowledge Discovery and Data Mining - 25th Pacific-Asia Conference, PAKDD 2021, Proceedings. ed. / Kamal Karlapalem; Hong Cheng; Naren Ramakrishnan; R. K. Agrawal; P. Krishna Reddy; Jaideep Srivastava; Tanmoy Chakraborty. Springer Science and Business Media Deutschland GmbH, 2021. p. 152-163 (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Vol. 12712 LNAI).

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Hanselle, J, Tornede, A, Wever, M & Hüllermeier, E 2021, Algorithm Selection as Superset Learning: Constructing Algorithm Selectors from Imprecise Performance Data. in K Karlapalem, H Cheng, N Ramakrishnan, RK Agrawal, PK Reddy, J Srivastava & T Chakraborty (eds), Advances in Knowledge Discovery and Data Mining - 25th Pacific-Asia Conference, PAKDD 2021, Proceedings. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 12712 LNAI, Springer Science and Business Media Deutschland GmbH, pp. 152-163, 25th Pacific-Asia Conference on Knowledge Discovery and Data Mining, PAKDD 2021, Virtual, Online, 11 May 2021. https://doi.org/10.1007/978-3-030-75762-5_13, https://doi.org/10.1007/978-3-030-75762-5_13
Hanselle, J., Tornede, A., Wever, M., & Hüllermeier, E. (2021). Algorithm Selection as Superset Learning: Constructing Algorithm Selectors from Imprecise Performance Data. In K. Karlapalem, H. Cheng, N. Ramakrishnan, R. K. Agrawal, P. K. Reddy, J. Srivastava, & T. Chakraborty (Eds.), Advances in Knowledge Discovery and Data Mining - 25th Pacific-Asia Conference, PAKDD 2021, Proceedings (pp. 152-163). (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Vol. 12712 LNAI). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-75762-5_13, https://doi.org/10.1007/978-3-030-75762-5_13
Hanselle J, Tornede A, Wever M, Hüllermeier E. Algorithm Selection as Superset Learning: Constructing Algorithm Selectors from Imprecise Performance Data. In Karlapalem K, Cheng H, Ramakrishnan N, Agrawal RK, Reddy PK, Srivastava J, Chakraborty T, editors, Advances in Knowledge Discovery and Data Mining - 25th Pacific-Asia Conference, PAKDD 2021, Proceedings. Springer Science and Business Media Deutschland GmbH. 2021. p. 152-163. (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)). doi: 10.1007/978-3-030-75762-5_13, 10.1007/978-3-030-75762-5_13
Hanselle, Jonas ; Tornede, Alexander ; Wever, Marcel et al. / Algorithm Selection as Superset Learning : Constructing Algorithm Selectors from Imprecise Performance Data. Advances in Knowledge Discovery and Data Mining - 25th Pacific-Asia Conference, PAKDD 2021, Proceedings. editor / Kamal Karlapalem ; Hong Cheng ; Naren Ramakrishnan ; R. K. Agrawal ; P. Krishna Reddy ; Jaideep Srivastava ; Tanmoy Chakraborty. Springer Science and Business Media Deutschland GmbH, 2021. pp. 152-163 (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)).
Download
@inproceedings{a88fe7d4edb14347b5854b1452eef64c,
title = "Algorithm Selection as Superset Learning: Constructing Algorithm Selectors from Imprecise Performance Data",
abstract = "Algorithm selection refers to the task of automatically selecting the most suitable algorithm for solving an instance of a computational problem from a set of candidate algorithms. Here, suitability is typically measured in terms of the algorithms{\textquoteright} runtimes. To allow the selection of algorithms on new problem instances, machine learning models are trained on previously observed performance data and then used to predict the algorithms{\textquoteright} performances. Due to the computational effort, the execution of such algorithms is often prematurely terminated, which leads to right-censored observations representing a lower bound on the actual runtime. While simply neglecting these censored samples leads to overly optimistic models, imputing them with precise though hypothetical values, such as the commonly used penalized average runtime, is a rather arbitrary and biased approach. In this paper, we propose a simple regression method based on so-called superset learning, in which right-censored runtime data are explicitly incorporated in terms of interval-valued observations, offering an intuitive and efficient approach to handling censored data. Benchmarking on publicly available algorithm performance data, we demonstrate that it outperforms the aforementioned na{\"i}ve ways of dealing with censored samples and is competitive to established methods for censored regression in the field of algorithm selection.",
keywords = "Algorithm selection, Censored data, Superset learning",
author = "Jonas Hanselle and Alexander Tornede and Marcel Wever and Eyke H{\"u}llermeier",
note = "Publisher Copyright: {\textcopyright} 2021, Springer Nature Switzerland AG.; 25th Pacific-Asia Conference on Knowledge Discovery and Data Mining, PAKDD 2021 ; Conference date: 11-05-2021 Through 14-05-2021",
year = "2021",
doi = "10.1007/978-3-030-75762-5_13",
language = "English",
isbn = "9783030757618",
series = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
publisher = "Springer Science and Business Media Deutschland GmbH",
pages = "152--163",
editor = "Kamal Karlapalem and Hong Cheng and Naren Ramakrishnan and Agrawal, {R. K.} and Reddy, {P. Krishna} and Jaideep Srivastava and Tanmoy Chakraborty",
booktitle = "Advances in Knowledge Discovery and Data Mining - 25th Pacific-Asia Conference, PAKDD 2021, Proceedings",
address = "Germany",

}

Download

TY - GEN

T1 - Algorithm Selection as Superset Learning

T2 - 25th Pacific-Asia Conference on Knowledge Discovery and Data Mining, PAKDD 2021

AU - Hanselle, Jonas

AU - Tornede, Alexander

AU - Wever, Marcel

AU - Hüllermeier, Eyke

N1 - Publisher Copyright: © 2021, Springer Nature Switzerland AG.

PY - 2021

Y1 - 2021

N2 - Algorithm selection refers to the task of automatically selecting the most suitable algorithm for solving an instance of a computational problem from a set of candidate algorithms. Here, suitability is typically measured in terms of the algorithms’ runtimes. To allow the selection of algorithms on new problem instances, machine learning models are trained on previously observed performance data and then used to predict the algorithms’ performances. Due to the computational effort, the execution of such algorithms is often prematurely terminated, which leads to right-censored observations representing a lower bound on the actual runtime. While simply neglecting these censored samples leads to overly optimistic models, imputing them with precise though hypothetical values, such as the commonly used penalized average runtime, is a rather arbitrary and biased approach. In this paper, we propose a simple regression method based on so-called superset learning, in which right-censored runtime data are explicitly incorporated in terms of interval-valued observations, offering an intuitive and efficient approach to handling censored data. Benchmarking on publicly available algorithm performance data, we demonstrate that it outperforms the aforementioned naïve ways of dealing with censored samples and is competitive to established methods for censored regression in the field of algorithm selection.

AB - Algorithm selection refers to the task of automatically selecting the most suitable algorithm for solving an instance of a computational problem from a set of candidate algorithms. Here, suitability is typically measured in terms of the algorithms’ runtimes. To allow the selection of algorithms on new problem instances, machine learning models are trained on previously observed performance data and then used to predict the algorithms’ performances. Due to the computational effort, the execution of such algorithms is often prematurely terminated, which leads to right-censored observations representing a lower bound on the actual runtime. While simply neglecting these censored samples leads to overly optimistic models, imputing them with precise though hypothetical values, such as the commonly used penalized average runtime, is a rather arbitrary and biased approach. In this paper, we propose a simple regression method based on so-called superset learning, in which right-censored runtime data are explicitly incorporated in terms of interval-valued observations, offering an intuitive and efficient approach to handling censored data. Benchmarking on publicly available algorithm performance data, we demonstrate that it outperforms the aforementioned naïve ways of dealing with censored samples and is competitive to established methods for censored regression in the field of algorithm selection.

KW - Algorithm selection

KW - Censored data

KW - Superset learning

UR - http://www.scopus.com/inward/record.url?scp=85111141266&partnerID=8YFLogxK

U2 - 10.1007/978-3-030-75762-5_13

DO - 10.1007/978-3-030-75762-5_13

M3 - Conference contribution

SN - 9783030757618

T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)

SP - 152

EP - 163

BT - Advances in Knowledge Discovery and Data Mining - 25th Pacific-Asia Conference, PAKDD 2021, Proceedings

A2 - Karlapalem, Kamal

A2 - Cheng, Hong

A2 - Ramakrishnan, Naren

A2 - Agrawal, R. K.

A2 - Reddy, P. Krishna

A2 - Srivastava, Jaideep

A2 - Chakraborty, Tanmoy

PB - Springer Science and Business Media Deutschland GmbH

Y2 - 11 May 2021 through 14 May 2021

ER -

By the same author(s)