License: Creative Commons Attribution 4.0 International license (CC BY 4.0)
When quoting this document, please refer to the following
DOI: 10.4230/LIPIcs.CP.2021.43
URN: urn:nbn:de:0030-drops-153346
URL: http://dagstuhl.sunsite.rwth-aachen.de/volltexte/2021/15334/
Matricon, Théo ;
Anastacio, Marie ;
Fijalkow, Nathanaël ;
Simon, Laurent ;
Hoos, Holger H.
Statistical Comparison of Algorithm Performance Through Instance Selection
Abstract
Empirical performance evaluations, in competitions and scientific publications, play a major role in improving the state of the art in solving many automated reasoning problems, including SAT, CSP and Bayesian network structure learning (BNSL). To empirically demonstrate the merit of a new solver usually requires extensive experiments, with computational costs of CPU years. This not only makes it difficult for researchers with limited access to computational resources to test their ideas and publish their work, but also consumes large amounts of energy. We propose an approach for comparing the performance of two algorithms: by performing runs on carefully chosen instances, we obtain a probabilistic statement on which algorithm performs best, trading off between the computational cost of running algorithms and the confidence in the result. We describe a set of methods for this purpose and evaluate their efficacy on diverse datasets from SAT, CSP and BNSL. On all these datasets, most of our approaches were able to choose the correct algorithm with about 95% accuracy, while using less than a third of the CPU time required for a full comparison; the best methods reach this level of accuracy within less than 15% of the CPU time for a full comparison.
BibTeX - Entry
@InProceedings{matricon_et_al:LIPIcs.CP.2021.43,
author = {Matricon, Th\'{e}o and Anastacio, Marie and Fijalkow, Nathana\"{e}l and Simon, Laurent and Hoos, Holger H.},
title = {{Statistical Comparison of Algorithm Performance Through Instance Selection}},
booktitle = {27th International Conference on Principles and Practice of Constraint Programming (CP 2021)},
pages = {43:1--43:21},
series = {Leibniz International Proceedings in Informatics (LIPIcs)},
ISBN = {978-3-95977-211-2},
ISSN = {1868-8969},
year = {2021},
volume = {210},
editor = {Michel, Laurent D.},
publisher = {Schloss Dagstuhl -- Leibniz-Zentrum f{\"u}r Informatik},
address = {Dagstuhl, Germany},
URL = {https://drops.dagstuhl.de/opus/volltexte/2021/15334},
URN = {urn:nbn:de:0030-drops-153346},
doi = {10.4230/LIPIcs.CP.2021.43},
annote = {Keywords: Performance assessment, early stopping, automated reasoning solvers}
}