This paper has been motivated by two observations. First, empirical comparison of algorithms is often carried out in an ad hoc manner. Second, performance data is abundantly generated, yet often not efficiently used. This second observation is particularly valid in the presence of evolutionary computing and other metaheuristic techniques. Inspired by these observations, a formal model for performance is introduced wherein the space of possible performances is modelled as a total order. On top of the total order, a quantification of the difference between performances is defined. The model is illustrated by formally defining the "penalised runtime" criterion for data from the 2014 SAT competition. Finally, the idea of defining questions in ...
Generalization performance comparison between the proposed model and the stat-of-the-art rival in th...
Introduction The conclusions drawn from a benchmark study of computer performance depend not only o...
International audienceBenchmarking aims to investigate the performance of one or several algorithms ...
This paper has been motivated by two observations. First, empirical comparison of algorithms is ofte...
International audienceEmpirical performance evaluations, in competitions and scientific publications...
We investigate the applicability of an existing framework for algorithm runtime prediction to the fi...
We investigate the applicability of an existing framework for algorithm runtime prediction to the fi...
Benchmark experiments nowadays are the method of choice to evaluate learn-ing algorithms in most res...
Comparison of performance obtained by our approach with other state-of-the-art algorithms.</p
Performance comparisons are ubiquitous in computer science. The proceedings of most conferences are ...
Traditionally, the performance of algorithms is evaluated using worst-case analysis. For a number of...
© Springer International Publishing AG 2018. This chapter presents the evaluation of the proposed al...
One of the primary methods employed by researchers to judge the merits of new heuristics and algorit...
In the field of cognitive science, the primary means of judging a model’s viability is made on the b...
The lack of a useful and accurate software infrastructure for measuring, modeling, and analyzing the...
Generalization performance comparison between the proposed model and the stat-of-the-art rival in th...
Introduction The conclusions drawn from a benchmark study of computer performance depend not only o...
International audienceBenchmarking aims to investigate the performance of one or several algorithms ...
This paper has been motivated by two observations. First, empirical comparison of algorithms is ofte...
International audienceEmpirical performance evaluations, in competitions and scientific publications...
We investigate the applicability of an existing framework for algorithm runtime prediction to the fi...
We investigate the applicability of an existing framework for algorithm runtime prediction to the fi...
Benchmark experiments nowadays are the method of choice to evaluate learn-ing algorithms in most res...
Comparison of performance obtained by our approach with other state-of-the-art algorithms.</p
Performance comparisons are ubiquitous in computer science. The proceedings of most conferences are ...
Traditionally, the performance of algorithms is evaluated using worst-case analysis. For a number of...
© Springer International Publishing AG 2018. This chapter presents the evaluation of the proposed al...
One of the primary methods employed by researchers to judge the merits of new heuristics and algorit...
In the field of cognitive science, the primary means of judging a model’s viability is made on the b...
The lack of a useful and accurate software infrastructure for measuring, modeling, and analyzing the...
Generalization performance comparison between the proposed model and the stat-of-the-art rival in th...
Introduction The conclusions drawn from a benchmark study of computer performance depend not only o...
International audienceBenchmarking aims to investigate the performance of one or several algorithms ...