The problem of selecting the best classification algorithm for a specific problem continues to be very relevant, especially since the number of classification algorithms keeps growing significantly. Testing all alternatives is not really a viable option: if we compare all pairs of algorithms, as is often advocated, the number of comparisons grows exponentially. To avoid this problem we suggest a method referred to as active testing, whose aim is to reduce the number of comparisons by carefully selecting which tests should be carried out. This method uses meta-knowledge concerning past experiments and proceeds in an iterative manner. It takes the form of a competition in which, in each iteration, the candidate best algorithm is pitted agains...