In this letter, we investigate the impact of choosing different loss functions from the viewpoint of statistical learning theory. We introduce a convexity assumption, which is met by all loss functions commonly used in the literature, and study how the bound on the estimation error changes with the loss. We also derive a general result on the minimizer of the expected risk for a convex loss function in the case of classification. The main outcome of our analysis is that for classification, the hinge loss appears to be the loss of choice. Other things being equal, the hinge loss leads to a convergence rate practically indistinguishable from the logistic loss rate and much better than the square loss rate. Furthermore, if the hypothesis space...
Abstract. We consider the problem of binary classification where the classifier can, for a particula...
We consider the problem of binary classification where the classifier can, for a particular cost, ch...
We consider the problem of binary classification where the classifier can, for a particular cost, ch...
In this letter, we investigate the impact of choosing different loss functions from the viewpoint of...
In this paper we investigate the impact of choosing different loss functions from the viewpoint of st...
In this paper we investigate the impact of choosing di\ufb00erent loss functions from the viewpoint ...
In many classification procedures, the classification function is obtained (or trained) by minimizi...
Many of the classification algorithms developed in the machine learning literature, including the s...
Many of the classification algorithms developed in the machine learning literature, including the su...
The logistic loss function is often advocated in machine learning and statistics as a smooth and str...
The combination of using loss functions that are both Bayes consistent and margin enforcing has lead...
Problems of data classification can be studied in the framework of regularization theory as ill-pose...
© 2018 IEEE. A loss function measures the discrepancy between the true values (observations) and the...
We study how closely the optimal Bayes error rate can be approximately reached using a classificatio...
© 2018 IEEE. A loss function measures the discrepancy between the true values (observations) and the...
Abstract. We consider the problem of binary classification where the classifier can, for a particula...
We consider the problem of binary classification where the classifier can, for a particular cost, ch...
We consider the problem of binary classification where the classifier can, for a particular cost, ch...
In this letter, we investigate the impact of choosing different loss functions from the viewpoint of...
In this paper we investigate the impact of choosing different loss functions from the viewpoint of st...
In this paper we investigate the impact of choosing di\ufb00erent loss functions from the viewpoint ...
In many classification procedures, the classification function is obtained (or trained) by minimizi...
Many of the classification algorithms developed in the machine learning literature, including the s...
Many of the classification algorithms developed in the machine learning literature, including the su...
The logistic loss function is often advocated in machine learning and statistics as a smooth and str...
The combination of using loss functions that are both Bayes consistent and margin enforcing has lead...
Problems of data classification can be studied in the framework of regularization theory as ill-pose...
© 2018 IEEE. A loss function measures the discrepancy between the true values (observations) and the...
We study how closely the optimal Bayes error rate can be approximately reached using a classificatio...
© 2018 IEEE. A loss function measures the discrepancy between the true values (observations) and the...
Abstract. We consider the problem of binary classification where the classifier can, for a particula...
We consider the problem of binary classification where the classifier can, for a particular cost, ch...
We consider the problem of binary classification where the classifier can, for a particular cost, ch...