Linear classification has achieved complexity linear to the data size. However, in many applications, large-scale data contains only a few samples that can improve the target objective. In this paper, we propose a sublinear-time algorithm that uses Nearest-Neighbor-based Coordinate Descent method to solve Linear SVM with truncated loss. In particular, we propose a sequential relaxation that solves prob-lem with general truncated-loss by a series of sparse convex programs. Then we solve each program using indexed dual coordinate descent, which avoids I/O of unnecessary data by searching coordinates of large gradient. We show the pro-posed algorithm has linear convergence rate, and has sublinear complexity for each iteration. We also demonstr...
Large-scale `1-regularized loss minimization problems arise in high-dimensional applications such as...
We consider convex-concave saddle point problems with a separable structure and non-strongly convex ...
Over recent years we have seen the appearance of huge datasets that do not fit into memory and do no...
Linear Classification has achieved complexity linear to the data size. However, in many applications...
Over the past decades, Linear Programming (LP) has been widely used in different areas and considere...
In many applications, data appear with a huge number of instances as well as features. Linear Suppor...
This manuscript describes a method for training linear SVMs (including binary SVMs, SVM regression, ...
In many machine learning problems such as the dual form of SVM, the objective function to be minimiz...
© 2016 IEEE. We present a sublinear version of the dual coordinate ascent method for solving a group...
© 2015, Springer Science+Business Media New York. In order to control the effects of outliers in tra...
For classification problems with millions of training examples or dimensions, accuracy, training and...
For classification problems with millions of training examples or dimensions, accuracy, training and...
Large-scale optimization problems appear quite frequently in data science and machine learning appli...
The main contribution of this dissertation is the development of a method to train a Support Vector ...
The main contribution of this dissertation is the development of a method to train a Support Vector ...
Large-scale `1-regularized loss minimization problems arise in high-dimensional applications such as...
We consider convex-concave saddle point problems with a separable structure and non-strongly convex ...
Over recent years we have seen the appearance of huge datasets that do not fit into memory and do no...
Linear Classification has achieved complexity linear to the data size. However, in many applications...
Over the past decades, Linear Programming (LP) has been widely used in different areas and considere...
In many applications, data appear with a huge number of instances as well as features. Linear Suppor...
This manuscript describes a method for training linear SVMs (including binary SVMs, SVM regression, ...
In many machine learning problems such as the dual form of SVM, the objective function to be minimiz...
© 2016 IEEE. We present a sublinear version of the dual coordinate ascent method for solving a group...
© 2015, Springer Science+Business Media New York. In order to control the effects of outliers in tra...
For classification problems with millions of training examples or dimensions, accuracy, training and...
For classification problems with millions of training examples or dimensions, accuracy, training and...
Large-scale optimization problems appear quite frequently in data science and machine learning appli...
The main contribution of this dissertation is the development of a method to train a Support Vector ...
The main contribution of this dissertation is the development of a method to train a Support Vector ...
Large-scale `1-regularized loss minimization problems arise in high-dimensional applications such as...
We consider convex-concave saddle point problems with a separable structure and non-strongly convex ...
Over recent years we have seen the appearance of huge datasets that do not fit into memory and do no...