Recent work has shown potential in using Mixed Integer Programming (MIP) solvers to optimize certain aspects of neural networks (NN). How- ever little research has gone into training NNs with solvers. State of the art methods to train NNs are typically gradient-based and require sig- nificant data, computation on GPUs and extensive hyper-parameter tuning. In contrast, training with MIP solvers should not require GPUs or hyper- parameter tuning but can likely not handle large amounts of data. Thus works builds on recent ad- vances that train binarized NNs using MIP solvers. We go beyond current work by formulating new MIP models to increase the amount of data that can be used and to train non-binary integer-valued net- works. Our results sho...
Recently, ReLU neural networks have been modelled as constraints in mixed integer linear programming...
As Machine Learning applications increase the demand for optimised implementations in both embedded ...
Due to limited size, cost and power, embedded devices do not offer the same computational throughput...
Recent work has shown potential in using Mixed Integer Programming (MIP) solvers to optimize certain...
Artificial Neural Networks (ANNs) are prevalent machine learning models that are applied across vari...
Mixed-integer programming (MIP) technology offers a generic way of formulating and solving combinato...
A recent Graph Neural Network (GNN) approach for learning to branch has been shown to successfully r...
The literature has shown how to optimize and analyze the parameters of different types of neural net...
Large Neighborhood Search (LNS) is a combinatorial optimization heuristic that starts with an assign...
In this work we present neural network train-ing algorithms, which are based on the differ-ential ev...
In this paper, low end Digital Signal Processors (DSPs) are applied to accelerate integer neural net...
The application of programmable devices to implement neural networks requires efficient training alg...
We consider the algorithmic problem of finding the optimal weights and biases for a two-layer fully ...
Abstract We present strong mixed-integer programming (MIP) formulations for high-dimensional piecew...
Mixed Integer Programming (MIP) is one of the most widely used modeling techniques for combinatorial...
Recently, ReLU neural networks have been modelled as constraints in mixed integer linear programming...
As Machine Learning applications increase the demand for optimised implementations in both embedded ...
Due to limited size, cost and power, embedded devices do not offer the same computational throughput...
Recent work has shown potential in using Mixed Integer Programming (MIP) solvers to optimize certain...
Artificial Neural Networks (ANNs) are prevalent machine learning models that are applied across vari...
Mixed-integer programming (MIP) technology offers a generic way of formulating and solving combinato...
A recent Graph Neural Network (GNN) approach for learning to branch has been shown to successfully r...
The literature has shown how to optimize and analyze the parameters of different types of neural net...
Large Neighborhood Search (LNS) is a combinatorial optimization heuristic that starts with an assign...
In this work we present neural network train-ing algorithms, which are based on the differ-ential ev...
In this paper, low end Digital Signal Processors (DSPs) are applied to accelerate integer neural net...
The application of programmable devices to implement neural networks requires efficient training alg...
We consider the algorithmic problem of finding the optimal weights and biases for a two-layer fully ...
Abstract We present strong mixed-integer programming (MIP) formulations for high-dimensional piecew...
Mixed Integer Programming (MIP) is one of the most widely used modeling techniques for combinatorial...
Recently, ReLU neural networks have been modelled as constraints in mixed integer linear programming...
As Machine Learning applications increase the demand for optimised implementations in both embedded ...
Due to limited size, cost and power, embedded devices do not offer the same computational throughput...