On Training Neural Networks with Mixed Integer Programming

T. Þorbjarnarson, N. Yorke-Smith

Research output: Contribution to conferencePaperpeer-review

85 Downloads (Pure)

Abstract

Recent work has shown potential in using Mixed Integer Programming (MIP) solvers to optimize certain aspects of neural networks (NN). How- ever little research has gone into training NNs with solvers. State of the art methods to train NNs are typically gradient-based and require sig- nificant data, computation on GPUs and extensive hyper-parameter tuning. In contrast, training with MIP solvers should not require GPUs or hyper- parameter tuning but can likely not handle large amounts of data. Thus works builds on recent ad- vances that train binarized NNs using MIP solvers. We go beyond current work by formulating new MIP models to increase the amount of data that can be used and to train non-binary integer-valued net- works. Our results show that comparable results to using gradient descent can be achieved when mini- mal data is available.
Original languageEnglish
Number of pages6
Publication statusPublished - 2021
EventIJCAI-PRICAI'20 Workshop on Data Science Meets Optimisation - Yokohama, Japan
Duration: 7 Jan 20218 Jan 2021

Workshop

WorkshopIJCAI-PRICAI'20 Workshop on Data Science Meets Optimisation
Country/TerritoryJapan
City Yokohama
Period7/01/218/01/21

Fingerprint

Dive into the research topics of 'On Training Neural Networks with Mixed Integer Programming'. Together they form a unique fingerprint.

Cite this