Authors

Stefano Massaroli, Michael Poli, Federico Califano, Angela Faragasso, Jinkyoo Park, Atsushi Yamashita, Hajime Asama

Abstract

Neural networks are discrete entities: subdivided into discrete layers and parametrized by weights which are iteratively optimized via difference equations. Recent work proposes networks with layer outputs which are no longer quantized but are solutions of an ordinary differential equation (ODE); however, these networks are still optimized via discrete methods (e.g. gradient descent). In this paper, we explore a different direction: namely, we propose a novel framework for learning in which the parameters themselves are solutions of ODEs. By viewing the optimization process as the evolution of a port-Hamiltonian system, we can ensure convergence to a minimum of the objective function. Numerical experiments have been performed to show the validity and effectiveness of the proposed methods.

Citation

  • Journal: 2019 IEEE 58th Conference on Decision and Control (CDC)
  • Year: 2019
  • Volume:
  • Issue:
  • Pages: 6799–6806
  • Publisher: IEEE
  • DOI: 10.1109/cdc40024.2019.9030017

BibTeX

@inproceedings{Massaroli_2019,
  title={{Port–Hamiltonian Approach to Neural Network Training}},
  DOI={10.1109/cdc40024.2019.9030017},
  booktitle={{2019 IEEE 58th Conference on Decision and Control (CDC)}},
  publisher={IEEE},
  author={Massaroli, Stefano and Poli, Michael and Califano, Federico and Faragasso, Angela and Park, Jinkyoo and Yamashita, Atsushi and Asama, Hajime},
  year={2019},
  pages={6799--6806}
}

Download the bib file

References