Manuela Bastidas

Assistant Professor


Curriculum vitae



Department of mathematics

Universidad Nacional de Colombia, Medellín

Medellín, Colombia



Optimizing Variational Physics-Informed Neural Networks Using Least Squares


Journal article


C. Uriarte, M. Bastidas, D. Pardo, J. M. Taylor, S. Rojas
Computers and Mathematics with Applications, 2024

Semantic Scholar ArXiv DBLP DOI
Cite

Cite

APA   Click to copy
Uriarte, C., Bastidas, M., Pardo, D., Taylor, J. M., & Rojas, S. (2024). Optimizing Variational Physics-Informed Neural Networks Using Least Squares. Computers and Mathematics with Applications.


Chicago/Turabian   Click to copy
Uriarte, C., M. Bastidas, D. Pardo, J. M. Taylor, and S. Rojas. “Optimizing Variational Physics-Informed Neural Networks Using Least Squares.” Computers and Mathematics with Applications (2024).


MLA   Click to copy
Uriarte, C., et al. “Optimizing Variational Physics-Informed Neural Networks Using Least Squares.” Computers and Mathematics with Applications, 2024.


BibTeX   Click to copy

@article{c2024a,
  title = {Optimizing Variational Physics-Informed Neural Networks Using Least Squares},
  year = {2024},
  journal = {Computers and Mathematics with Applications},
  author = {Uriarte, C. and Bastidas, M. and Pardo, D. and Taylor, J. M. and Rojas, S.}
}

Abstract

Variational Physics-Informed Neural Networks often suffer from poor convergence when using stochastic gradient-descent-based optimizers. By introducing a Least Squares solver for the weights of the last layer of the neural network, we improve the convergence of the loss during training in most practical scenarios. This work analyzes the computational cost of the resulting hybrid Least-Squares/Gradient-Descent optimizer and explains how to implement it efficiently. In particular, we show that a traditional implementation based on backward-mode automatic differentiation leads to a prohibitively expensive algorithm. To remedy this, we propose using either forward-mode automatic differentiation or an ultraweak-type scheme that avoids the differentiation of trial functions in the discrete weak formulation. The proposed alternatives are up to one hundred times faster than the traditional one, recovering a computational cost-per-iteration similar to that of a conventional gradient-descent-based optimizer alone. To support our analysis, we derive computational estimates and conduct numerical experiments in one- and two-dimensional problems.


Share

Tools
Translate to