Stable likelihood computation for machine learning of linear differential operators with Gaussian processes

Omid Chatrabgoun, Mohsen Esmaeilbeigi, Maysam Cheraghi, Alireza Daneshkhah

Research output: Contribution to journalArticlepeer-review

6 Citations (Scopus)
78 Downloads (Pure)

Abstract

In many applied sciences, the main aim is to learn the parameters in the operational equations which best fit the observed data. A framework for solving such problems is to employ Gaussian process (GP) emulators which are well-known as nonparametric Bayesian machine learning techniques. GPs are among a class of methods known as kernel machines which can be used to approximate rather complex problems by tuning their hyperparameters. The maximum likelihood estimation (MLE) has widely been used to estimate the parameters of the operators and kernels. However, the MLE-based and Bayesian inference in the standard form are usually involved in setting up a covariance matrix which is generally ill-conditioned. As a result, constructing and inverting the covariance matrix using the standard form will become unstable to learn the parameters in the operational equations. In this paper, we propose a novel approach to tackle these computational complexities and also resolve the ill-conditioning problem by forming the covariance matrix using alternative bases via the Hilbert−Schmidt SVD (HS-SVD) approach. Applying this approach yields a novel matrix factorization of the block-structured covariance matrix which can be implemented stably by isolating the main source of the ill-conditioning. In contrast to standard matrix decompositions which start with a matrix and produce the resulting factors, the HS-SVD is constructed from the Hilbert−Schmidt eigenvalues and eigenvectors without the need to ever form the potentially ill-conditioned matrix. We also provide stable MLE and Bayesian inference to adaptively estimate hyperparameters, and the corresponding operators can then be efficiently predicted at some new points using the proposed HS-SVD bases. The efficiency and stability of the proposed HS-SVD method will be compared with the existing methods by several illustrations of the parametric linear equations, such as ordinary and partial differential equations, and integro-differential and fractional order operators.
Original languageEnglish
Pages (from-to)75-99
Number of pages25
JournalInternational Journal for Uncertainty Quantification
Volume12
Issue number3
DOIs
Publication statusPublished - 20 Apr 2022

Bibliographical note

Copyright © and Moral Rights are retained by the author(s) and/ or other copyright owners. A copy can be downloaded for personal non-commercial research or study, without prior permission or charge. This item cannot be reproduced or quoted extensively from without first obtaining permission in writing from the copyright holder(s). The content must not be changed in any way or sold commercially in any format or medium without the formal permission of the copyright holders.

This document is the author’s post-print version, incorporating any revisions agreed during the peer-review process. Some differences between the published version and this version may remain and you are advised to consult the published version if you wish to cite from it.

Keywords

  • Gaussian processes
  • Hilbert–Schmidt’s theory
  • probabilistic machine learning
  • stable computation
  • uncertainty quantification

ASJC Scopus subject areas

  • Statistics and Probability
  • Modelling and Simulation
  • Discrete Mathematics and Combinatorics
  • Control and Optimization

Fingerprint

Dive into the research topics of 'Stable likelihood computation for machine learning of linear differential operators with Gaussian processes'. Together they form a unique fingerprint.

Cite this