# Low-rank tensor decompositions for sampling of high-dimensional probability distributions.

Sergey Dolgov

## Affiliation:

University of Bath, UK

## Date:

Tue, 25/06/2019 - 11:05am

## Venue:

RC-4082, The Red Centre, UNSW

## Abstract:

Uncertainty quantification and inverse problems in many variables are pressingly needed tasks, yet high-dimensional functions are notoriously difficult to integrate in order to compute desired quantities of interest. Functional approximations, in particular the low-rank separation of variables into tensor product decompositions, have become popular for reducing the computational cost of high-dimensional integration down to linear scaling in the number of variables. However, tensor approximations may be inefficient for non-smooth functions. Sampling based Monte Carlo methods are more general, but they may exhibit a very slow convergence, overlooking a hidden structure of the function.

In this talk we review tensor product approximations for the problem of uncertainty quantification and Bayesian inference. This allows efficient integration of smooth PDE solutions, posterior density functions and quantities of interest. Moreover, we can use the low-rank approximation of the density function to construct efficient proposals in the MCMC algorithm, as well as in the importance weighting.

This combined tensor approximation - MCMC method is more accurate also if the quantity of interest is not smooth, such as the indicator function of an event.