Projects per year
Abstract
Low rank tensor approximations have been employed successfully, for example, to build surrogate models that can be used to speed up large-scale inference problems in high dimensions. The success of this depends critically on the rank that is necessary to represent or approximate the underlying distribution. In this paper, we develop a-priori rank bounds for approximations in the functional Tensor-Train representation for the case of a Gaussian (normally distributed) model. We show that under suitable conditions on the precision matrix, we can represent the Gaussian density to high accuracy without suffering from an exponential growth of complexity as the dimension increases. Our results provide evidence of the suitability and limitations of low rank tensor methods in a simple but important model case. Numerical experiments confirm that the rank bounds capture the qualitative behavior of the rank structure when varying the parameters of the precision matrix and the accuracy of the approximation.
Original language | English |
---|---|
Pages (from-to) | 1191-1224 |
Number of pages | 34 |
Journal | SIAM/ASA Journal on Uncertainty Quantification |
Volume | 10 |
Issue number | 3 |
Early online date | 28 Sep 2022 |
DOIs | |
Publication status | Published - 30 Sep 2022 |
Keywords
- math.NA
- cs.NA
- math.ST
- stat.TH
- 15A23, 15A69, 65C60, 65D32, 65D15, 41A10
Fingerprint
Dive into the research topics of 'Rank Bounds for Approximating Gaussian Densities in the Tensor-Train Format'. Together they form a unique fingerprint.Projects
- 1 Active
-
Tensor decomposition sampling algorithms for Bayesian inverse problems
Engineering and Physical Sciences Research Council
1/03/21 → 28/02/24
Project: Research council