Abstract
In this work we investigate the practicality of stochastic gradient descent and its variants with variance-reduction techniques in imaging inverse problems. Such algorithms have been shown in the large-scale optimization and machine learning literature to have optimal complexity in theory, and to provide great improvement empirically over the deterministic gradient methods. However, in some tasks such as image deblurring, many of such methods fail to converge faster than the deterministic gradient methods, even in terms of epoch counts. We investigate this phenomenon and propose a theory-inspired mechanism for the practitioners to efficiently characterize whether it is beneficial for an inverse problem to be solved by stochastic optimization techniques or not. Using standard tools in numerical linear algebra, we derive conditions on the spectral structure of the inverse problem for being a suitable application of stochastic gradient methods. Particularly, if the Hessian matrix of an imaging inverse problem has a fast-decaying eigenspectrum, then our theory suggests that the stochastic gradient methods can be more advantageous than deterministic methods for solving such a problem. Our results also provide guidance on choosing appropriately the partition minibatch schemes, showing that a good minibatch scheme typically has relatively low correlation within each of the minibatches. Finally, we present numerical studies which validate our results.
Original language | English |
---|---|
Article number | 9234749 |
Pages (from-to) | 1471-1485 |
Number of pages | 15 |
Journal | IEEE Transactions on Computational Imaging |
Volume | 6 |
DOIs | |
Publication status | Published - 21 Oct 2020 |
Keywords
- Imaging inverse problems
- large-scale optimization
- stochastic optimization
ASJC Scopus subject areas
- Signal Processing
- Computer Science Applications
- Computational Mathematics