Abstract
We present CRefNet, a hybrid transformer-convolutional deep neural network for consistent reflectance estimation in intrinsic image decomposition. Estimating consistent reflectance is particularly challenging when the same material appears differently due to changes in illumination. Our method achieves enhanced global reflectance consistency via a novel transformer module that converts image features to reflectance features. At the same time, this module also exploits long-range data interactions. We introduce reflectance reconstruction as a novel auxiliary task that shares a common decoder with the reflectance estimation task, and which substantially improves the quality of reconstructed reflectance maps. Finally, we improve local reflectance consistency via a new rectified gradient filter that effectively suppresses small variations in predictions without any overhead at inference time. Our experiments show that our contributions enable CRefNet to predict highly consistent reflectance maps and to outperform the state of the art by 10% WHDR.
Original language | English |
---|---|
Number of pages | 14 |
Journal | IEEE Transactions on Visualization and Computer Graphics |
DOIs | |
Publication status | Published - 1 Dec 2023 |
Bibliographical note
This work was supported by the China Scholarship Council, an EPSRC-UKRI Innovation Fellowship (EP/S001050/1) andEPSRC grant CAMERA (EP/M023281/1, EP/T022523/1).