Multi-resolution distillation for self-supervised monocular depth estimation

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 79
  • Download : 0
Obtaining dense depth ground-truth is not trivial, which leads to the introduction of self-supervised monocular depth estimation. Most self-supervised methods utilize the photometric loss as the primary supervisory signal to optimize a depth network. However, such self-supervised training often falls into an undesirable local minimum due to the ambiguity of the photometric loss. In this paper, we propose a novel self-distillation training scheme that provides a new self-supervision signal, depth consistency among different input resolutions, to the depth network. We further introduce a gradient masking strategy that adjusts the self-supervision signal of the depth consistency during back-propagation to boost the effectiveness of our depth consistency. Experiments demonstrate that our method brings meaningful performance improvements when applied to various depth network architectures. Furthermore, our method outperforms the existing self-supervised methods on KITTI, Cityscapes, and DrivingStereo datasets by a noteworthy margin. © 2023 Elsevier B.V.
Publisher
ELSEVIER
Issue Date
2023-12
Language
English
Article Type
Article
Citation

PATTERN RECOGNITION LETTERS, v.176, pp.215 - 222

ISSN
0167-8655
DOI
10.1016/j.patrec.2023.11.001
URI
http://hdl.handle.net/10203/316116
Appears in Collection
CS-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0