Learning Monocular Depth in Dynamic Scenes via Instance-Aware Projection Consistency

Cited 34 time in webofscience Cited 0 time in scopus
  • Hit : 75
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorLee, Seokjuko
dc.contributor.authorIm, Sunghoonko
dc.contributor.authorLin, Stephenko
dc.contributor.authorKweon, In Soko
dc.date.accessioned2021-10-27T08:50:21Z-
dc.date.available2021-10-27T08:50:21Z-
dc.date.created2021-10-27-
dc.date.created2021-10-27-
dc.date.issued2021-02-
dc.identifier.citation35th AAAI Conference on Artificial Intelligence / 33rd Conference on Innovative Applications of Artificial Intelligence / 11th Symposium on Educational Advances in Artificial Intelligence, pp.1863 - 1872-
dc.identifier.issn2159-5399-
dc.identifier.urihttp://hdl.handle.net/10203/288362-
dc.description.abstractWe present an end-to-end joint training framework that explicitly models 6-DoF motion of multiple dynamic objects, ego-motion, and depth in a monocular camera setup without supervision. Our technical contributions are three-fold. First, we highlight the fundamental difference between inverse and forward projection while modeling the individual motion of each rigid object, and propose a geometrically correct projection pipeline using a neural forward projection module. Second, we design a unified instance-aware photometric and geometric consistency loss that holistically imposes self-supervisory signals for every background and object region. Lastly, we introduce a general-purpose auto-annotation scheme using any off-the-shelf instance segmentation and optical flow models to produce video instance segmentation maps that will be utilized as input to our training pipeline. These proposed elements are validated in a detailed ablation study. Through extensive experiments conducted on the KITTI and Cityscapes dataset, our framework is shown to outperform the state-of-the-art depth and motion estimation methods. Our code, dataset, and models are publicly available.-
dc.languageEnglish-
dc.publisherASSOC ADVANCEMENT ARTIFICIAL INTELLIGENCE-
dc.titleLearning Monocular Depth in Dynamic Scenes via Instance-Aware Projection Consistency-
dc.typeConference-
dc.identifier.wosid000680423501106-
dc.identifier.scopusid2-s2.0-85130038874-
dc.type.rimsCONF-
dc.citation.beginningpage1863-
dc.citation.endingpage1872-
dc.citation.publicationname35th AAAI Conference on Artificial Intelligence / 33rd Conference on Innovative Applications of Artificial Intelligence / 11th Symposium on Educational Advances in Artificial Intelligence-
dc.identifier.conferencecountryUS-
dc.identifier.conferencelocationVirtual-
dc.contributor.localauthorKweon, In So-
dc.contributor.nonIdAuthorLee, Seokju-
dc.contributor.nonIdAuthorIm, Sunghoon-
dc.contributor.nonIdAuthorLin, Stephen-
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 34 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0