Semantic-guided de-attention with sharpened triplet marginal loss for visual place recognition

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 471
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorChoi, Seung-Minko
dc.contributor.authorLee, Seung-Ikko
dc.contributor.authorLee, Jae-Yeongko
dc.contributor.authorKweon, In Soko
dc.date.accessioned2023-06-21T07:00:12Z-
dc.date.available2023-06-21T07:00:12Z-
dc.date.created2023-06-21-
dc.date.issued2023-09-
dc.identifier.citationPATTERN RECOGNITION, v.141-
dc.identifier.issn0031-3203-
dc.identifier.urihttp://hdl.handle.net/10203/307421-
dc.description.abstractThanks to Earth-level Street View images from Google Maps, a visual image geo-localization can estimate the coarse location of a query image with a visual place recognition process. However, this can get very challenging when non-static objects change with time, severely degrading image retrieval accuracy. We address the problem of city-scale visual place recognition in complex urban environments crowded with non-static clutters. To this end, we first analyze what clutters degrade similarity matching between the query and database images. Second, we design a self-supervised trainable de-attention module that pre-vents the network from focusing on non-static objects in an input image. In addition, we propose a novel triplet marginal loss called sharpened triplet marginal loss to make feature descriptors more discriminative. Lastly, due to the lack of geo-tagged public datasets with a high density of non-static objects, we propose a clutter augmentation method to evaluate our approach. The experimental results show that our model has notably improved over the existing attention methods in geo-localization tasks on the public bench-mark datasets and on their augmented versions with high population and traffic. Our code is available at https://github.com/ccsmm78/deattention _ with _ stml _ for _ vpr .(c) 2023 The Author(s). Published by Elsevier Ltd.This is an open access article under the CC BY-NC-ND license ( http://creativecommons.org/licenses/by-nc-nd/4.0/ )-
dc.languageEnglish-
dc.publisherELSEVIER SCI LTD-
dc.titleSemantic-guided de-attention with sharpened triplet marginal loss for visual place recognition-
dc.typeArticle-
dc.identifier.wosid000999041000001-
dc.identifier.scopusid2-s2.0-85156187545-
dc.type.rimsART-
dc.citation.volume141-
dc.citation.publicationnamePATTERN RECOGNITION-
dc.identifier.doi10.1016/j.patcog.2023.109645-
dc.contributor.localauthorKweon, In So-
dc.contributor.nonIdAuthorLee, Seung-Ik-
dc.contributor.nonIdAuthorLee, Jae-Yeong-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorVisual place recognition-
dc.subject.keywordAuthorImage retrieval-
dc.subject.keywordAuthorTriplet marginal loss-
dc.subject.keywordAuthorAttention-
dc.subject.keywordAuthorDe-attention-
dc.subject.keywordAuthorSemantic guidance-
dc.subject.keywordAuthorSemantic segmentation-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0