Ambiguity-aware robust teacher (ART): Enhanced self-knowledge distillation framework with pruned teacher network

Cited 7 time in webofscience Cited 0 time in scopus
  • Hit : 236
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorCho, Yucheolko
dc.contributor.authorHam, Gyeongdoko
dc.contributor.authorLee, Jae-Hyeokko
dc.contributor.authorKim, Daeshikko
dc.date.accessioned2023-04-24T07:01:15Z-
dc.date.available2023-04-24T07:01:15Z-
dc.date.created2023-04-24-
dc.date.created2023-04-24-
dc.date.issued2023-03-
dc.identifier.citationPATTERN RECOGNITION, v.140-
dc.identifier.issn0031-3203-
dc.identifier.urihttp://hdl.handle.net/10203/306376-
dc.description.abstractSelf-knowledge distillation (self-KD) methods, which use a student model itself as the teacher model instead of a large and complex teacher model, are currently a subject of active study. Since most previ-ous self-KD approaches relied on the knowledge of a single teacher model, if the teacher model incor-rectly predicted confusing samples, poor-quality knowledge was transferred to the student model. Unfor-tunately, natural images are often ambiguous for teacher models due to multiple objects, mislabeling, or low quality. In this paper, we propose a novel knowledge distillation framework named ambiguity-aware robust teacher knowledge distillation (ART-KD) that provides refined knowledge, that reflects the ambigu-ity of the samples with network pruning. Since the pruned teacher model is simply obtained by copying and pruning the teacher model, re-training process is unnecessary in ART-KD. The key insight of ART-KD lies in the predictions of a teacher model and pruned teacher model for ambiguous samples providing different distributions with low similarity. From these two distributions, we obtain a joint distribution considering the ambiguity of the samples as teacher's knowledge for distillation. We comprehensively evaluate our method on public classification benchmarks, as well as more challenging benchmarks for fine-grained visual recognition (FGVR), achieving much superior performance to state-of-the-art counter-parts.(c) 2023 Elsevier Ltd. All rights reserved.-
dc.languageEnglish-
dc.publisherELSEVIER SCI LTD-
dc.titleAmbiguity-aware robust teacher (ART): Enhanced self-knowledge distillation framework with pruned teacher network-
dc.typeArticle-
dc.identifier.wosid000966289200001-
dc.identifier.scopusid2-s2.0-85150935746-
dc.type.rimsART-
dc.citation.volume140-
dc.citation.publicationnamePATTERN RECOGNITION-
dc.identifier.doi10.1016/j.patcog.2023.109541-
dc.contributor.localauthorKim, Daeshik-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorKnowledge distillation-
dc.subject.keywordAuthorSelf-knowledge distillation-
dc.subject.keywordAuthorNetwork pruning-
dc.subject.keywordAuthorTeacher -student model-
dc.subject.keywordAuthorLong -tail samples-
dc.subject.keywordAuthorAmbiguous samples-
dc.subject.keywordAuthorSample ambiguity-
dc.subject.keywordAuthorData augmentation-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 7 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0