Multi-task vision transformer using low-level chest X-ray feature corpus for COVID-19 diagnosis and severity quantification

Cited 49 time in webofscience Cited 0 time in scopus
  • Hit : 403
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorPark, Sangjoonko
dc.contributor.authorKim, Gwanghyunko
dc.contributor.authorOh, Yujinko
dc.contributor.authorSeo, Joon Beomko
dc.contributor.authorLee, Sang Minko
dc.contributor.authorKim, Jin Hwanko
dc.contributor.authorMoon, Sungjunko
dc.contributor.authorLim, Jae-Kwangko
dc.contributor.authorYe, Jong Chulko
dc.date.accessioned2022-02-08T06:44:13Z-
dc.date.available2022-02-08T06:44:13Z-
dc.date.created2022-02-08-
dc.date.created2022-02-08-
dc.date.created2022-02-08-
dc.date.created2022-02-08-
dc.date.created2022-02-08-
dc.date.issued2022-01-
dc.identifier.citationMEDICAL IMAGE ANALYSIS, v.75-
dc.identifier.issn1361-8415-
dc.identifier.urihttp://hdl.handle.net/10203/292128-
dc.description.abstractDeveloping a robust algorithm to diagnose and quantify the severity of the novel coronavirus disease 2019 (COVID-19) using Chest X-ray (CXR) requires a large number of well-curated COVID-19 datasets, which is difficult to collect under the global COVID-19 pandemic. On the other hand, CXR data with other findings are abundant. This situation is ideally suited for the Vision Transformer (ViT) architecture, where a lot of unlabeled data can be used through structural modeling by the self-attention mechanism. However, the use of existing ViT may not be optimal, as the feature embedding by direct patch flattening or ResNet backbone in the standard ViT is not intended for CXR. To address this problem, here we propose a novel Multi-task ViT that leverages low-level CXR feature corpus obtained from a backbone network that extracts common CXR findings. Specifically, the backbone network is first trained with large public datasets to detect common abnormal findings such as consolidation, opacity, edema, etc. Then, the embedded features from the backbone network are used as corpora for a versatile Transformer model for both the diagnosis and the severity quantification of COVID-19. We evaluate our model on various external test datasets from totally different institutions to evaluate the generalization capability. The experimental results confirm that our model can achieve state-of-the-art performance in both diagnosis and severity quantification tasks with outstanding generalization capability, which are sine qua non of widespread deployment. (c) 2021 Elsevier B.V. All rights reserved.-
dc.languageEnglish-
dc.publisherELSEVIER-
dc.titleMulti-task vision transformer using low-level chest X-ray feature corpus for COVID-19 diagnosis and severity quantification-
dc.typeArticle-
dc.identifier.wosid000744256900001-
dc.identifier.scopusid2-s2.0-85119406601-
dc.type.rimsART-
dc.citation.volume75-
dc.citation.publicationnameMEDICAL IMAGE ANALYSIS-
dc.identifier.doi10.1016/j.media.2021.102299-
dc.contributor.localauthorYe, Jong Chul-
dc.contributor.nonIdAuthorPark, Sangjoon-
dc.contributor.nonIdAuthorKim, Gwanghyun-
dc.contributor.nonIdAuthorOh, Yujin-
dc.contributor.nonIdAuthorSeo, Joon Beom-
dc.contributor.nonIdAuthorLee, Sang Min-
dc.contributor.nonIdAuthorKim, Jin Hwan-
dc.contributor.nonIdAuthorMoon, Sungjun-
dc.contributor.nonIdAuthorLim, Jae-Kwang-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorCoronavirus disease-19-
dc.subject.keywordAuthorChest X-ray-
dc.subject.keywordAuthorVision transformer-
dc.subject.keywordAuthorMulti-task learning-
Appears in Collection
AI-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 49 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0