KLI

Self-evolving vision transformer for chest X-ray diagnosis through knowledge distillation

Metadata Downloads
Abstract
Although deep learning-based computer-aided diagnosis systems have recently achieved expert-level performance, developing a robust model requires large, high-quality data with annotations that are expensive to obtain. This situation poses a conundrum that annually-collected chest x-rays cannot be utilized due to the absence of labels, especially in deprived areas. In this study, we present a framework named distillation for self-supervision and self-train learning (DISTL) inspired by the learning process of the radiologists, which can improve the performance of vision transformer simultaneously with self-supervision and self-training through knowledge distillation. In external validation from three hospitals for diagnosis of tuberculosis, pneumothorax, and COVID-19, DISTL offers gradually improved performance as the amount of unlabeled data increase, even better than the fully supervised model with the same amount of labeled data. We additionally show that the model obtained with DISTL is robust to various real-world nuisances, offering better applicability in clinical setting.
Author(s)
Sangjoon ParkGwanghyun KimYujin OhJoon Beom SeoSang Min LeeJin Hwan KimSungjun MoonJae-Kwang LimChang Min ParkJong Chul Ye
Issued Date
2022
Type
Article
Keyword
ChestCOVID-19 Pandemic, 2020-DiagnosisDistillationImage processingPneumothoraxPublic healthSupervisionTrainingTuberculosis
DOI
10.1038/s41467-022-31514-x
URI
https://oak.ulsan.ac.kr/handle/2021.oak/15076
Publisher
NATURE COMMUNICATIONS
Language
영어
ISSN
2041-1723
Citation Volume
13
Citation Number
1
Citation Start Page
1
Citation End Page
11
Appears in Collections:
Engineering > Medical Engineering
공개 및 라이선스
  • 공개 구분공개
파일 목록
  • 관련 파일이 존재하지 않습니다.

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.