Improving Medical Image Classification in Noisy Labels Using only Self-supervised Pretraining

Bidur Khanal*, Binod Bhattarai, Bishesh Khanal, Cristian A. Linte

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingPublished conference contribution

1 Citation (Scopus)

Abstract

Noisy labels hurt deep learning-based supervised image classification performance as the models may overfit the noise and learn corrupted feature extractors. For natural image classification training with noisy labeled data, model initialization with contrastive self-supervised pretrained weights has shown to reduce feature corruption and improve classification performance. However, no works have explored: i) how other self-supervised approaches, such as pretext task-based pretraining, impact the learning with noisy label, and ii) any self-supervised pretraining methods alone for medical images in noisy label settings. Medical images often feature smaller datasets and subtle inter-class variations, requiring human expertise to ensure correct classification. Thus, it is not clear if the methods improving learning with noisy labels in natural image datasets such as CIFAR would also help with medical images. In this work, we explore contrastive and pretext task-based self-supervised pretraining to initialize the weights of a deep learning classification model for two medical datasets with self-induced noisy labels—NCT-CRC-HE-100K tissue histological images and COVID-QU-Ex chest X-ray images. Our results show that models initialized with pretrained weights obtained from self-supervised learning can effectively learn better features and improve robustness against noisy labels.

Original languageEnglish
Title of host publicationDEMI 2023: Data Engineering in Medical Imaging
EditorsBinod Bhattarai, Sharib Ali, Anita Rau, Anh Nguyen, Ana Namburete, Razvan Caramalau, Danail Stoyanov
PublisherSpringer Science and Business Media Deutschland GmbH
Pages78-90
Number of pages13
ISBN (Print)9783031449918
DOIs
Publication statusPublished - 1 Oct 2023
Event1st MICCAI Workshop on Data Engineering in Medical Imaging, DEMI 2023 - Vancouver, Canada
Duration: 8 Oct 20238 Oct 2023

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume14314 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference1st MICCAI Workshop on Data Engineering in Medical Imaging, DEMI 2023
Country/TerritoryCanada
CityVancouver
Period8/10/238/10/23

Bibliographical note

Acknowledgments. Research reported in this publication was supported by the National Institute of General Medical Sciences Award No. R35GM128877 of the National Institutes of Health, the Office of Advanced Cyber Infrastructure Award No. 1808530 of the National Science Foundation, and the Division Of Chemistry, Bioengineering, Environmental, and Transport Systems Award No. 2245152 of the National Science Foundation. We would like to thank the Research Computing team [31] at the Rochester Institute of Technology for proving computing resources for this research.

Keywords

  • feature extraction
  • label noise
  • learning with noisy labels
  • medical image classification
  • self-supervised pretraining
  • warm-up obstacle

Fingerprint

Dive into the research topics of 'Improving Medical Image Classification in Noisy Labels Using only Self-supervised Pretraining'. Together they form a unique fingerprint.

Cite this