Evaluating task-specific augmentations in self-supervised pre-training for 3D medical image analysis

Cris H.B. Claessens, J.J.M. Hamm, Christiaan G.A. Viviers, Joost Nederend, D.J. Grünhagen, Pieter J. Tanis, Peter H.N. de With, Fons van der Sommen

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

2 Citations (Scopus)
32 Downloads (Pure)

Abstract

Self-supervised learning (SSL) has become a crucial approach for pre-training deep learning models in natural and medical image analysis. However, applying transformations designed for natural images to three-dimensional (3D) medical data poses challenges. This study explores the efficacy of specific augmentations in the context of self-supervised pre-training for volumetric medical images. A 3D non-contrastive framework is proposed for in-domain self-supervised pre-training on 3D gray-scale thorax CT data, incorporating four spatial and two intensity augmentations commonly used in 3D medical image analysis. The pre-trained models, adapted versions of ResNet-50 and Vision Transformer (ViT)-S, are evaluated on lung nodule classification and lung tumor segmentation tasks. The results indicate a significant impact of SSL, with a remarkable increase in AUC and DSC as compared to training from scratch. For classification, random scalings and random rotations play a fundamental role in achieving higher downstream performance, while intensity augmentations show limited contribution and may even degrade performance. For segmentation, random intensity histogram shifting enhances robustness, while other augmentations have marginal or negative impacts. These findings underscore the necessity of tailored data augmentations within SSL for medical imaging, emphasizing the importance of task-specific transformations for optimal model performance in complex 3D medical datasets.
Original languageEnglish
Title of host publicationMedical Imaging 2024
Subtitle of host publicationImage Processing
EditorsOlivier Colliot, Jhimli Mitra
PublisherSPIE
Number of pages8
ISBN (Electronic)9781510671560
DOIs
Publication statusPublished - 2 Apr 2024
EventSPIE Medical Imaging 2024 - San Diego, United States
Duration: 18 Feb 202423 Feb 2024

Publication series

NameProceedings of SPIE
Volume12926
ISSN (Print)1605-7422
ISSN (Electronic)2410-9045

Conference

ConferenceSPIE Medical Imaging 2024
Country/TerritoryUnited States
CitySan Diego
Period18/02/2423/02/24

Funding

Data used in this research were obtained from The Cancer Imaging Archive (TCIA),23 the National Lung Screening Trial (NLST),15,16 the LUNA16 challenge,21 and the Medical Segmentation Decathlon.22 Code used in this research was partially adapted from DINO\u2020.

Keywords

  • augmentations
  • medical imaging
  • pre-training
  • self-distillation
  • self-supervised learning
  • three-dimensional

Fingerprint

Dive into the research topics of 'Evaluating task-specific augmentations in self-supervised pre-training for 3D medical image analysis'. Together they form a unique fingerprint.

Cite this