What Is the Best Way to Fine-Tune Self-supervised Medical Imaging Models?

Muhammad Osama Khan, Yi Fang

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

In recent years, self-supervised learning (SSL) has enabled significant breakthroughs via training large foundation models. These self-supervised pre-trained models are typically utilized for downstream tasks via end-to-end fine-tuning. However, it remains unclear whether end-to-end fine-tuning is truly optimal for effectively leveraging the pre-trained knowledge. This is especially true considering the diverse categories of SSL that capture distinct features, potentially requiring varied fine-tuning approaches. To bridge this research gap, we present the first comprehensive study discovering optimal fine-tuning strategies for self-supervised learning in medical imaging. Firstly, we develop strong contrastive and restorative SSL baselines that outperform SOTA methods on four diverse downstream tasks. Next, we conduct an extensive fine-tuning analysis across multiple pre-training and fine-tuning datasets, as well as various fine-tuning dataset sizes. Contrary to the conventional wisdom of fine-tuning only the last few layers of a pre-trained network, we show that fine-tuning intermediate layers is much more effective. Specifically, fine-tuning the second quarter (25–50%) of the network is optimal for contrastive SSL whereas fine-tuning the third quarter (50–75%) of the network is optimal for restorative SSL. Moreover, compared to the de-facto standard of end-to-end fine-tuning, our best fine-tuning strategy, which fine-tunes a shallower network consisting of the first three quarters (0–75%) of the pre-trained network, yields improvements of as much as 5.48%. Additionally, using these insights, we propose a simple yet effective method to leverage the complementary strengths of multiple SSL models, resulting in enhancements of up to 3.57%. Given the rapid progress in SSL, we hope these fine-tuning techniques will significantly improve the utility of self-supervised medical imaging models.

Original languageEnglish (US)
Title of host publicationMedical Image Understanding and Analysis - 28th Annual Conference, MIUA 2024, Proceedings
EditorsMoi Hoon Yap, Connah Kendrick, Ardhendu Behera, Timothy Cootes, Reyer Zwiggelaar
PublisherSpringer Science and Business Media Deutschland GmbH
Pages267-281
Number of pages15
ISBN (Print)9783031669545
DOIs
StatePublished - 2024
Event28th Annual Conference on Medical Image Understanding and Analysis, MIUA 2024 - Manchester, United Kingdom
Duration: Jul 24 2024Jul 26 2024

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume14859 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference28th Annual Conference on Medical Image Understanding and Analysis, MIUA 2024
Country/TerritoryUnited Kingdom
CityManchester
Period7/24/247/26/24

ASJC Scopus subject areas

  • Theoretical Computer Science
  • General Computer Science

Fingerprint

Dive into the research topics of 'What Is the Best Way to Fine-Tune Self-supervised Medical Imaging Models?'. Together they form a unique fingerprint.

Cite this