Annotation-efficient deep learning for automatic medical image segmentation

Shanshan Wang*, Cheng Li*, Rongpin Wang, Zaiyi Liu, Meiyun Wang, Hongna Tan, Yaping Wu, Xinfeng Liu, Hui Sun, Rui Yang, Xin Liu, Jie Chen, Huihui Zhou, Ismail Ben Ayed, Hairong Zheng*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

149 Citations (Scopus)

Abstract

Automatic medical image segmentation plays a critical role in scientific research and medical care. Existing high-performance deep learning methods typically rely on large training datasets with high-quality manual annotations, which are difficult to obtain in many clinical applications. Here, we introduce Annotation-effIcient Deep lEarning (AIDE), an open-source framework to handle imperfect training datasets. Methodological analyses and empirical evaluations are conducted, and we demonstrate that AIDE surpasses conventional fully-supervised models by presenting better performance on open datasets possessing scarce or noisy annotations. We further test AIDE in a real-life case study for breast tumor segmentation. Three datasets containing 11,852 breast images from three medical centers are employed, and AIDE, utilizing 10% training annotations, consistently produces segmentation maps comparable to those generated by fully-supervised counterparts or provided by independent radiologists. The 10-fold enhanced efficiency in utilizing expert labels has the potential to promote a wide range of biomedical applications.

Original languageEnglish
Article number5915
JournalNature Communications
Volume12
Issue number1
DOIs
Publication statusPublished - 1 Dec 2021
Externally publishedYes

Fingerprint

Dive into the research topics of 'Annotation-efficient deep learning for automatic medical image segmentation'. Together they form a unique fingerprint.

Cite this