Abstract

The last few years have seen a boom in using generative models to augment real datasets, as synthetic data can effectively model real data distributions and provide privacy-preserving, shareable datasets that can be used to train deep learning models. However, most of these methods are 2D and provide synthetic datasets that come, at most, with categorical annotations. The generation of paired images and segmentation samples that can be used in downstream, supervised segmentation tasks remains fairly uncharted territory. This work proposes a two-stage generative model capable of producing 2D and 3D semantic label maps and corresponding multi-modal images. We use a latent diffusion model for label synthesis and a VAE-GAN for semantic image synthesis. Synthetic datasets provided by this model are shown to work in a wide variety of segmentation tasks, supporting small, real datasets or fully replacing them while maintaining good performance. We also demonstrate its ability to improve downstream performance on out-of-distribution data.

Original languageEnglish
Article number103278
Pages (from-to)103278
JournalMedical Image Analysis
Volume97
Early online date18 Jul 2024
DOIs
Publication statusPublished - Oct 2024

Fingerprint

Dive into the research topics of 'Generating multi-pathological and multi-modal images and labels for brain MRI'. Together they form a unique fingerprint.

Cite this