Generating multi-pathological and multi-modal images and labels for brain MRI

Med Image Anal. 2024 Oct:97:103278. doi: 10.1016/j.media.2024.103278. Epub 2024 Jul 18.

Abstract

The last few years have seen a boom in using generative models to augment real datasets, as synthetic data can effectively model real data distributions and provide privacy-preserving, shareable datasets that can be used to train deep learning models. However, most of these methods are 2D and provide synthetic datasets that come, at most, with categorical annotations. The generation of paired images and segmentation samples that can be used in downstream, supervised segmentation tasks remains fairly uncharted territory. This work proposes a two-stage generative model capable of producing 2D and 3D semantic label maps and corresponding multi-modal images. We use a latent diffusion model for label synthesis and a VAE-GAN for semantic image synthesis. Synthetic datasets provided by this model are shown to work in a wide variety of segmentation tasks, supporting small, real datasets or fully replacing them while maintaining good performance. We also demonstrate its ability to improve downstream performance on out-of-distribution data.

Keywords: Brain MRI; Generative modelling; Medical imaging segmentation.

MeSH terms

  • Algorithms
  • Brain / diagnostic imaging
  • Deep Learning
  • Humans
  • Image Interpretation, Computer-Assisted / methods
  • Image Processing, Computer-Assisted / methods
  • Imaging, Three-Dimensional / methods
  • Magnetic Resonance Imaging* / methods
  • Multimodal Imaging / methods