Mask-conditioned latent diffusion for generating gastrointestinal polyp images
Macháček, Roman; Mozaffari, Leila; Sepasdar, Zahra; Parasa, Sravanthi; Halvorsen, Pål; Riegler, Michael; Thambawita, Vajira L B; Machacek, Roman
Chapter, Peer reviewed, Conference object
Accepted version
View/ Open
Date
2023Metadata
Show full item recordCollections
Original version
https://doi.org/10.1145/3592571.3592978Abstract
In order to take advantage of artificial intelligence (AI) solutions in endoscopy diagnostics, we must overcome the issue of limited annotations. These limitations are caused by the high privacy concerns in the medical field and the requirement of getting aid from experts for the time-consuming and costly medical data annotation process. In computer vision, image synthesis has made a significant contribution in recent years, as a result of the progress of generative adversarial networks (GANs) and diffusion probabilistic models (DPMs). Novel DPMs have outperformed GANs in text, image, and video generation tasks. Therefore, this study proposes a conditional DPM framework to generate synthetic gastrointestinal (GI) polyp images conditioned on given generated segmentation masks. Our experimental results show that our system can generate an unlimited number of high-fidelity synthetic polyp images with the corresponding ground truth masks of polyps. To test the usefulness of the generated data we trained binary image segmentation models to study the effect of using synthetic data. Results show that the best micro-imagewise intersection over union (IOU) of 0.7751 was achieved from DeepLabv3+ when the training data consists of both real data and synthetic data. However, the results reflect that achieving good segmentation performance with synthetic data heavily depends on model architectures.