• Media type: E-Article
  • Title: Dilated conditional GAN for bone suppression in chest radiographs with enforced semantic features
  • Contributor: Zhou, Zhizhen; Zhou, Luping; Shen, Kaikai
  • Published: Wiley, 2020
  • Published in: Medical Physics, 47 (2020) 12, Seite 6207-6215
  • Language: English
  • DOI: 10.1002/mp.14371
  • ISSN: 0094-2405; 2473-4209
  • Origination:
  • Footnote:
  • Description: PurposeThe purpose of this essay is to improve computer‐aided diagnosis of lung diseases by the removal of bone structures imagery such as ribs and clavicles, which may shadow a clinical view of lesions. This paper aims to develop an algorithm to suppress the imaging of bone structures within clinical x‐ray images, leaving a residual portrayal of lung tissue; such that these images can be used to better serve applications, such as lung nodule detection or pathology based on the radiological reading of chest x rays.MethodsWe propose a conditional Adversarial Generative Network (cGAN) (Mirza and Osindero, Conditional generative adversarial nets, 2014.) model, consisting of a generator and a discriminator, for the task of bone shadow suppression. The proposed model utilizes convolutional operations to expand the size of the receptive field of the generator without losing contextual information while downsampling the image. It is trained by enforcing both the pixel‐wise intensity similarity and the semantic‐level visual similarity between the generated x‐ray images and the ground truth, via optimizing an L‐1 loss of the pixel intensity values on the generator side and a feature matching loss on the discriminator side, respectively.ResultsThe framework we propose is trained and tested on an open‐access chest radiograph dataset for benchmark. Results show that our model is capable of generating bone‐suppressed images of outstanding quality with a limited number of training samples (N = 272).ConclusionsOur approach outperforms current state‐of‐the‐art bone suppression methods using x‐ray images. Instead of simply downsampling images at different scales, our proposed method mitigates the loss of contextual information by utilizing dilated convolutions, which gains a noticeable quality improvement for the outputs. On the other hand, our experiment shows that enforcing the semantic similarity between the generated and the ground truth images assists the adversarial training process and achieves better perceptual quality.