Improving Quality of Medical Scans using GANs
Improving Quality of Medical Scans using GANs
ISSN No:-2456-2165
Abstract:- Improving the quality of medical images is Despite even the medical imaging is crucial for
essential for precise diagnosis and treatment planning. diagnosis, therapy planning, obtaining representative and
When low quality images are used to train the neural diverse datasets for machine learning model training is still
network model, the good accuracy cannot be achieved. difficult because of privacy issues and restricted access to
Nowadays, Generative Adversarial Networks (GANs) uncommon cases. In order to enhance current datasets, this
have become a potent image enhancement tool that can research investigates the possibility of using Generative
provide a fresh method for raising the caliber of medical Adversarial Networks (GANs) to generate artificial medical
images. In order to improve medical images, this paper scans [3]. We explore the difficulties in medical imaging, the
presents a GAN-based framework that reduces noise, shortcomings of conventional data augmentation methods,
increases resolution, and corrects artifacts. The suggested and the requirement for intelligent data augmentation
technique makes use of a generator network to convert strategies. A major challenge in developing models fit for
low-quality images into their high-quality equivalents, clinical use is the lack of sufficient diverse labelled training
and a discriminator network to assess the veracity of the data [4]. Additionally, class disparity frequently occurs in
improved images. To ensure robustness across various medical data. The idea behind the GANs is that it has a
modalities, the model is trained on a diverse dataset of generator and discriminator built in, which makes it useful for
medical images, including MRI, CT, and X-ray scans. comparing human scans. It is helpful in analysis for
Our experimental results show that GAN-based method improvement of the medical scans. The practical and research
significantly improves the image quality when compared was significantly improved, suggesting that GAN-based data
to conventional methods, as evidenced by enhanced peak augmentation holds promise for medical applications [5].
signal-to-noise ratio (PSNR) and structural similarity
index (SSIM) according to quantitative evaluations. This II. LITERATURE REVIEW AND
study emphasizes the value of incorporating deep NEED OF THE STUDY
learning methods into medical image processing pipelines
and the potential of GANs to advance medical imaging Medical image quality can now be improved with
technology so that a robust neural network model can be greater effectiveness thanks to Generative Adversarial
designed. Networks (GANs), which can handle tasks like noise
reduction, resolution enhancement, and image reconstruction.
Keywords:- Medical Images Quality, Convolutional Neural Many researchers did work on the GAN, some most cited
Networks, Generative Adversarial Networks (GANs), Peak research work is taken here. Qiaoying Yang et al [6] showed
Signal to Noise Ratio (PSNR), Structural Similarity Index that GANs could reduce noise while maintaining diagnostic
(SSIM). accuracy, resulting in a significant improvement in low-dose
CT scan quality.
I. INTRODUCTION
Moreover, super-resolution has been achieved with
Medical imaging technologies, which provide non- GANs, improving the resolution of medical images. The
invasive observation of internal body components, are work of Chunyuan Li et al. [7] demonstrated that GAN-based
essential to contemporary healthcare. Examples of these approaches perform better in this context than traditional
technologies include CT, MRI, and X-ray. These pictures are methods by reconstructing high-resolution images with better
useful diagnostic tools that help doctors find anomalies, preservation of fine details. Specifically, the Super-
monitor the course of diseases, and schedule treatments. Resolution GAN (SRGAN) produced high-quality images
However, the availability of varied and representative that were more accurate and aesthetically pleasing. Moreover,
datasets for training is a major factor in how well machine GANs have been used in medical image synthesis, which
learning models perform in medical image interpretation [1]. makes it possible to produce high-quality images from
The part of Convolutional Neural Networks (CNNs) have imperfect or low-quality data.
demonstrated remarkable performance in segmenting images
and recognizing objects in recent years. Utilizing these Yibin Song et al [8] did work on Liver lesion detection
networks for clinical tasks, such as classifying medical and classification with novel neural network architectures.
pictures, and segmenting organs and diseases, should Zhu Jun-Yan et al. [9] develop a model for image to image
improve medical judgment [2]. translation and created high-quality MRI images from CT
scans using GANs. Zhang, et al [10] proposed a translating
Generator Network: This generator learns to produce Convolutional Layers: DCGANs employ convolutional
visuals that mimic actual data by using random noise as layers for the generator and discriminator rather than fully
input. Usually, it is made up of convolutional layers, linked layers. Convolutional layers work effectively for
which are followed by upsampling layers such as applications like picture production because they can
transposed or nearest-neighbor convolutions. The capture spatial patterns in images.
generator's output is an image.
Batch Normalization: To stabilize training and quicken
Discriminator Network: The discriminator is a convergence, batch normalization is frequently used in
convolutional neural network (CNN) that learns to both the discriminator and generator networks. It makes
distinguish between real images from the dataset and fake the decreasing the internal covariate shift issue through
ones from the generator. It accepts an image as input. It the activations of each layer.
produces a likelihood score that indicates the authenticity
of the provided image.
The main work or the duty is the image to image Pix2Pix is a popular tool for many computer vision
transition or Pix2Pix, is a conditional generative adversarial applications, including semantic segmentation, style transfer,
network (GAN) designed especially for image-to-image image colorization, and more. It is a useful tool for many
translation applications. This is how Pix2Pix functions and image alteration applications because of its capacity to learn
what sets it apart: mappings across various visual domains from paired training
data
Conditional GAN Framework: Pix2Pix expands the
capabilities of the GAN framework by including a StarGAN
conditional setting. In a traditional GAN, the generator With just one model, StarGAN is a flexible GAN
generates images from random noise as input, while the architecture that can translate images between different
discriminator looks for differences between genuine and domains. It makes it possible to create a variety of medical
fake images. The generator in Pix2Pix gains the capability images with various features and qualities. StarGAN, or "Star
to translate images between distinct domains because both Generative Adversarial Network," is a generative adversarial
the discriminator and generator are conditioned on input network (GAN) architecture designed for multi-domain
images [15]. image-to- image translation. This is how StarGAN functions
and what makes it unique:
Translation of Images to Images: Pix2Pix is especially
made for jobs in which the input and "Image-to-Image Multi-Domain Image Translation: StarGAN can handle
Translation with Conditional Adversarial Networks," or many domains inside a single model, in contrast to
Pix2Pix for short, is a sort of conditional generative standard image-to-image translation techniques that call
adversarial network (GAN) in which the output images for distinct models for each translation assignment. It
are paired. This covers tasks such as mapping satellite supports several target domains at once and has the ability
imagery to maps, creating realistic graphics from to convert images across them. It may, for instance,
sketches, transforming daytime views into nighttime convert pictures of human faces into many facial styles,
sceneries, and more. The network gains the ability to representing various ages, genders, and races [17].
translate input images from one domain into equivalent
output images in another. Single Generator and Discriminator: For all domain
translations, StarGAN has a single generator and
Generator and Discriminator Networks: In Pix2Pix, an discriminator architecture. This indicates that the
encoder-decoder architecture is commonly used as the accountable generator is the same for creating images in
generator. The input image is encoded by the encoder into every target domain, and the discriminator gains the
a latent representation, which is then decoded by the ability to discern between images produced by the
decoder to produce the output image. The part of the generator that are false and genuine images from any
Convolutional neural networks are used in the domain.
Fig 3 Testing and Updating the Curve Comparison of GAN-based Augment Training against no Augment
Table 2: AUC Results for Augmented Classes across Dataset Regimens and Augmentation Strategies
Dataset Size Pathology Not the augment Standard Augment Generative adversarial network augment
The Lung part 1.727 1.728 1.758
1% Pleural 1.566 1.550 1.594
Broken part 1.583 1.601 1.656
The Lung part 2.809 1.796 2.852
10% Pleural 1.632 1.655 1.670
Broken part 1.700 1.723 1.742
The Lung part 1.826 1.822 1.828
50% Pleural 1.710 1.696 1.706
Broken part 1.789 1.780 1.793
The Lung part 1.835 2.945 1.834
100% Pleural 1.721 1.712 1.727
Broken part 0031.811 1.793 1.807