<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xml:lang="en" article-type="review-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CMES</journal-id>
<journal-id journal-id-type="nlm-ta">CMES</journal-id>
<journal-id journal-id-type="publisher-id">CMES</journal-id>
<journal-title-group>
<journal-title>Computer Modeling in Engineering &#x0026; Sciences</journal-title>
</journal-title-group>
<issn pub-type="epub">1526-1506</issn>
<issn pub-type="ppub">1526-1492</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">67108</article-id>
<article-id pub-id-type="doi">10.32604/cmes.2025.067108</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Review</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>A Survey of Generative Adversarial Networks for Medical Images</article-title>
<alt-title alt-title-type="left-running-head">A Survey of Generative Adversarial Networks for Medical Images</alt-title>
<alt-title alt-title-type="right-running-head">A Survey of Generative Adversarial Networks for Medical Images</alt-title>
</title-group>
<contrib-group>
<contrib id="author-1" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Sagheer</surname><given-names>Sameera V. Mohd</given-names></name><xref ref-type="aff" rid="aff-1">1</xref><xref ref-type="author-notes" rid="afn1">#</xref><email>sameeravm@gmail.com</email></contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western"><surname>Nimitha</surname><given-names>U.</given-names></name><xref ref-type="aff" rid="aff-2">2</xref><xref ref-type="author-notes" rid="afn1">#</xref></contrib>
<contrib id="author-3" contrib-type="author">
<name name-style="western"><surname>Ameer</surname><given-names>P. M.</given-names></name><xref ref-type="aff" rid="aff-2">2</xref></contrib>
<contrib id="author-4" contrib-type="author">
<name name-style="western"><surname>Parayangat</surname><given-names>Muneer</given-names></name><xref ref-type="aff" rid="aff-3">3</xref></contrib>
<contrib id="author-5" contrib-type="author">
<name name-style="western"><surname>Abbas</surname><given-names>Mohamed</given-names></name><xref ref-type="aff" rid="aff-3">3</xref></contrib>
<contrib id="author-6" contrib-type="author">
<name name-style="western"><surname>Arunachalam</surname><given-names>Krishna Prakash</given-names></name><xref ref-type="aff" rid="aff-4">4</xref></contrib>
<aff id="aff-1"><label>1</label><institution>Department of Biomedical Engineering, KMCT College of Engineering for Women</institution>, <addr-line>Kozhikode, 673601, Kerala</addr-line>, <country>India</country></aff>
<aff id="aff-2"><label>2</label><institution>Department of Electronics and Communication Engineering, National Institute of Technology Calicut</institution>, <addr-line>Kozhikode, 673601, Kerala</addr-line>, <country>India</country></aff>
<aff id="aff-3"><label>3</label><institution>Electrical Engineering Department, College of Engineering, King Khalid University</institution>, <addr-line>Abha, 61413</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff-4"><label>4</label><institution>Departamento de Ciencias de la Construcci&#x00F3;n, Facultad de Ciencias de la Construcci&#x00F3;n Ordenamiento Territorial, Universidad Tecnol&#x00F3;gica Metropolitana</institution>, <addr-line>Santiago, 7800002</addr-line>, <country>Chile</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label>Corresponding Author: Sameera V. Mohd Sagheer. Email: <email>sameeravm@gmail.com</email></corresp>
<fn id="afn1">
<p><sup>#</sup>These authors contributed equally to this work</p>
</fn>
</author-notes>
<pub-date date-type="collection" publication-format="electronic">
<year>2026</year>
</pub-date>
<pub-date date-type="pub" publication-format="electronic">
<day>26</day><month>2</month><year>2026</year>
</pub-date>
<volume>146</volume>
<issue>2</issue>
<elocation-id>4</elocation-id>
<history>
<date date-type="received">
<day>25</day>
<month>4</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>08</day>
<month>8</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2026 The Authors.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Published by Tech Science Press.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CMES_67108.pdf"></self-uri>
<abstract>
<p>Over the years, Generative Adversarial Networks (<inline-formula id="ieqn-1"><mml:math id="mml-ieqn-1"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>) have revolutionized the medical imaging industry for applications such as image synthesis, denoising, super resolution, data augmentation, and cross-modality translation. The objective of this review is to evaluate the advances, relevances, and limitations of <inline-formula id="ieqn-2"><mml:math id="mml-ieqn-2"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical imaging. An organised literature review was conducted following the guidelines of <italic>PRISMA</italic> (Preferred Reporting Items for Systematic Reviews and Meta-Analyses). The literature considered included peer-reviewed papers published between <inline-formula id="ieqn-3"><mml:math id="mml-ieqn-3"><mml:mn>2020</mml:mn></mml:math></inline-formula> and <inline-formula id="ieqn-4"><mml:math id="mml-ieqn-4"><mml:mn>2025</mml:mn></mml:math></inline-formula> across databases including PubMed, <italic>IEEE</italic> Xplore, and Scopus. The studies related to applications of <italic>GAN</italic> architectures in medical imaging with reported experimental outcomes and published in English in reputable journals and conferences were considered for the review. Thesis, white papers, communication letters, and non-English articles were not included for the same. <italic>CLAIM</italic> based quality assessment criteria were applied to the included studies to assess the quality. The study classifies diverse <italic>GAN</italic> architectures, summarizing their clinical applications, technical performances, and their implementation hardships. Key findings reveal the increasing applications of <inline-formula id="ieqn-5"><mml:math id="mml-ieqn-5"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> for enhancing diagnostic accuracy, reducing data scarcity through synthetic data generation, and supporting modality translation. However, concerns such as limited generalizability, lack of clinical validation, and regulatory constraints persist. This review provides a comprehensive study of the prevailing scenario of <inline-formula id="ieqn-6"><mml:math id="mml-ieqn-6"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical imaging and highlights crucial research gaps and future directions. Though <inline-formula id="ieqn-7"><mml:math id="mml-ieqn-7"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> hold transformative capability for medical imaging, their integration into clinical use demands further validation, interpretability, and regulatory alignment.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Generative adversarial networks</kwd>
<kwd>medical images</kwd>
<kwd>denoising</kwd>
<kwd>segmentation</kwd>
<kwd>translation</kwd>
</kwd-group>
<funding-group>
<award-group id="awg1">
<funding-source>King Khalid University</funding-source>
<award-id>RGP2/540/46</award-id>
</award-group>
</funding-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>Medical imaging is a fundamental component of modern healthcare, offering non-invasive methods to visualize the internal structures of the human body. It supports diagnosing, planning treatment, and monitoring a range of medical conditions, utilizing common imaging techniques such as <italic>X</italic>-rays, Magnetic Resonance <inline-formula id="ieqn-8"><mml:math id="mml-ieqn-8"><mml:mo stretchy="false">(</mml:mo><mml:mi>M</mml:mi><mml:mi>R</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, Computed Tomography <inline-formula id="ieqn-9"><mml:math id="mml-ieqn-9"><mml:mo stretchy="false">(</mml:mo><mml:mi>C</mml:mi><mml:mi>T</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, Ultrasound, and Positron Emission Tomography <inline-formula id="ieqn-10"><mml:math id="mml-ieqn-10"><mml:mo stretchy="false">(</mml:mo><mml:mi>P</mml:mi><mml:mi>E</mml:mi><mml:mi>T</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>.</p>
<p>These imaging methods, such as <italic>MR</italic>, <italic>CT</italic>, <italic>PET</italic>, and ultrasound, serve various diagnostic purposes, offering detailed insights into the body&#x2019;s internal structures. Each modality is suited for specific clinical applications, with <italic>MR</italic> excelling in soft tissue imaging, <italic>CT</italic> providing high-resolution bone images, <italic>PET</italic> detecting metabolic activity, and ultrasound enabling real-time visualization of soft tissues. An outline of the different medical imaging techniques is provided below, outlining their specific uses and advantages in healthcare.
<list list-type="simple">
<list-item><label>1.</label><p><bold>Ultrasound (US) Images:</bold> <italic>US</italic> imaging is widely utilized in diagnostic fields like cardiology, obstetrics, and gynecology due to its ability to generate high-resolution images without subjecting patients to ionizing radiation [<xref ref-type="bibr" rid="ref-1">1</xref>]. The technique works by emitting high-frequency sound waves (typically between <inline-formula id="ieqn-11"><mml:math id="mml-ieqn-11"><mml:mn>1</mml:mn></mml:math></inline-formula> and <inline-formula id="ieqn-12"><mml:math id="mml-ieqn-12"><mml:mn>5</mml:mn></mml:math></inline-formula> MHz) from a probe into the body. As these sound waves pass through the body, they interact with different tissue boundaries. Some of the waves are reflected back to the probe. The probe captures these reflected waves and transmits the data to the ultrasound machine. The distance between the probe and the organ boundaries information is then used to create a two-dimensional image on the screen, showing the distances and intensities of the reflections [<xref ref-type="bibr" rid="ref-2">2</xref>]. <xref ref-type="fig" rid="fig-1">Fig. 1</xref> illustrates the formation of an <italic>US</italic> image. Diagnostic ultrasound typically operates at frequencies ranging from <inline-formula id="ieqn-13"><mml:math id="mml-ieqn-13"><mml:mn>2</mml:mn></mml:math></inline-formula> to <inline-formula id="ieqn-14"><mml:math id="mml-ieqn-14"><mml:mn>15</mml:mn></mml:math></inline-formula> MHz. Higher frequencies produce better image resolution but have lower penetration depth due to increased absorption and attenuation. For this reason, high-frequency ultrasound is used to visualize superficial structures like the thyroid, while lower frequencies are employed for imaging deeper organs. Ultrasound is a non-invasive imaging technique, making it a preferred option in many medical procedures. However, it does have limitations. They cannot be used to image bones cannot be imaged as they block or absorb the ultrasound waves [<xref ref-type="bibr" rid="ref-3">3</xref>&#x2013;<xref ref-type="bibr" rid="ref-6">6</xref>].</p>
</list-item>

<list-item><label>2.</label><p><bold>Magnetic Resonance (MR) Images:</bold> <italic>MR</italic> imaging uses magnetic fields and radio waves to generate detailed images of internal body structures that are difficult to capture with other imaging modalities. The human body is composed of billions of hydrogen atoms, which align with the magnetic field when exposed to it. This alignment causes the hydrogen atoms, which are positively charged, to orient uniformly. A pulse of radio frequency energy is applied to disrupt this alignment, causing the protons to shift. The protons emit energy when they return to the initial position. The intensity of this released energy is measured and displayed on a gray scale, forming cross-sectional images of the body. <italic>MR</italic> images are created using complex values that correspond to the Fourier transform of the magnetization distribution [<xref ref-type="bibr" rid="ref-7">7</xref>&#x2013;<xref ref-type="bibr" rid="ref-9">9</xref>]. <xref ref-type="fig" rid="fig-2">Fig. 2</xref> illustrates the formation of <italic>MR</italic> image.</p>
</list-item>

<list-item><label>3.</label><p><bold>Computed Tomography (CT) Images:</bold> A <italic>CT</italic> scan employs computer algorithms to process multiple <italic>X</italic>-ray images procured from various angles. The combination of these images generate cross-sectional (tomographic) images of a given region within the scanned object. This technique is particularly useful in detecting hemorrhages and other conditions that may resemble a stroke, such as tumors or subdural/extradural hematomas [<xref ref-type="bibr" rid="ref-10">10</xref>]. However, <italic>CT</italic> imaging relies on ionizing radiation, and the exposure from this radiation accumulates over time. To minimize the impact of ionizing radiation, Low Dose Computed Tomography (<italic>LDCT</italic>) images are generated as an alternative [<xref ref-type="bibr" rid="ref-11">11</xref>&#x2013;<xref ref-type="bibr" rid="ref-14">14</xref>].</p></list-item>
<list-item><label>4.</label><p><bold>Positron Emission Tomography (PET) Images:</bold> <italic>PET</italic> is a molecular imaging method that has quickly become a vital tool for functional imaging. <italic>PET</italic> works by generating images of the body based on the radiation emitted by radioactive substances introduced into the body. These substances, often tagged with short-lived radioactive isotopes like Carbon-<inline-formula id="ieqn-15"><mml:math id="mml-ieqn-15"><mml:mn>11</mml:mn></mml:math></inline-formula>, are created by bombarding standard chemicals with neutrons. When a positron emitted by the radioactive material interacts with an electron in the tissue, gamma rays are released, which the <italic>PET</italic> scanner detects. <italic>PET</italic> can visualize blood flow and biochemical processes. Unlike structural imaging methods, <italic>PET</italic> focuses on the functionality of organs and the nervous system. Despite its valuable applications, the technology is costly and not widely available [<xref ref-type="bibr" rid="ref-15">15</xref>&#x2013;<xref ref-type="bibr" rid="ref-17">17</xref>]. The gamma-ray detectors are used to identify pairs of gamma photons emitted in opposite directions, which are captured by two corresponding detector elements. Once these events are detected, the analog front-end circuitry produces an event signal. If the amplitude of the incoming gamma-ray pulse exceeds a predefined threshold, a trigger signal is generated by the analog front-end. This signal is then sent to the time-to-digital converters <inline-formula id="ieqn-16"><mml:math id="mml-ieqn-16"><mml:mo stretchy="false">(</mml:mo><mml:mi>T</mml:mi><mml:mi>D</mml:mi><mml:mi>C</mml:mi><mml:mi>s</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, which convert the time interval of the detected event into a digital representation.</p></list-item>
</list><fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>Formation of <italic>US</italic> image [<xref ref-type="bibr" rid="ref-2">2</xref>]</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-1.tif"/>
</fig></p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>Formation of <italic>MR</italic> image [<xref ref-type="bibr" rid="ref-8">8</xref>]</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-2.tif"/>
</fig>
<p>Each imaging modality provides unique understanding, helping clinicians in making better decisions. The complexity and quantity of medical imaging data emphasizes for advanced computational tools to support analysis and interpretation. Generative Adversarial Networks <inline-formula id="ieqn-17"><mml:math id="mml-ieqn-17"><mml:mo stretchy="false">(</mml:mo><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> are well-suited to the unique expectations of medical imaging due to their ability to generate highly realistic images unlike the traditional discriminative models that only classify patterns. This capability is especially important in medical imaging, where annotated datasets are often limited, imbalanced, or expensive to obtain thus improving model robustness and performance.</p>
<p>Moreover, <inline-formula id="ieqn-18"><mml:math id="mml-ieqn-18"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> excel at image to image translation tasks or enhancing image quality thus making them ideal for applications where clarity and detail are critical for diagnosis. These strengths collectively make <inline-formula id="ieqn-19"><mml:math id="mml-ieqn-19"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> a powerful tool for advancing medical imaging. The following section gives a detailed introduction to <inline-formula id="ieqn-20"><mml:math id="mml-ieqn-20"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>.</p>
<sec id="s1_1">
<label>1.1</label>
<title>Overview of Generative Adverserial Network (GAN)</title>
<p>Recent advancements in computing power and big data analysis have significantly boosted the development of Artificial Intelligence (<italic>AI</italic>)&#x2014;systems that mimic human cognitive abilities, such as learning, problem-solving, and decision-making [<xref ref-type="bibr" rid="ref-18">18</xref>&#x2013;<xref ref-type="bibr" rid="ref-22">22</xref>]. <italic>AI</italic> can process and analyze large datasets efficiently. Machine Learning (<italic>ML</italic>), a subset of <italic>AI</italic>, learns from data by identifying patterns and features [<xref ref-type="bibr" rid="ref-23">23</xref>]. Two fundamental types of machine learning are supervised learning and unsupervised learning. While supervised learning [<xref ref-type="bibr" rid="ref-24">24</xref>,<xref ref-type="bibr" rid="ref-25">25</xref>] requires labeled data for training, unsupervised learning discovers patterns in unlabeled data, making it more applicable in scenarios where labeling is infeasible. Among these, supervised learning is the most widely utilized and successful approach. In supervised learning, algorithms are provided with a data set comprising pairs of input and output examples. The algorithm learns to map each input with its corresponding output, effectively associating input examples to output examples. A widely used form of supervised learning is classification. Once trained, supervised learning algorithms can achieve accuracy levels that exceed human performance, making them essential in various products and services. Despite these advancements, the learning process has limitations compared to human abilities. Current supervised learning approaches typically require millions of training examples [<xref ref-type="bibr" rid="ref-26">26</xref>]. To address these challenges, researchers are increasingly focusing on unsupervised learning, to reduce dependence on extensive human supervision and decrease the number of training examples needed. In general, the purpose of unsupervised learning is to extract meaningful information from a data set containing unlabeled input examples. Unlike supervised learning, unsupervised learning seeks to uncover useful patterns from unlabeled data [<xref ref-type="bibr" rid="ref-27">27</xref>,<xref ref-type="bibr" rid="ref-28">28</xref>]. Two well-known applications of unsupervised learning are clustering and dimensionality reduction.</p>
<p>A significant approach in unsupervised learning is generative modeling. Generative modeling aims to approximate the true data distribution <inline-formula id="ieqn-21"><mml:math id="mml-ieqn-21"><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> with a model distribution <inline-formula id="ieqn-22"><mml:math id="mml-ieqn-22"><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>o</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. This is achieved by designing a function <inline-formula id="ieqn-23"><mml:math id="mml-ieqn-23"><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>o</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo>;</mml:mo><mml:mi>y</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> with parameters <inline-formula id="ieqn-24"><mml:math id="mml-ieqn-24"><mml:mi>y</mml:mi></mml:math></inline-formula> and optimizing these parameters to make <inline-formula id="ieqn-25"><mml:math id="mml-ieqn-25"><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>o</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> as similar to <inline-formula id="ieqn-26"><mml:math id="mml-ieqn-26"><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> as possible. A common technique for generative modeling is maximum likelihood estimation, which minimizes the Kullback-Leibler (<italic>KL</italic>) divergence between <inline-formula id="ieqn-27"><mml:math id="mml-ieqn-27"><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> and <inline-formula id="ieqn-28"><mml:math id="mml-ieqn-28"><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>o</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. Traditionally, explicit density models with simple probabilistic forms were used for such tasks. However, with advances in machine learning, more complex models have been developed to handle high-dimensional data. Diffusion models have recently gained attention in the field of computer vision for their impressive performance in generative tasks. These models consist of a two-phase process: a forward noise injection phase and a reverse reconstruction phase. During the forward phase, the original data is corrupted by adding Gaussian noise. The reverse phase consists of a neural network trained to reconstruct the original data by denoising it. Although these models can produce high quality and diverse outputs, they are computationally intensive and often suffer from slow generation times [<xref ref-type="bibr" rid="ref-29">29</xref>]. To address these limitations, Goodfellow et al. [<xref ref-type="bibr" rid="ref-30">30</xref>] introduced Generative Adversarial Networks (<inline-formula id="ieqn-29"><mml:math id="mml-ieqn-29"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>) [<xref ref-type="bibr" rid="ref-31">31</xref>], a novel generative model that has become a prominent tool for tasks like image denoising, translation, segmentation, and reconstruction. Generative Adversarial Networks (<inline-formula id="ieqn-30"><mml:math id="mml-ieqn-30"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>) [<xref ref-type="bibr" rid="ref-32">32</xref>&#x2013;<xref ref-type="bibr" rid="ref-34">34</xref>] enables unsupervised learning by generating new data samples from an existing data distribution.</p>
<p>Formally, a <italic>GAN</italic> consists of two models: a generator <inline-formula id="ieqn-31"><mml:math id="mml-ieqn-31"><mml:mi>G</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>z</mml:mi><mml:mo>;</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> and a discriminator <inline-formula id="ieqn-32"><mml:math id="mml-ieqn-32"><mml:mi>D</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo>;</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mi>d</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. The generator maps noise vectors <inline-formula id="ieqn-33"><mml:math id="mml-ieqn-33"><mml:mi>z</mml:mi><mml:mo>&#x223C;</mml:mo><mml:msub><mml:mi>p</mml:mi><mml:mi>z</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>z</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> to the data space, creating synthetic samples <inline-formula id="ieqn-34"><mml:math id="mml-ieqn-34"><mml:mi>G</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>z</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. The discriminator outputs a probability <inline-formula id="ieqn-35"><mml:math id="mml-ieqn-35"><mml:mi>D</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2208;</mml:mo><mml:mo stretchy="false">[</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">]</mml:mo></mml:math></inline-formula> representing the likelihood that <inline-formula id="ieqn-36"><mml:math id="mml-ieqn-36"><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> came from the real data distribution <inline-formula id="ieqn-37"><mml:math id="mml-ieqn-37"><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mtext>data</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> rather than <inline-formula id="ieqn-38"><mml:math id="mml-ieqn-38"><mml:mi>G</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>z</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. The two models are trained simultaneously in a minimax game and the equation as follows [<xref ref-type="bibr" rid="ref-30">30</xref>]:
<disp-formula id="eqn-1"><label>(1)</label><mml:math id="mml-eqn-1" display="block"><mml:munder><mml:mo movablelimits="true" form="prefix">min</mml:mo><mml:mi>G</mml:mi></mml:munder><mml:munder><mml:mo movablelimits="true" form="prefix">max</mml:mo><mml:mi>D</mml:mi></mml:munder><mml:mi>V</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>D</mml:mi><mml:mo>,</mml:mo><mml:mi>G</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="double-struck">E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x223C;</mml:mo><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mrow><mml:mtext>data</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mi>log</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mi>D</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>]</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="double-struck">E</mml:mi></mml:mrow><mml:mrow><mml:mi>z</mml:mi><mml:mo>&#x223C;</mml:mo><mml:msub><mml:mi>p</mml:mi><mml:mi>z</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>z</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mi>log</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mi>D</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>G</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>z</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo>)</mml:mo></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>This adversarial process drives the generator to improve its outputs such that they are indistinguishable from real samples, while the discriminator becomes better at detection.</p>
<p>In general, the two neural networks: a generator and a discriminator, are designed to compete with each other [<xref ref-type="bibr" rid="ref-35">35</xref>]. <xref ref-type="fig" rid="fig-3">Fig. 3</xref> illustrates the structure of a <italic>GAN</italic>. The generator and discriminator architectures generally consist of multi-layer convolutional or fully connected layers. The generator learns the statistical properties of the training data and generates new images, while the discriminator evaluates and distinguishes between real and synthetic images [<xref ref-type="bibr" rid="ref-30">30</xref>]. Both networks serve as mappings between data domains [<xref ref-type="bibr" rid="ref-36">36</xref>]. The generator, without direct access to the real dataset, aims to create convincing synthetic images to deceive the discriminator. If the discriminator makes an incorrect classification, an error signal is generated to refine the generator&#x2019;s output, progressively enhancing the quality of generated images. The generator transforms a latent space into the data space, while the discriminator maps image data to a probability score, indicating whether an image is real or synthetic. If the discriminator identifies an image as real, it outputs a value close to <inline-formula id="ieqn-39"><mml:math id="mml-ieqn-39"><mml:mn>1</mml:mn></mml:math></inline-formula>, whereas for a synthetic image, it outputs a value near <inline-formula id="ieqn-40"><mml:math id="mml-ieqn-40"><mml:mn>0</mml:mn></mml:math></inline-formula>. <xref ref-type="fig" rid="fig-4">Fig. 4</xref> shows the training process of a <italic>GAN</italic>.</p>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>Block diagram of <italic>GAN</italic></title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-3.tif"/>
</fig><fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>Flowchart of <italic>GAN</italic> training process</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-4.tif"/>
</fig>
<p>Over the years, <inline-formula id="ieqn-42"><mml:math id="mml-ieqn-42"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> have undergone substantial development, starting from the original adversarial learning concept, although initial versions struggled with fully capturing the data distribution [<xref ref-type="bibr" rid="ref-42">42</xref>]. A comparison of some popular <italic>GAN</italic> models is mentioned in <xref ref-type="table" rid="table-1">Table 1</xref>. The original <italic>GAN</italic>, commonly referred to as the Vanilla <italic>GAN</italic> [<xref ref-type="bibr" rid="ref-37">37</xref>], utilizes random noise as input to the generator, which synthesizes photorealistic images. The discriminator differentiates between real and generated (fake) images. In the absence of ground truth images, the generator is trained solely using adversarial loss, while the discriminator is optimized using classification loss. However, the Vanilla <italic>GAN</italic> is limited in its ability to generate images across diverse classes effectively. Training Vanilla <inline-formula id="ieqn-43"><mml:math id="mml-ieqn-43"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> is inherently unstable and often results in generators yielding outputs that lack coherence structure. A set of architectural constraints was proposed to overcome the un-stability issue and evaluated for Convolutional <inline-formula id="ieqn-44"><mml:math id="mml-ieqn-44"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, termed Deep Convolutional <inline-formula id="ieqn-45"><mml:math id="mml-ieqn-45"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> (<italic>DCGAN</italic>) [<xref ref-type="bibr" rid="ref-38">38</xref>]. Trained discriminators demonstrate competitive performance in image classification tasks compared to other unsupervised methods. Visualization of learned filters reveals that specific filters specialize in generating distinct objects. Compared to Vanilla <italic>GAN</italic>, <italic>DCGAN</italic> replaces pooling layers with strided convolutions in the discriminator and fractional-strided convolutions in the generator. Additionally, it incorporates batch normalization in both the generator and discriminator. The emergence of <italic>DCGAN</italic> brought improvements in image fidelity [<xref ref-type="bibr" rid="ref-38">38</xref>], and <italic>WGAN</italic> later addressed challenges related to mode collapse and training instability [<xref ref-type="bibr" rid="ref-40">40</xref>]. Vanilla <italic>GAN</italic> and <italic>DCGAN</italic> exhibit several limitations, one of which is the inability to control the generated outputs. For example, while a <italic>GAN</italic> can train a generator to produce images of digits <inline-formula id="ieqn-46"><mml:math id="mml-ieqn-46"><mml:mo stretchy="false">(</mml:mo><mml:mn>0</mml:mn><mml:mrow><mml:mo>&#x2212;</mml:mo></mml:mrow><mml:mn>9</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> from random noise, practical applications often require generating a specific image. This limitation can be addressed by incorporating an additional input to guide the generation process. Previously, the generative model was <inline-formula id="ieqn-47"><mml:math id="mml-ieqn-47"><mml:msub><mml:mi>p</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. Now, it is designed to produce <inline-formula id="ieqn-48"><mml:math id="mml-ieqn-48"><mml:msub><mml:mi>p</mml:mi><mml:mi>g</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mi>c</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, where <italic>c</italic> is a conditional input used to control the generation process. This input <italic>c</italic> can be a string of codes representing the desired output or intent. <italic>GAN</italic> models often encounter challenges like mode collapse, which hinder their ability to provide meaningful learning curves that are crucial for debugging and hyper-parameter tuning. This issue can be addressed by using Wasserstein <italic>GAN</italic> <inline-formula id="ieqn-49"><mml:math id="mml-ieqn-49"><mml:mo stretchy="false">(</mml:mo><mml:mi>W</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-40">40</xref>], which incorporates the Wasserstein distance, also known as the Earth Mover&#x2019;s distance. In the <italic>WGAN</italic> architecture, the discriminator is replaced with a critic, which evaluates the Wasserstein distance rather than performing binary classification.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>Comparison of the advantages and limitations of specific <italic>GAN</italic> variants across different medical imaging tasks</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th></th>
<th><italic>Vanilla GAN</italic> [<xref ref-type="bibr" rid="ref-37">37</xref>]</th>
<th><italic>DCGAN</italic> [<xref ref-type="bibr" rid="ref-38">38</xref>]</th>
<th>cGAN [<xref ref-type="bibr" rid="ref-39">39</xref>]</th>
<th><italic>WGAN</italic> [<xref ref-type="bibr" rid="ref-40">40</xref>]</th>
<th><italic>CycleGAN</italic> [<xref ref-type="bibr" rid="ref-41">41</xref>]</th>
</tr>
</thead>
<tbody>
<tr>
<td><bold>Training stability</bold></td>
<td>Low-prone to mode collapse and convergence issues</td>
<td>Moderate-more stable due to architectural constraints (e.g., use of <inline-formula id="ieqn-41"><mml:math id="mml-ieqn-41"><mml:mi>C</mml:mi><mml:mi>N</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>)</td>
<td>Moderate-improved with conditioning but can still be unstable</td>
<td>High-uses Wasserstein loss for better convergence</td>
<td>High-designed for unpaired image to image translation</td>
</tr>
<tr>
<td><bold>Data requirement</bold></td>
<td>Moderate-requires substantial labeled data</td>
<td>Moderate-unlabeled data sufficient for many tasks</td>
<td>High-requires paired data for training</td>
<td>Moderate-flexible with labeled/ unlabeled data</td>
<td>Low-does not require paired data</td>
</tr>
<tr>
<td><bold>Performance metrics</bold></td>
<td>Variable-depends heavily on task and tuning</td>
<td>Good for image synthesis; less suitable for segmentation</td>
<td>High accuracy in segmentation and paired synthesis tasks</td>
<td>Superior generative quality and stable training</td>
<td>Excellent in unpaired synthesis (e.g., <italic>MR</italic> to <italic>CT</italic>), not suitable for segmentation</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><inline-formula id="ieqn-50"><mml:math id="mml-ieqn-50"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> generate images by sampling from a probability distribution, while <inline-formula id="ieqn-51"><mml:math id="mml-ieqn-51"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-41">41</xref>] perform image-to-image translation between two domains. <inline-formula id="ieqn-52"><mml:math id="mml-ieqn-52"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> establishes a mapping <inline-formula id="ieqn-53"><mml:math id="mml-ieqn-53"><mml:mi>G</mml:mi><mml:mo>:</mml:mo><mml:mi>X</mml:mi><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mi>Y</mml:mi></mml:math></inline-formula> such that the output <inline-formula id="ieqn-54"><mml:math id="mml-ieqn-54"><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mo>=</mml:mo><mml:mi>G</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo><mml:mspace width="thinmathspace" /><mml:mi>x</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mi>X</mml:mi></mml:math></inline-formula>, is indistinguishable from real images <inline-formula id="ieqn-55"><mml:math id="mml-ieqn-55"><mml:mi>y</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mi>Y</mml:mi></mml:math></inline-formula> by an adversary trained to differentiate <inline-formula id="ieqn-56"><mml:math id="mml-ieqn-56"><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula> from <inline-formula id="ieqn-57"><mml:math id="mml-ieqn-57"><mml:mi>y</mml:mi></mml:math></inline-formula>. Unlike conventional <inline-formula id="ieqn-58"><mml:math id="mml-ieqn-58"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, <inline-formula id="ieqn-59"><mml:math id="mml-ieqn-59"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> employs two generators and two discriminators to enable bidirectional translation between the domains. To enhance translation quality, <inline-formula id="ieqn-60"><mml:math id="mml-ieqn-60"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> incorporates cycle consistency loss in addition to adversarial loss, ensuring that a translated image can be mapped back to its original domain. <inline-formula id="ieqn-61"><mml:math id="mml-ieqn-61"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> enabled transformations between image domains without the need for paired datasets [<xref ref-type="bibr" rid="ref-41">41</xref>], and <italic>PGAN</italic> introduced a stepwise training method to generate increasingly detailed images [<xref ref-type="bibr" rid="ref-43">43</xref>]. Subsequent models like <italic>SAGAN</italic> concentrated on identifying important image areas and capturing long-range dependencies [<xref ref-type="bibr" rid="ref-41">41</xref>]. More recent innovations include <italic>RANDGAN</italic>, which improves segmentation for anomaly detection and outperforms earlier <italic>GAN</italic> frameworks in the medical context [<xref ref-type="bibr" rid="ref-44">44</xref>]. <italic>DGGAN</italic> focuses on generating anonymized brain vascular imagery using <italic>MRA</italic> patches [<xref ref-type="bibr" rid="ref-45">45</xref>], and <italic>ED</italic>&#x2212;<italic>GAN</italic> integrates <inline-formula id="ieqn-62"><mml:math id="mml-ieqn-62"><mml:mi>V</mml:mi><mml:mi>A</mml:mi><mml:mi>E</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> with <inline-formula id="ieqn-63"><mml:math id="mml-ieqn-63"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> to enhance generative performance [<xref ref-type="bibr" rid="ref-46">46</xref>]. These advancements collectively demonstrate the substantial progress of <inline-formula id="ieqn-64"><mml:math id="mml-ieqn-64"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, broadening their applicability in areas such as medical image synthesis, anomaly detection, and complex data representation.</p>
<p>With the variety of <italic>GAN</italic> models available for different applications, the base model or its specific variants can be selected based on the task. For instance, <italic>DCGAN</italic> is suitable for image generation, <italic>SRGAN</italic> for super-resolution, U-Net-based <inline-formula id="ieqn-65"><mml:math id="mml-ieqn-65"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> for segmentation, and <inline-formula id="ieqn-66"><mml:math id="mml-ieqn-66"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> for image translation.</p>
<p><inline-formula id="ieqn-67"><mml:math id="mml-ieqn-67"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-47">47</xref>&#x2013;<xref ref-type="bibr" rid="ref-50">50</xref>] have demonstrated their potential in various fields, including image generation, style transfer, and data augmentation. Their ability to generate high-fidelity synthetic images without extensive labeled datasets makes them valuable for medical imaging. The following are the applications of <inline-formula id="ieqn-68"><mml:math id="mml-ieqn-68"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-51">51</xref>] in medical imaging:
<list list-type="bullet">
<list-item>
<p>Image Denoising: Medical images, such as low-dose <italic>CT</italic> scans or accelerated <italic>MRI</italic>, often have lower quality due to noise or reduced resolution. <inline-formula id="ieqn-69"><mml:math id="mml-ieqn-69"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> enhance these images by reducing noise and increasing resolution [<xref ref-type="bibr" rid="ref-39">39</xref>,<xref ref-type="bibr" rid="ref-52">52</xref>&#x2013;<xref ref-type="bibr" rid="ref-54">54</xref>].</p></list-item>
<list-item>
<p>Image Segmentation: Segmentation is a crucial task in medical imaging, as it helps in determining and extracting areas of interest from medical images which has lately been well achieved using <inline-formula id="ieqn-70"><mml:math id="mml-ieqn-70"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-55">55</xref>&#x2013;<xref ref-type="bibr" rid="ref-57">57</xref>].</p></list-item>
<list-item>
<p>Image Super Resolution: Super-Resolution enhances the resolution thereby improving the clarity and detail of anatomical structures. <inline-formula id="ieqn-71"><mml:math id="mml-ieqn-71"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> consist of the generator and the discriminator that work in opposition to refine the super resolution process [<xref ref-type="bibr" rid="ref-58">58</xref>,<xref ref-type="bibr" rid="ref-59">59</xref>].</p></list-item>
<list-item>
<p>Image Translation: Translation involves converting images from one modality to another. <inline-formula id="ieqn-72"><mml:math id="mml-ieqn-72"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> have demonstrated great ability in medical image translation by generating realistic images while preserving anatomical accuracy [<xref ref-type="bibr" rid="ref-60">60</xref>&#x2013;<xref ref-type="bibr" rid="ref-63">63</xref>].</p></list-item>
<list-item>
<p>Image Reconstruction: <inline-formula id="ieqn-73"><mml:math id="mml-ieqn-73"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> enable the generation of medical images from limited data. For instance, they can generate <inline-formula id="ieqn-74"><mml:math id="mml-ieqn-74"><mml:mn>3</mml:mn><mml:mi>D</mml:mi></mml:math></inline-formula> images from <inline-formula id="ieqn-75"><mml:math id="mml-ieqn-75"><mml:mn>2</mml:mn><mml:mi>D</mml:mi></mml:math></inline-formula> slices or convert <italic>MR</italic> [<xref ref-type="bibr" rid="ref-64">64</xref>&#x2013;<xref ref-type="bibr" rid="ref-67">67</xref>] into <italic>CT</italic> images, facilitating multi-modality analysis.</p></list-item>
<list-item>
<p>Data Augmentation: <inline-formula id="ieqn-76"><mml:math id="mml-ieqn-76"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> can create synthetic medical images that capture realistic variations, helping to balance datasets and improve model training. This approach is particularly useful when dealing with rare diseases or underrepresented classes [<xref ref-type="bibr" rid="ref-68">68</xref>&#x2013;<xref ref-type="bibr" rid="ref-72">72</xref>].</p></list-item>
<list-item>
<p>Anomaly Detection: By learning the distribution of normal anatomical structures, <inline-formula id="ieqn-77"><mml:math id="mml-ieqn-77"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> can identify deviations from this norm, aiding in the detection of anomalies [<xref ref-type="bibr" rid="ref-73">73</xref>,<xref ref-type="bibr" rid="ref-74">74</xref>] and for early disease detection and screening [<xref ref-type="bibr" rid="ref-75">75</xref>&#x2013;<xref ref-type="bibr" rid="ref-78">78</xref>].</p></list-item>
<list-item>
<p>Domain Adaptation: Variations in medical images resulting from differences in imaging devices, acquisition protocols, or healthcare institutions can affect model performance. <inline-formula id="ieqn-78"><mml:math id="mml-ieqn-78"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> facilitate domain adaptation, allowing models trained on one dataset to generalize effectively to other datasets [<xref ref-type="bibr" rid="ref-79">79</xref>&#x2013;<xref ref-type="bibr" rid="ref-81">81</xref>].</p></list-item>
</list></p>
<p>The use of <inline-formula id="ieqn-79"><mml:math id="mml-ieqn-79"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical imaging represents a significant improvement in healthcare. <inline-formula id="ieqn-80"><mml:math id="mml-ieqn-80"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> offer distinct advantages in addressing the challenges of medical imaging, primarily due to their capacity to produce highly realistic images, even with limited supervision which is in contrast to conventional models that focus solely on classification or detection tasks. This is important in the medical domain, where labeled datasets are often scarce, imbalanced, or costly to acquire. In addition, <inline-formula id="ieqn-81"><mml:math id="mml-ieqn-81"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> are effective in tasks that require one form of medical image to be transformed into another, such as improving the resolution of scans, converting between imaging modalities. These applications are vital for accurate clinical assessment.</p>
<p>Ongoing research and technological progress continue to expand the scope and impact of <inline-formula id="ieqn-82"><mml:math id="mml-ieqn-82"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in healthcare [<xref ref-type="bibr" rid="ref-82">82</xref>]. For instance, <inline-formula id="ieqn-83"><mml:math id="mml-ieqn-83"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> is employed for domain adaptation, facilitating the translation of images across different modalities [<xref ref-type="bibr" rid="ref-41">41</xref>], while pix2pix supports tasks like resolution improvement and denoising through image-to-image translation [<xref ref-type="bibr" rid="ref-83">83</xref>]. A further example is <italic>UNITGAN</italic>, which allows cross-modal image integration by establishing a shared latent space between modalities. In contrast, <inline-formula id="ieqn-84"><mml:math id="mml-ieqn-84"><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>o</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> contributes to the generation of high-resolution images [<xref ref-type="bibr" rid="ref-43">43</xref>]. These <italic>GAN</italic> models have played a key role in enhancing both the diversity and quality of medical images, which in turn supports improved healthcare analysis and outcomes. <xref ref-type="fig" rid="fig-5">Fig. 5</xref> summarizes the application of <italic>GAN</italic> in medical imaging.</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>Applications of <italic>GAN</italic> in medical imaging</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-5.tif"/>
</fig>
<p>Despite their potential, <inline-formula id="ieqn-85"><mml:math id="mml-ieqn-85"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> present certain challenges. Training <inline-formula id="ieqn-86"><mml:math id="mml-ieqn-86"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> can be computationally demanding, and achieving stable training requires careful model design and parameter tuning. The interpretability of <italic>GAN</italic>-based models is another concern, particularly in high-stakes medical applications where model transparency is essential. Ethical considerations are also crucial, especially regarding patient privacy and the potential misuse of synthetic medical images.</p>
</sec>
<sec id="s1_2">
<label>1.2</label>
<title>Methodology</title>
<p>This research adopts a Systematic Literature Review <inline-formula id="ieqn-87"><mml:math id="mml-ieqn-87"><mml:mo stretchy="false">(</mml:mo><mml:mi>S</mml:mi><mml:mi>L</mml:mi><mml:mi>R</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> approach, guided by the methodologies introduced by Kitchenham et al. [<xref ref-type="bibr" rid="ref-84">84</xref>]. The study design follows the structured process outlined in the <italic>PRISMA</italic> (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) framework. This ensures a clear, consistent, and reproducible process for gathering and analyzing data. All key literature included in this review was selected in accordance with <italic>PRISMA's</italic> widely accepted standards for conducting systematic reviews. The following steps were included in the search strategy.
<list list-type="simple">
<list-item><label>1.</label><p><bold>Databases Searched:</bold> The review targets scholarly, peer-reviewed articles sourced from well-established databases, including ScienceDirect, SpringerLink, the <italic>ACM</italic> Digital Library, <italic>IEEE</italic> Xplore, PubMed, Scopus and Web of Science. At the outset of this study, a total of <inline-formula id="ieqn-88"><mml:math id="mml-ieqn-88"><mml:mn>500</mml:mn></mml:math></inline-formula> articles were initially selected for review. The articles were sourced from literature published between <inline-formula id="ieqn-89"> <mml:math id="mml-ieqn-89"><mml:mn>2020</mml:mn></mml:math></inline-formula> and <inline-formula id="ieqn-90"><mml:math id="mml-ieqn-90"><mml:mn>2025</mml:mn></mml:math></inline-formula>, with the final selection process carried out between <inline-formula id="ieqn-91"><mml:math id="mml-ieqn-91"><mml:mi>N</mml:mi><mml:mi>o</mml:mi><mml:mi>v</mml:mi><mml:mi>e</mml:mi><mml:mi>m</mml:mi><mml:mi>b</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mn>2024</mml:mn></mml:math></inline-formula> and <inline-formula id="ieqn-92"><mml:math id="mml-ieqn-92"><mml:mi>J</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>u</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi><mml:mi>y</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mn>2025</mml:mn></mml:math></inline-formula>. Following a thorough evaluation, <inline-formula id="ieqn-93"><mml:math id="mml-ieqn-93"><mml:mn>167</mml:mn></mml:math></inline-formula> articles were shortlisted based on their relevance to applications, challenges, and recent advancements. The deliberate focus on recently published literature reflects our commitment to providing a forward-looking and state-of-the-art analysis.</p></list-item>
<list-item><label>2.</label><p><bold>Search Terms and Keyword Strategy:</bold> To capture the full range of applications of <inline-formula id="ieqn-94"><mml:math id="mml-ieqn-94"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical imaging, we developed a structured list of keywords representing both the core concept (<inline-formula id="ieqn-95"><mml:math id="mml-ieqn-95"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>) and specific application domains in medical imaging. <xref ref-type="table" rid="table-2">Table 2</xref> indicates the keywords used for selecting the data and the number of selected papers under each category.</p>
</list-item>
<list-item><label>3.</label><p><bold>Boolean Operators and Search String Construction:</bold> Boolean operators were applied to systematically combine the two groups of keywords. The operator <italic>OR</italic> was used within each group to capture synonyms and variations, while the operator <italic>AND</italic> was used to combine the technology-related terms with the application-related terms. The following is an example of the search string used:</p>
<p>(&#x201C;Generative Adversarial Networks&#x201D; or &#x201C;<italic>GAN</italic>&#x201D; or &#x201C;<inline-formula id="ieqn-96"><mml:math id="mml-ieqn-96"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>&#x201D;) <italic>AND</italic> (&#x201C;Medical Imaging&#x201D; or &#x201C;Medical Image Denoising&#x201D; or &#x201C;Medical Image Super Resolution&#x201D; or &#x201C;Medical Image Segmentation&#x201D; or &#x201C;Medical Image Translation&#x201D; or &#x201C;Medical Image Reconstruction&#x201D; or &#x201C;Medical Data Augmentation&#x201D;).</p>
<p>This logic ensures that the search retrieves studies that discuss <inline-formula id="ieqn-97"><mml:math id="mml-ieqn-97"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in any of the specified application domains within medical imaging.</p></list-item>
<list-item><label>4.</label><p><bold>Search Execution and Documentation:</bold> The search was performed using both keyword and <inline-formula id="ieqn-98"><mml:math id="mml-ieqn-98"><mml:mi>M</mml:mi><mml:mi>e</mml:mi><mml:mi>S</mml:mi><mml:mi>H</mml:mi></mml:math></inline-formula> term combinations (where applicable, such as in PubMed). Filters were applied to include only peer-reviewed journal articles and conference papers published in English. Search results were exported to a reference management tool (e.g., Zotero or EndNote), and duplicates were removed prior to screening.</p></list-item>
<list-item><label>5.</label><p><bold>Screening Process:</bold> Following retrieval, titles and abstracts were screened independently by two reviewers. Full texts were then assessed for eligibility based on predefined inclusion and exclusion criteria. <xref ref-type="table" rid="table-3">Table 3</xref> indicates the inclusion and exclusion criteria used for selecting the articles.</p>
</list-item>
<list-item><label>6.</label><p><bold>Quality Assessment Method:</bold> To assess the methodological quality and reporting transparency of included studies, the <italic>CLAIM</italic> (Checklist for Artificial Intelligence in Medical Imaging) guideline was used. This checklist evaluates critical domains relevant to <italic>AI</italic> studies, including dataset characteristics, model evaluation procedures, validation methodology, and reproducibility. <xref ref-type="table" rid="table-4">Table 4</xref> adapted from the <italic>CLAIM</italic>, indicates the criterion used to evaluate the methodological quality of included studies. Only studies meeting key <italic>CLAIM</italic> criteria were retained for final synthesis to ensure reliability of the review findings. Each study was assessed independently by two reviewers. Disagreements were resolved through discussion or by consulting a third reviewer.</p>
</list-item>
</list><table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>Keywords used for article selection</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Keyword</th>
<th>Paper count</th>
</tr>
</thead>
<tbody>
<tr>
<td><inline-formula id="ieqn-99"><mml:math id="mml-ieqn-99"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical imaging</td>
<td>145</td>
</tr>
<tr>
<td><inline-formula id="ieqn-100"><mml:math id="mml-ieqn-100"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical image denoising</td>
<td>25</td>
</tr>
<tr>
<td><inline-formula id="ieqn-101"><mml:math id="mml-ieqn-101"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical image super resolution</td>
<td>27</td>
</tr>
<tr>
<td><inline-formula id="ieqn-102"><mml:math id="mml-ieqn-102"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical image segmentation</td>
<td>23</td>
</tr>
<tr>
<td><inline-formula id="ieqn-103"><mml:math id="mml-ieqn-103"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical image translation</td>
<td>28</td>
</tr>
<tr>
<td><inline-formula id="ieqn-104"><mml:math id="mml-ieqn-104"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical image reconstruction</td>
<td>30</td>
</tr>
<tr>
<td><inline-formula id="ieqn-105"><mml:math id="mml-ieqn-105"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical data augmentation</td>
<td>10</td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>Inclusion and exclusion criteria for selected articles</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th></th>
<th>Inclusion criteria</th>
<th>Exclusion criteria</th>
</tr>
</thead>
<tbody>
<tr>
<td>Types of study</td>
<td>Original and review article</td>
<td>Thesis, white papers, communication letters and editorials</td>
</tr>
<tr>
<td>Language</td>
<td>Research published in English language</td>
<td>Duplicate and non-English articles</td>
</tr>
<tr>
<td>Publication Year (For result analysis and application)</td>
<td>Articles published in <inline-formula id="ieqn-106"><mml:math id="mml-ieqn-106"><mml:mn>2020</mml:mn><mml:mrow><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /></mml:mrow><mml:mn>2025</mml:mn></mml:math></inline-formula></td>
<td>&#x2013;</td>
</tr>
<tr>
<td>Source</td>
<td>Publications in peer reviewed journals and conferences</td>
<td>&#x2013;</td>
</tr>
<tr>
<td>Settings</td>
<td><inline-formula id="ieqn-107"><mml:math id="mml-ieqn-107"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical imaging</td>
<td>Not related to medical imaging</td>
</tr>
<tr>
<td>Intervention</td>
<td><italic>ML</italic> and <italic>GAN</italic></td>
<td>Traditional methods</td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-4">
<label>Table 4</label>
<caption>
<title><italic>CLAIM</italic> based quality assessment criteria applied to included studies</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th>S. No.</th>
<th><italic>CLAIM</italic> Domain</th>
<th>Assessment Criteria</th>
</tr>
</thead>
<tbody>
<tr>
<td>1</td>
<td>Study design &#x0026; objective</td>
<td>Clear research objective; rationale for using GANs clearly stated</td>
</tr>
<tr>
<td>2</td>
<td>Dataset description</td>
<td>Data source and acquisition protocols described; dataset size, diversity, and labeling provided</td>
</tr>
<tr>
<td>3</td>
<td>Data splitting &#x0026; Preprocessing</td>
<td>Clear description of training/validation/test splits; preprocessing steps explained</td>
</tr>
<tr>
<td>4</td>
<td>Model architecture</td>
<td><italic>GAN</italic> architecture detailed (e.g., type of <italic>GAN</italic>, loss functions, training parameters)</td>
</tr>
<tr>
<td>5</td>
<td>Evaluation metrics</td>
<td>Appropriate and consistent metrics used (e.g., <italic>PSNR</italic>, <italic>SSIM</italic>, <inline-formula id="ieqn-108"><mml:math id="mml-ieqn-108"><mml:mi>D</mml:mi><mml:mi>i</mml:mi><mml:mi>c</mml:mi><mml:mi>e</mml:mi></mml:math></inline-formula>, <italic>AUC</italic>)</td>
</tr>
<tr>
<td>6</td>
<td>Validation strategy</td>
<td>Cross-validation or external validation employed to avoid overfitting</td>
</tr>
<tr>
<td>7</td>
<td>Ground truth &#x0026; annotation</td>
<td>Reference standards and annotation protocols clearly described</td>
</tr>
<tr>
<td>8</td>
<td>Reproducibility &#x0026; code availability</td>
<td>Whether code, models, or datasets were publicly shared</td>
</tr>
<tr>
<td>9</td>
<td>Clinical relevance &#x0026; interpretation</td>
<td>Results discussed in the context of clinical utility, risks, or limitations</td>
</tr>
<tr>
<td>10</td>
<td>Bias and limitations disclosure</td>
<td>Potential sources of bias and study limitations explicitly acknowledged</td>
</tr>
</tbody>
</table>
</table-wrap></p>
<p>This approach ensured a comprehensive, reproducible, and methodologically sound search process in alignment with <italic>PRISMA</italic> standards. The <italic>PRISMA</italic> flow diagram of the article selection procedure is shown in <xref ref-type="fig" rid="fig-6">Fig. 6</xref>. <italic>PRISMA</italic> checklists can be found in the supplementary files.</p>
<fig id="fig-6">
<label>Figure 6</label>
<caption>
<title><italic>PRISMA</italic> flow diagram of the article selection procedure</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-6.tif"/>
</fig>
<p>The remaining sections of this paper explore in detail how <inline-formula id="ieqn-109"><mml:math id="mml-ieqn-109"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> are applied within the field of medical imaging. <xref ref-type="sec" rid="s2">Section 2</xref> discusses the various applications of <inline-formula id="ieqn-110"><mml:math id="mml-ieqn-110"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical images. <xref ref-type="sec" rid="s4">Section 4</xref> details the different challenges of <inline-formula id="ieqn-111"><mml:math id="mml-ieqn-111"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> for clinical use under regulatory frameworks. <xref ref-type="sec" rid="s5">Section 5</xref> gives an insight to the other genrative models present. Finally, the <xref ref-type="sec" rid="s6">Section 6</xref> concludes the paper providing an insight into the limitations and future of <inline-formula id="ieqn-112"><mml:math id="mml-ieqn-112"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in the medical field.</p>
</sec>
</sec>
<sec id="s2">
<label>2</label>
<title>Medical Applications</title>
<p>Medical imaging application uses <italic>GAN</italic> in two separate ways; one as generator which examines the underlying data distribution and generates new (synthetic) images. The discriminator section can classify normal and abnormal images. An overview of usage of <inline-formula id="ieqn-113"><mml:math id="mml-ieqn-113"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in various medical applications is presented in <xref ref-type="table" rid="table-5">Tables 5</xref>&#x2013;<xref ref-type="table" rid="table-7">7</xref>. The following subsection describes the medical imaging applications of <italic>GAN</italic> for denoising, segmentation, super resolution, translation, reconstruction and data augmentation.</p>
<table-wrap id="table-5">
<label>Table 5</label>
<caption>
<title>Summary of methods &#x0026; applications (Arranged by Year)</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th align="center">S. No.</th>
<th>Name</th>
<th>Year</th>
<th>Purpose</th>
<th>Dataset</th>
<th>Metrics</th>
<th>Remark</th>
</tr>
</thead>
<tbody>
<tr>
<td>1.</td>
<td>Ali et al. [<xref ref-type="bibr" rid="ref-85">85</xref>]</td>
<td>2024</td>
<td>Medical image processing</td>
<td>&#x2013;</td>
<td></td>
<td>Overview of <italic>GAN</italic> for medical image processing.</td>
</tr>
<tr>
<td>2.</td>
<td>Tu et al. [<xref ref-type="bibr" rid="ref-86">86</xref>]</td>
<td>2024</td>
<td>Image translation</td>
<td>summer2winter, facade dataset</td>
<td>FID and EGF</td>
<td>Introduces the unpaired image-to-image translation method using a diffusion adversarial network (UNDAN). The method captures seasonal features well and enhances clarity but occasionally misinterprets details or leaves inconsistencies.</td>
</tr>
<tr>
<td>3.</td>
<td>Fan et al. [<xref ref-type="bibr" rid="ref-66">66</xref>]</td>
<td>2023</td>
<td>Image fusion</td>
<td>AANLIB</td>
<td>SSIM, RMSE, MI and Gradient Evaluation</td>
<td>U-Patch GAN, built on a GAN framework to achieve self-supervised fusion of multimodal brain images, aiming to improve the quality of the fusion process. Effectively fuses various modal images while preserving rich texture and functional information, however, but there is potential to develop more effective mechanisms to improve the generator structure.</td>
</tr>
<tr>
<td>4.</td>
<td>Wang et al. [<xref ref-type="bibr" rid="ref-87">87</xref>]</td>
<td>2023</td>
<td>Image translation</td>
<td>Cityscapes, Facades, CUHK Face Sketch, Day-Night, and Oil-Chinese Paintings</td>
<td>FCN-score, semantic segmentation metrics, FID, and LPIPS</td>
<td>Introduces an unsupervised approach to image-to-image translation using long-short cycle-consistent adversarial networks. Enhanced image quality by reducing information loss, lacks from computational overhead without proportional performance gains.</td>
</tr>
<tr>
<td>5.</td>
<td>Zhou et al. [<xref ref-type="bibr" rid="ref-48">48</xref>]</td>
<td>2023</td>
<td>Medical image fusion</td>
<td>&#x2013;</td>
<td>&#x2013;</td>
<td>Survey of Generative Adversarial Networks (GANs).</td>
</tr>
<tr>
<td>6.</td>
<td>Esmaeili et al. [<xref ref-type="bibr" rid="ref-75">75</xref>]</td>
<td>2023</td>
<td>Anomaly detection</td>
<td>Head-Ct, Brain Tumor MRI, Br35h-MRI, MIAS-MAMMO, MIAS-Patches MAMMO, C-NMC-Leukemia, and Retinal OCT</td>
<td>Precision, Recall, Specificity, F1-Score, ROC AUC, and PR AUC</td>
<td>Overview of the application of GANs in Alzheimer&#x2019;s Disease (AD), along with an exploration of cutting-edge GAN-based methods for AD in biomedical imaging. Examines the outcomes from both data and model perspectives reveals key limitations of existing techniques, but struggled to detect breast cancer lesions due to their complexity and similarity to normal tissue.</td>
</tr>
<tr>
<td>7.</td>
<td>Fu et al. [<xref ref-type="bibr" rid="ref-52">52</xref>]</td>
<td>2022</td>
<td>Medical image denoising</td>
<td></td>
<td></td>
<td>Denoising low-dose CT images (LDCT) through a two-phase deep learning approach known as the Noisy Generation-Removal Network (NGRNet).</td>
</tr>
<tr>
<td>8.</td>
<td>Zhang et al. [<xref ref-type="bibr" rid="ref-78">78</xref>]</td>
<td>2022</td>
<td>Anomaly detection</td>
<td>CIFAR-10, ILD, HAM 10000</td>
<td>AUC, Accuracy, Precision, Recall and F1-Score</td>
<td>An unsupervised approach for deep anomaly detection using an enhanced adversarial auto-encoder. The unsupervised anomaly detection method using an improved adversarial auto-encoder with convolution block chains (CCB) replacing skip connections, offering more stable training, but struggles with anomaly detection involving temporal variations.</td>
</tr>
<tr>
<td>9.</td>
<td>Sun et al. [<xref ref-type="bibr" rid="ref-88">88</xref>]</td>
<td>2022</td>
<td>Medical image synthesis</td>
<td>COPDGene, GSP</td>
<td>FID, MMD and IS</td>
<td>Introduce a new end-to-end GAN architecture designed to generate high-resolution 3D images. HA-GAN serves as plug and play module for all existing GANs; but higher memory utilization with increase in image resolutions.</td>
</tr>
<tr>
<td>10.</td>
<td>Xia et al. [<xref ref-type="bibr" rid="ref-74">74</xref>]</td>
<td>2022</td>
<td>Anomaly detection</td>
<td>&#x2013;</td>
<td>&#x2013;</td>
<td>Survey of GAN based anomaly detection methods.</td>
</tr>
<tr>
<td>11.</td>
<td>Xun et al. [<xref ref-type="bibr" rid="ref-89">89</xref>]</td>
<td>2022</td>
<td>Medical image synthesis</td>
<td>&#x2013;</td>
<td>&#x2013;</td>
<td>Presents the origins, operating principles, and extended versions of GAN, along with a review of the most recent advancements in GAN-based methods for medical image segmentation.</td>
</tr>
<tr>
<td>12.</td>
<td>Singh et al. [<xref ref-type="bibr" rid="ref-65">65</xref>]</td>
<td>2021</td>
<td>Medical image generation</td>
<td>&#x2013;</td>
<td>&#x2013;</td>
<td>Survey of GAN based medical image generation methods.</td>
</tr>
<tr>
<td>13.</td>
<td>Suganthi et al. [<xref ref-type="bibr" rid="ref-47">47</xref>]</td>
<td>2021</td>
<td>Medical image synthesis</td>
<td>&#x2013;</td>
<td>&#x2013;</td>
<td>Survey of medical image synthesis using GAN.</td>
</tr>
<tr>
<td>14.</td>
<td>Ma et al. [<xref ref-type="bibr" rid="ref-49">49</xref>]</td>
<td>2021</td>
<td>Medical image enhancement</td>
<td>CORN-2, Fundus Multi-Disease Diagnosis, EASE</td>
<td>SNR, AUC, Accuracy, Sensitivity, G-mean, Dice</td>
<td>Two novel constraints, illumination regularization to enhance illumination uniformity and structure loss to preserve structural details, are incorporated into the objective function to improve clinical interpretability and subsequent analysis. StillGAN, carries the risk of incorrect translations, such as altering colors or introducing lesion-like artifacts, during image enhancement.</td>
</tr>
<tr>
<td>15.</td>
<td>Li et al. [<xref ref-type="bibr" rid="ref-53">53</xref>]</td>
<td>2021</td>
<td>Medical image denoising</td>
<td>JSRT LIDC-IDRI or LIDC</td>
<td>PSNR, SSIM</td>
<td>A medical image denoising technique utilizing a conditional generative adversarial network (CGAN) to address various types of unknown noise. The method handles spatially varying and unknown noises, using noise and gradient images as conditional inputs to enhance structural specificity and preserve image details. Selecting an optimal threshold to accurately derive gradient information poses a significant challenge due to the variability in image intensity and structural patterns.</td>
</tr>
<tr>
<td>16.</td>
<td>Goodfellow et al. [<xref ref-type="bibr" rid="ref-30">30</xref>]</td>
<td>2020</td>
<td>Overview of GAN</td>
<td>&#x2013;</td>
<td>&#x2013;</td>
<td>Overview of generative adversarial networks.</td>
</tr>
<tr>
<td>17.</td>
<td>Xie et al. constraint [<xref ref-type="bibr" rid="ref-80">80</xref>]</td>
<td>2020</td>
<td>Domain adaptation</td>
<td>Colonoscopic, REFUGE</td>
<td>Dice score</td>
<td>A GAN, referred to as MIGAN, is employed to preserve image content during cross-domain image-to-image (I2I) translation. The model preserves image content by disentangling content features from domain information and maximizing mutual information between the content features of source and translated images.</td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-6">
<label>Table 6</label>
<caption>
<title>Summary of methods &#x0026; applications (Arranged by Year) (Contd.)</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th align="center">S. No.</th>
<th>Name</th>
<th>Year</th>
<th>Purpose</th>
<th>Dataset</th>
<th>Metrics</th>
<th>Remark</th>
</tr>
</thead>
<tbody>
<tr>
<td>18.</td>
<td>Sun et al. [<xref ref-type="bibr" rid="ref-71">71</xref>]</td>
<td>2020</td>
<td>Medical image segmentation</td>
<td>BRATS17, LIVER100</td>
<td>Dice Score</td>
<td>MM-GAN, simulates the distribution of real data and generates new samples from the limited data distribution to enrich the training set. This model can translate label maps into 3D MR images without the concern of violating the underlying pathology. MM-GAN enables the translation of label maps into 3D MR images while ensuring the integrity of pathological features.</td>
</tr>
<tr>
<td>19.</td>
<td>Kazeminia et al. [<xref ref-type="bibr" rid="ref-51">51</xref>]</td>
<td>2020</td>
<td>Medical image analysis</td>
<td>&#x2013;</td>
<td>&#x2013;</td>
<td>Overview of <inline-formula id="ieqn-114"><mml:math id="mml-ieqn-114"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> for medical image analysis.</td>
</tr>
<tr>
<td>20.</td>
<td>Armanious et al. [<xref ref-type="bibr" rid="ref-90">90</xref>]</td>
<td>2020</td>
<td>Image translation</td>
<td>Anonymized datasets</td>
<td>SSIM, PSNR, MSE, VIF, UQI, LPIPS</td>
<td>MedGAN is utilized for medical image-to-image translation, functioning at the image level in a fully end-to-end manner. Model integrates a conditional adversarial framework with novel non-adversarial losses and a CasNET generator architecture to improve global consistency and preserve high-frequency details in the results, but is unable to process 3D medical volumes effectively.</td>
</tr>
<tr>
<td>21.</td>
<td>Kim et al. [<xref ref-type="bibr" rid="ref-39">39</xref>]</td>
<td>2020</td>
<td>Medical imaging</td>
<td>LIDC&#x2013;IDRI, AAPM</td>
<td>SSIM</td>
<td>Conditional Generative Adversarial Networks (<italic>CGAN</italic>) are applied to low-dose chest imaging, a widely used diagnostic tool in medical imaging. The technique offers universal applicability to various medical images without complex parameter tuning, image information loss, or high modality dependence, while enabling rapid inference through optimized convolutional operations. Using batch normalization in the generator can hinder the production of diverse and realistic images.</td>
</tr>
<tr>
<td>22.</td>
<td>Zhang et al. [<xref ref-type="bibr" rid="ref-91">91</xref>]</td>
<td>2019</td>
<td>Medical image denoising</td>
<td>4D Extended Cardiac Torso (XCAT) phantom</td>
<td>Qualitative Analysis</td>
<td><italic>SPECT</italic> Image Denoising: A study assessing the application of Generative Adversarial Networks (<italic>GAN</italic>) for denoising static SPECT images. The proposed method can reduce SPECT image noise, enabling lower injection doses or shorter acquisition times while preserving image quality suitable for clinical diagnosis. The technique addresses only a single problem in low-level medical image vision, resulting in limited clinical applicability and significant residual noise after SR processing.</td>
</tr>
<tr>
<td>23.</td>
<td>Pan et al. [<xref ref-type="bibr" rid="ref-23">23</xref>]</td>
<td>2019</td>
<td>Survey on <italic>GAN</italic></td>
<td>&#x2013;</td>
<td>&#x2013;</td>
<td>Overview on Generative Adversarial Networks (<inline-formula id="ieqn-115"><mml:math id="mml-ieqn-115"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>).</td>
</tr>
<tr>
<td>24.</td>
<td>Dar et al. [<xref ref-type="bibr" rid="ref-61">61</xref>]</td>
<td>2019</td>
<td><italic>MRI</italic> Synthesis</td>
<td>MIDAS, IXI</td>
<td>SSIM, PSNR</td>
<td><italic>MRI</italic> synthesis using a conditional generative adversarial network. The model synthesize distinct contrasts from a single modality, with applications in multi-contrast brain MRI and improved accuracy using cross-sectional correlations. The model have the disadvantage of multi-modal imaging for large cohorts due to its high economic and time costs.</td>
</tr>
<tr>
<td>25.</td>
<td>Mahapatra et al. [<xref ref-type="bibr" rid="ref-92">92</xref>]</td>
<td>2019</td>
<td>Medical image analysis</td>
<td>EYEPACS</td>
<td>SSIM, PSNR, S3, P</td>
<td>Progressive Generative Adversarial Networks (<inline-formula id="ieqn-116"><mml:math id="mml-ieqn-116"><mml:mi>P</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>) based super resolution technique. The superior performance of the proposed method is evident in multiple medical image, modalities like retinal vessel segmentation, micro aneurysm detection, MRI super-resolution, and cardiac LV segmentation. It struggles with minor anatomy and pathologies, limiting early disease detection.</td>
</tr>
<tr>
<td>26.</td>
<td>You et al. [<xref ref-type="bibr" rid="ref-93">93</xref>]</td>
<td>2019</td>
<td>Image reconstruction</td>
<td>Tibia, Abdominal</td>
<td>PSNR, SSIM, IFC</td>
<td>Deep learning method designed to effectively reconstruct <italic>HRCT</italic> images from their <italic>LRCT</italic> images. Achieved promising results in preserving anatomical details and reducing image noise using both supervised and semi-supervised learning approaches. The model requires significantly longer training than standard GAN methods (1&#x2013;2 days) and struggles to recover subtle features.</td>
</tr>
<tr>
<td>27.</td>
<td>Creswell et al. [<xref ref-type="bibr" rid="ref-36">36</xref>]</td>
<td>2018</td>
<td>Study of <italic>GAN</italic></td>
<td>&#x2013;</td>
<td>&#x2013;</td>
<td>Survey of generative adversarial networks.</td>
</tr>
<tr>
<td>28.</td>
<td>Yang et al. [<xref ref-type="bibr" rid="ref-94">94</xref>]</td>
<td>2018</td>
<td>Medical image denoising</td>
<td>2016 NIH-AAPM Mayo Clinic Low Dose CT Grand Challenge</td>
<td>PSNR, SSIM</td>
<td>A <italic>CT</italic> image denoising approach utilizing a Generative Adversarial Network (<italic>GAN</italic>) with Wasserstein distance and perceptual similarity. The WGAN framework reduces over-smoothing, yielding images with higher PSNR and more faithful statistical properties. The model risks losing critical features and requires re-tuning for datasets with varying noise properties.</td>
</tr>
<tr>
<td>29.</td>
<td>Xue et al. [<xref ref-type="bibr" rid="ref-56">56</xref>]</td>
<td>2018</td>
<td>Medical image segmentation</td>
<td>MICCAI BRATS</td>
<td>Dice score</td>
<td>An end-to-end adversarial neural network, known as <inline-formula id="ieqn-117"><mml:math id="mml-ieqn-117"><mml:mi>S</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>, designed for medical image segmentation tasks. The multi-scale loss function integrates CNN feature differences across multiple layers, enabling more effective training compared to classic GANs with separate losses.</td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-7">
<label>Table 7</label>
<caption>
<title>Summary of methods &#x0026; applications (Arranged by Year) (Contd.)</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th align="center">S. No.</th>
<th>Name</th>
<th>Year</th>
<th>Purpose</th>
<th>Dataset</th>
<th>Metrics</th>
<th>Remark</th>
</tr>
</thead>
<tbody>
<tr>
<td>30.</td>
<td>Han et al. [<xref ref-type="bibr" rid="ref-57">57</xref>]</td>
<td>2018</td>
<td>Medical image segmentation</td>
<td>Private dataset</td>
<td>Average pixel-level accuracy, Dice coefficient</td>
<td>Spine-<italic>GAN</italic>, a Recurrent Generative Adversarial Network, is employed for the segmentation of various spinal structures. The model integrates atrous convolution and autoencoder for fine-grained semantic representation, employs long short-term memory module (LSTM) to model spatial pathological relationships, and uses an auxiliary CNN for error correction. This model is limited to segmenting and classifying spinal structures, falling short of enabling human-like understanding of MR images.</td>
</tr>
<tr>
<td>31.</td>
<td>Li et al. [<xref ref-type="bibr" rid="ref-95">95</xref>]</td>
<td>2018</td>
<td>Medical image segmentation</td>
<td>I3A, MIVIA</td>
<td>Precision, Recall, SEG</td>
<td>Generative adversarial networks (<inline-formula id="ieqn-118"><mml:math id="mml-ieqn-118"><mml:mi>c</mml:mi><mml:mi>C</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>) are utilized for the robust segmentation of various HEp-2 datasets. The framework leverages auxiliary classifiers to tackle HEp-2 cell segmentation, reducing over-fitting and enhancing transfer capacity in deep learning networks.</td>
</tr>
<tr>
<td>32.</td>
<td>Sanchez et al. [<xref ref-type="bibr" rid="ref-59">59</xref>]</td>
<td>2018</td>
<td><italic>MR</italic> image generation</td>
<td>ADNI</td>
<td>SSIM, PSNR</td>
<td>An adversarial learning method for generating high-resolution MRI scans from low-resolution images. The model incorporate adversarial loss for realism and content loss to minimize differences between real and generated images, achieving promising results in high-factor downsampling for 3D medical imaging super-resolution. A limitation is the need for improved architectures and the integration of additional perceptual terms in the loss function to enhance the quality of the generated volumes.</td>
</tr>
<tr>
<td>33.</td>
<td>Hu et al. [<xref ref-type="bibr" rid="ref-96">96</xref>]</td>
<td>2018</td>
<td>Medical image synthesis</td>
<td>ICPR</td>
<td>Precision, Recall, F1-score</td>
<td><italic>GAN</italic> based method with a novel loss formulation to enable robust cell-level visual representation learning in an unsupervised environment. The approach relies on a novel pipeline combining cell-level visual representation learning and nuclei segmentation to distinguish cellular elements. Incorporating the gradient penalty into the network architecture necessitates computing second-order derivatives, which significantly increases training time.</td>
</tr>
<tr>
<td>34.</td>
<td>Nie et al. [<xref ref-type="bibr" rid="ref-97">97</xref>]</td>
<td>2018</td>
<td>Medical image synthesis</td>
<td>ADNI, Pelvic</td>
<td>PSNR, MAE, Dice Index</td>
<td>An adversarial learning strategy applied to improve the modelling of the <italic>FCN</italic>. The supervised deep convolutional model is using adversarial learning to estimate target images from source images, even across modalities, with an image gradient difference loss to reduce blurriness in generated outputs. A limitation is the dependency on paired source-target images from a large cohort, as acquiring multi-modal imaging at this scale is often constrained by significant economic and time costs.</td>
</tr>
<tr>
<td>35.</td>
<td>Bissoto et al. [<xref ref-type="bibr" rid="ref-98">98</xref>]</td>
<td>2018</td>
<td>Skin lesion synthesis</td>
<td>2018 ISIC Challenge</td>
<td>AUC</td>
<td>Focuses on skin lesion synthesis with GAN. A visual comparison showcasing high-resolution samples with fine-grained details, including accurately placed and sharply defined malignancy markers, resulting in visually compelling images. The model requires annotated data for image generation.</td>
</tr>
<tr>
<td>36.</td>
<td>Han et al. [<xref ref-type="bibr" rid="ref-64">64</xref>]</td>
<td>2018</td>
<td><italic>MR</italic> image generation</td>
<td>BRATS 2016</td>
<td>Visual turing test, accuracy</td>
<td>Generates synthetic multi-sequence brain Magnetic Resonance (<italic>MR</italic>) images through the use of Generative Adversarial Networks (<inline-formula id="ieqn-119"><mml:math id="mml-ieqn-119"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>). This research demonstrates the potential of WGAN to generate realistic multi-sequence brain MR images, enabling applications like data augmentation and enhanced physician training. A limitation of this work is its reliance solely on sagittal MR images, with coronal and transverse images yet to be generated.</td>
</tr>
<tr>
<td>37.</td>
<td>Udrea et al. [<xref ref-type="bibr" rid="ref-99">99</xref>]</td>
<td>2017</td>
<td>Skin lesion detection</td>
<td>Skin lesion dataset</td>
<td>Accuracy</td>
<td>Accuracy an artificial intelligence approach for segmenting pigmented and non-pigmented lesions using <italic>GAN</italic>. The ability to achieve these results using smartphone-acquired images from laypersons highlights its value for mHealth and eHealth dermatological applications. The approach requires improved GAN architecture, a more diverse dataset, and analysis on color channels beyond RGB.</td>
</tr>
</tbody>
</table>
</table-wrap>
<sec id="s2_1">
<label>2.1</label>
<title>Denoising</title>
<p>Image denoising is a vital preprocessing act in analysis, as all types of medical images are susceptible to noise [<xref ref-type="bibr" rid="ref-100">100</xref>&#x2013;<xref ref-type="bibr" rid="ref-104">104</xref>]. The sources of noise in medical imaging can be categorized into sensor-related, acquisition-related, and radiation-related factors. Computed Tomography (<italic>CT</italic>) is a widely utilized technique for disease diagnosis, but it carries the potential risk of radiation exposure [<xref ref-type="bibr" rid="ref-105">105</xref>,<xref ref-type="bibr" rid="ref-106">106</xref>]. Reducing radiation levels can lead to increased noise<xref ref-type="fig" rid="fig-7"> </xref> in <italic>CT</italic> images. Reconstruction of Low-Dose <italic>CT</italic> (<italic>LDCT</italic>) images offers an effective approach to address this issue. <xref ref-type="fig" rid="fig-7">Fig. 7</xref> shows a GAN based framework for medical image denoising. A <italic>GAN</italic> utilizing Wasserstein distance (<italic>WGAN</italic>) and perceptual similarity was applied to denoise <italic>CT</italic> images in [<xref ref-type="bibr" rid="ref-94">94</xref>]. The perceptual loss minimized noise by aligning output and ground truth features in a defined space, while the <italic>GAN</italic> shifted the noise distribution from strong to weak. Wasserstein distance was used to compare the distributions of normal-dose <italic>CT</italic> (<italic>NDCT</italic>) and <italic>LDCT</italic> data. Feature extraction was performed using Convolutional Neural Networks (<italic>CNN</italic>) based on the Visual Geometry Group (<italic>VGG</italic>). Performance metrics such as <italic>PSNR</italic> and <italic>SSIM</italic> were employed to evaluate the outputs of different networks, with <italic>WGAN</italic>&#x2212;<italic>VGG</italic> achieving superior performance. Huang et al. in [<xref ref-type="bibr" rid="ref-107">107</xref>] presented a denoising method (<italic>DU</italic>&#x2212;<italic>GAN</italic>) which utilized U-Net-based discriminators to assess the global and local variations in the denoised and normal images. A denoising method based on Conditional <italic>GAN</italic> was popularized by Li et al. in [<xref ref-type="bibr" rid="ref-53">53</xref>] in which the image context relationship and structural information was preserved. The method was tested on <italic>LIDC</italic> dataset and was seen to outperform the state of the art works. Zhu et al. had introduced a denoising method based on <italic>GAN</italic> in [<xref ref-type="bibr" rid="ref-108">108</xref>]. Molecular activity in human tissues was captured using Single-Photon Emission Computed Tomography (<italic>SPECT</italic>), which relies on gamma rays for image acquisition in [<xref ref-type="bibr" rid="ref-91">91</xref>]. High-noise <italic>SPECT</italic> images are input into the generator, while the discriminator assesses the generated images against real samples, specifically low-noise <italic>SPECT</italic> images. The loss, which quantifies the difference between generated and real images, is used to update both the generator and discriminator simultaneously. Both components were optimized using the Adam optimizer, with a learning rate of <inline-formula id="ieqn-120"><mml:math id="mml-ieqn-120"><mml:mn>0.00001</mml:mn></mml:math></inline-formula> and <inline-formula id="ieqn-121"><mml:math id="mml-ieqn-121"><mml:mn>800</mml:mn></mml:math></inline-formula> training epochs. Noise levels were assessed using the normalized standard deviation (<italic>NSD</italic>), enabling a comparison of results with and without conditional <italic>GAN</italic> denoising applied to the reconstructed <italic>SPECT</italic> images.</p>
<fig id="fig-7">
<label>Figure 7</label>
<caption>
<title><italic>GAN</italic>-based framework for medical image denoising: enhancing image quality with adversarial training</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-7.tif"/>
</fig>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Segmentation</title>
<p>Segmentation plays a crucial role in medical image analysis. Automating the segmentation process is highly challenging due to variations in anatomical structures across different patients [<xref ref-type="bibr" rid="ref-55">55</xref>,<xref ref-type="bibr" rid="ref-109">109</xref>&#x2013;<xref ref-type="bibr" rid="ref-111">111</xref>]. Skin cancer, common among individuals with fair skin, is<xref ref-type="fig" rid="fig-8"> </xref> classified into melanoma (pigmented lesions) and non-melanoma (non-pigmented lesions). Early detection of melanoma is critical. Dermoscopic images captured via smartphones can be analyzed to detect pigmented lesions. A <italic>GAN</italic> was trained using <inline-formula id="ieqn-122"><mml:math id="mml-ieqn-122"><mml:mn>3000</mml:mn></mml:math></inline-formula> color images in [<xref ref-type="bibr" rid="ref-99">99</xref>], with the generator employing a <inline-formula id="ieqn-123"><mml:math id="mml-ieqn-123"><mml:mi>U</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>N</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula> architecture and the discriminator using the Adam optimizer. Training was conducted in two phases: without image rotation and with image rotation. The highest segmentation accuracy was achieved when training included image rotation. However, the proposed method lacked pre-processing stages. Future improvements in the same work were involved in enhancing the <italic>GAN</italic> architecture, utilizing larger datasets, and analyzing specific color channels beyond conventional <italic>RGB</italic> spaces. <italic>X</italic>-ray imaging, which uses high-frequency electromagnetic waves, depends on the radiological density of tissues to determine the level of absorption. Chest <italic>X</italic>-rays can detect infections, tuberculosis, cancer, and other chronic chest conditions. Segmenting postero-anterior chest <italic>X</italic>-ray images involves isolating the left and right lung fields and the heart. Since chest <italic>X</italic>-rays are <inline-formula id="ieqn-124"><mml:math id="mml-ieqn-124"><mml:mn>3</mml:mn><mml:mi>D</mml:mi></mml:math></inline-formula> projections onto <inline-formula id="ieqn-125"><mml:math id="mml-ieqn-125"><mml:mn>2</mml:mn><mml:mi>D</mml:mi></mml:math></inline-formula> images, overlapping structures complicate segmentation. The Structure Correcting Adversarial Network (<italic>SCAN</italic>) framework [<xref ref-type="bibr" rid="ref-99">99</xref>] applied adversarial techniques for semantic segmentation, comprising a segmentation network and a critic network that were trained jointly. Both the segmentation and critic networks in <italic>SCAN</italic> employed Fully Convolutional Networks (<inline-formula id="ieqn-126"><mml:math id="mml-ieqn-126"><mml:mi>F</mml:mi><mml:mi>C</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>). While <inline-formula id="ieqn-127"><mml:math id="mml-ieqn-127"><mml:mi>F</mml:mi><mml:mi>C</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> were effective for <italic>RGB</italic> images, <italic>SCAN</italic> used <inline-formula id="ieqn-128"><mml:math id="mml-ieqn-128"><mml:mi>F</mml:mi><mml:mi>C</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> for grayscale chest <italic>X</italic>-ray images, enabling extraction of high-level data representations. Metrics such as Intersection-over-Union (<inline-formula id="ieqn-129"><mml:math id="mml-ieqn-129"><mml:mi>I</mml:mi><mml:mi>o</mml:mi><mml:mi>U</mml:mi></mml:math></inline-formula>) and Dice coefficient were used to evaluate performance. The <italic>SCAN</italic> framework was trained on <inline-formula id="ieqn-130"><mml:math id="mml-ieqn-130"><mml:mn>247</mml:mn></mml:math></inline-formula> chest <italic>X</italic>-ray images from the Japanese Society of Radiological Technology (<italic>JSRT</italic>) dataset, which included <inline-formula id="ieqn-131"><mml:math id="mml-ieqn-131"><mml:mn>154</mml:mn></mml:math></inline-formula> images with lung nodules and <inline-formula id="ieqn-132"><mml:math id="mml-ieqn-132"><mml:mn>93</mml:mn></mml:math></inline-formula> without. Additionally, <inline-formula id="ieqn-133"><mml:math id="mml-ieqn-133"><mml:mn>138</mml:mn></mml:math></inline-formula> chest <italic>X</italic>-ray images from the Montgomery dataset were used, with <inline-formula id="ieqn-134"><mml:math id="mml-ieqn-134"><mml:mn>117</mml:mn></mml:math></inline-formula> for training and <inline-formula id="ieqn-135"><mml:math id="mml-ieqn-135"><mml:mn>21</mml:mn></mml:math></inline-formula> for testing. Approximate <inline-formula id="ieqn-136"><mml:math id="mml-ieqn-136"><mml:mi>I</mml:mi><mml:mi>o</mml:mi><mml:mi>U</mml:mi></mml:math></inline-formula> values for the <italic>JSRT</italic> and Montgomery datasets were <inline-formula id="ieqn-137"><mml:math id="mml-ieqn-137"><mml:mn>95.1</mml:mn><mml:mi mathvariant="normal">&#x0025;</mml:mi></mml:math></inline-formula> and <inline-formula id="ieqn-138"><mml:math id="mml-ieqn-138"><mml:mn>93</mml:mn><mml:mi mathvariant="normal">&#x0025;</mml:mi></mml:math></inline-formula>, respectively. This method represented a successful application of convolutional neural networks for accurate segmentation results. Spinal <italic>MR</italic> image enabled the detection of defects in the spinal cord, vertebrae, intervertebral discs, and more. The <inline-formula id="ieqn-139"><mml:math id="mml-ieqn-139"><mml:mi>S</mml:mi><mml:mi>p</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>e</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-57">57</xref>] method employed a dynamic optimization algorithm combined with a hybrid learning strategy. The <inline-formula id="ieqn-140"><mml:math id="mml-ieqn-140"><mml:mi>S</mml:mi><mml:mi>p</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>e</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> included a segmentation network with three components: an encoder, a local long short-term memory module (<italic>LSTM</italic>), and a decoder. An Atrous Convolutional Autoencoder (<italic>ACAE</italic>), was designed for spinal image representation and pixel-level classification. The <italic>LSTM</italic>, a recurrent neural network (<italic>RNN</italic>), modelled spinal structural details by leveraging spatial correlations, thereby enhancing network performance. A Convolutional Neural Network (<italic>CNN</italic>) discriminator mitigated overfitting and incorporated a robust learning strategy with a flexible optimization algorithm. The discriminator accepted inputs from either the Segmentor or the ground truth. Spinal <italic>MR</italic> data from <inline-formula id="ieqn-141"><mml:math id="mml-ieqn-141"><mml:mn>253</mml:mn></mml:math></inline-formula> patients, captured using <inline-formula id="ieqn-142"><mml:math id="mml-ieqn-142"><mml:mn>1.5</mml:mn><mml:mspace width="thinmathspace" /><mml:mi>T</mml:mi></mml:math></inline-formula> equipment, was used for training and testing, comprising <inline-formula id="ieqn-143"><mml:math id="mml-ieqn-143"><mml:mn>5343</mml:mn></mml:math></inline-formula> samples of neural foramen images, <inline-formula id="ieqn-144"><mml:math id="mml-ieqn-144"><mml:mn>1818</mml:mn></mml:math></inline-formula> disc images, and vertebrae images. Data was split into five subgroups, with one group used for training and the others for testing. Performance metrics included pixel-level accuracy, Dice coefficient, specificity, and sensitivity, yielding an accuracy of <inline-formula id="ieqn-145"><mml:math id="mml-ieqn-145"><mml:mn>96.2</mml:mn><mml:mi mathvariant="normal">&#x0025;</mml:mi></mml:math></inline-formula>, specificity of <inline-formula id="ieqn-146"><mml:math id="mml-ieqn-146"><mml:mn>89.1</mml:mn><mml:mi mathvariant="normal">&#x0025;</mml:mi></mml:math></inline-formula>, and sensitivity of <inline-formula id="ieqn-147"><mml:math id="mml-ieqn-147"><mml:mn>86</mml:mn><mml:mi mathvariant="normal">&#x0025;</mml:mi></mml:math></inline-formula>. Limitations include its inapplicability to all <italic>MR</italic> image types and the lack of prior clinical spine knowledge in the diagnosis. Future improvements could address these limitations.</p>
<fig id="fig-8">
<label>Figure 8</label>
<caption>
<title>Framework for medical image segmentation using <italic>GAN</italic> for lesion, infection detection, etc.</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-8.tif"/>
</fig>
<p><xref ref-type="fig" rid="fig-8">Fig. 8</xref> shows a framework for medical image segmentation using GAN. A healthy immune system defends against foreign bodies in the body. In autoimmune disorders, the immune system attacks healthy tissues. Human epithelial type <inline-formula id="ieqn-148"><mml:math id="mml-ieqn-148"><mml:mn>2</mml:mn></mml:math></inline-formula> (<inline-formula id="ieqn-149"><mml:math id="mml-ieqn-149"><mml:mi>H</mml:mi><mml:mi>E</mml:mi><mml:mi>p</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mn>2</mml:mn></mml:math></inline-formula>) cell images can be analyzed to diagnose autoimmune disorders [<xref ref-type="bibr" rid="ref-95">95</xref>]. Due to the large variety of patterns, the available datasets were limited. <inline-formula id="ieqn-150"><mml:math id="mml-ieqn-150"><mml:mi>c</mml:mi><mml:mi>C</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>, a transfer learning framework in <italic>GAN</italic>, is a powerful segmentation approach for <inline-formula id="ieqn-151"><mml:math id="mml-ieqn-151"><mml:mi>H</mml:mi><mml:mi>E</mml:mi><mml:mi>p</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mn>2</mml:mn></mml:math></inline-formula> datasets, addressing overfitting and improving transfer capacity. The <inline-formula id="ieqn-152"><mml:math id="mml-ieqn-152"><mml:mi>c</mml:mi><mml:mi>C</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> used three losses: <inline-formula id="ieqn-153"><mml:math id="mml-ieqn-153"><mml:mi>L</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> Loss for label prediction, <italic>GAN</italic> Loss to help the discriminator distinguish between outputs, and Softmax Loss for classification. The generator in <inline-formula id="ieqn-154"><mml:math id="mml-ieqn-154"><mml:mi>c</mml:mi><mml:mi>C</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> employed Residual <inline-formula id="ieqn-155"><mml:math id="mml-ieqn-155"><mml:mi>U</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>N</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula> (<inline-formula id="ieqn-156"><mml:math id="mml-ieqn-156"><mml:mi>R</mml:mi><mml:mi>U</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>N</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula>) architecture, while an additional epoch training scheme (<inline-formula id="ieqn-157"><mml:math id="mml-ieqn-157"><mml:mi>A</mml:mi><mml:mi>E</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula>) was used for stable training of the generator and discriminator. Two datasets were used: the first dataset contained <inline-formula id="ieqn-158"><mml:math id="mml-ieqn-158"><mml:mn>252</mml:mn></mml:math></inline-formula> specimens from six categories (Homogeneous, Speckled, Nucleolar, Centromere, Golgi, Nuclear membrane, and Mitotic spindle), while the second dataset contained <inline-formula id="ieqn-159"><mml:math id="mml-ieqn-159"><mml:mn>28</mml:mn></mml:math></inline-formula> green-channel <inline-formula id="ieqn-160"><mml:math id="mml-ieqn-160"><mml:mi>H</mml:mi><mml:mi>E</mml:mi><mml:mi>p</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mn>2</mml:mn></mml:math></inline-formula> images with six categories (Centromere, Homogeneous, Fine Speckled, Coarse Speckled, Nucleolar, and Cytoplasmic). The smaller size of the second dataset impacted fine-tuning, making it less effective for segmentation. Segmentation accuracy and precision were used for evaluation, with <inline-formula id="ieqn-161"><mml:math id="mml-ieqn-161"><mml:mi>c</mml:mi><mml:mi>C</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> achieving <inline-formula id="ieqn-162"><mml:math id="mml-ieqn-162"><mml:mn>86.15</mml:mn><mml:mi mathvariant="normal">&#x0025;</mml:mi></mml:math></inline-formula> accuracy compared to Fully Convolutional ResNet (<italic>FCRN</italic>), which achieved <inline-formula id="ieqn-163"><mml:math id="mml-ieqn-163"><mml:mn>87.29</mml:mn><mml:mi mathvariant="normal">&#x0025;</mml:mi></mml:math></inline-formula>. The <italic>FCRN</italic>&#x2019;s deeper network architecture improved accuracy but also increased the risk of overfitting due to a higher number of parameters. <inline-formula id="ieqn-164"><mml:math id="mml-ieqn-164"><mml:mi>c</mml:mi><mml:mi>C</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> effectively distinguished between real and synthetic samples and classified cells. A multi-scale <inline-formula id="ieqn-165"><mml:math id="mml-ieqn-165"><mml:mi>L</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> loss function was effective for semantic segmentation tasks [<xref ref-type="bibr" rid="ref-56">56</xref>]. Training for <inline-formula id="ieqn-166"><mml:math id="mml-ieqn-166"><mml:mi>S</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> followed a conventional <italic>GAN</italic> min-max framework, involving a Segmentor network (<italic>S</italic>) and a Critic network (<italic>C</italic>). Both networks were trained using back propagation, optimizing the multi-scale <inline-formula id="ieqn-167"><mml:math id="mml-ieqn-167"><mml:mi>L</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> loss. The Segmentor is a fully convolutional encoder-decoder with a kernel size of <inline-formula id="ieqn-168"><mml:math id="mml-ieqn-168"><mml:mn>4</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>4</mml:mn></mml:math></inline-formula> and a stride of <inline-formula id="ieqn-169"><mml:math id="mml-ieqn-169"><mml:mn>2</mml:mn></mml:math></inline-formula> for downsampling. The same loss function was applied for training both the Segmentor and the Critic. The method was evaluated using <inline-formula id="ieqn-170"><mml:math id="mml-ieqn-170"><mml:mn>220</mml:mn></mml:math></inline-formula> high-grade and <inline-formula id="ieqn-171"><mml:math id="mml-ieqn-171"><mml:mn>54</mml:mn></mml:math></inline-formula> low-grade brain <italic>MR</italic> images, with Dice, Precision, and Sensitivity serving as evaluation metrics. <inline-formula id="ieqn-172"><mml:math id="mml-ieqn-172"><mml:mi>S</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> compared the segmented image with the ground truth across multiple critic layers. Xue et al. in [<xref ref-type="bibr" rid="ref-112">112</xref>] had introduced a method for multi organ thorax segmentation using <italic>U</italic>&#x2212;<italic>GAN</italic>.</p>
<p>The neural network architecture proposed in [<xref ref-type="bibr" rid="ref-113">113</xref>] is composed of two main components: a Segmentor and a Critic. To enhance semantic feature extraction, a novel model named <inline-formula id="ieqn-173"><mml:math id="mml-ieqn-173"><mml:mi>T</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi><mml:mi>f</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>m</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>C</mml:mi><mml:mi>V</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>U</mml:mi><mml:mi>n</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>T</mml:mi><mml:mi>C</mml:mi><mml:mi>U</mml:mi><mml:mi>n</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> is introduced. <inline-formula id="ieqn-174"><mml:math id="mml-ieqn-174"><mml:mi>T</mml:mi><mml:mi>C</mml:mi><mml:mi>U</mml:mi><mml:mi>n</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula> was utilized as the generator within a <italic>GAN</italic> framework to perform image segmentation, improving both robustness and computational efficiency. Once the generator segmented the target images, the Critic integrated the latent features with hierarchical information from various modalities. Additionally, a hybrid adversarial mechanism incorporating a multi-phase <italic>CV</italic> energy functional is developed. This integrated framework, referred to as <inline-formula id="ieqn-175"><mml:math id="mml-ieqn-175"><mml:mi>A</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mi>T</mml:mi><mml:mi>C</mml:mi><mml:mi>U</mml:mi><mml:mi>n</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula>, leverages the strengths of both the generator and critic modules. Comprehensive evaluations on the <inline-formula id="ieqn-176"><mml:math id="mml-ieqn-176"><mml:mi>B</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>T</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> <inline-formula id="ieqn-177"><mml:math id="mml-ieqn-177"><mml:mn>2019</mml:mn><mml:mrow><mml:mo>&#x2212;</mml:mo></mml:mrow><mml:mn>2021</mml:mn></mml:math></inline-formula> datasets demonstrate that the proposed method outperforms current leading approaches in brain tumor segmentation on <italic>MRI</italic> scans, achieving Dice Similarity Coefficients of <inline-formula id="ieqn-178"><mml:math id="mml-ieqn-178"><mml:mn>0.8642</mml:mn></mml:math></inline-formula> for Enhancing Tumor (ET), <inline-formula id="ieqn-179"><mml:math id="mml-ieqn-179"><mml:mn>0.9303</mml:mn></mml:math></inline-formula> for Whole Tumor (WT), and <inline-formula id="ieqn-180"><mml:math id="mml-ieqn-180"><mml:mn>0.9060</mml:mn></mml:math></inline-formula> for Tumor Core (TC) on <inline-formula id="ieqn-181"><mml:math id="mml-ieqn-181"><mml:mi>B</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>T</mml:mi><mml:mi>s</mml:mi><mml:mn>2021</mml:mn></mml:math></inline-formula>. A Local Cross-Attention Unet (<inline-formula id="ieqn-182"><mml:math id="mml-ieqn-182"><mml:mi>L</mml:mi><mml:mi>C</mml:mi><mml:mi>A</mml:mi><mml:mi>U</mml:mi><mml:mi>n</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula>) was introduced by Wang et al. in [<xref ref-type="bibr" rid="ref-114">114</xref>] for the segmentation of skin lesions using edge and body features.</p>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>Super Resolution</title>
<p>High-resolution images play a crucial role in accurate disease diagnosis. Despite advancements in medical imaging techniques, factors such as imaging conditions, equipment limitations, and environmental obstructions can result in low-resolution images [<xref ref-type="bibr" rid="ref-115">115</xref>&#x2013;<xref ref-type="bibr" rid="ref-119">119</xref>]. Resolution can be improved by various methods such as enhancing the spatial resolution, interpolation, multi image super resolution methods. <italic>MR</italic> image resolution can be improved using <italic>GAN</italic> technique by taking advantage of the volumetric information in the <inline-formula id="ieqn-183"><mml:math id="mml-ieqn-183"><mml:mn>3</mml:mn><mml:mi>D</mml:mi><mml:mi>M</mml:mi><mml:mi>R</mml:mi></mml:math></inline-formula> image [<xref ref-type="bibr" rid="ref-59">59</xref>,<xref ref-type="bibr" rid="ref-120">120</xref>]. The proposed architecture to improve the <italic>MRI</italic> resolution was based on Super Resolution <italic>GAN</italic> (<italic>SRGAN</italic>). The difference between conventional <italic>GAN</italic> and <italic>SRGAN</italic> is in the convolutional filters, they are <inline-formula id="ieqn-184"><mml:math id="mml-ieqn-184"><mml:mn>3</mml:mn><mml:mi>D</mml:mi></mml:math></inline-formula> convolutional layers capable to handle volumetric information. To overcome the problem of vanishing gradient in cross entropy loss function, adversarial loss was used in the discriminator and in generator network, combination of adversarial and content loss is used. Generator network contains had six residual blocks with <inline-formula id="ieqn-185"><mml:math id="mml-ieqn-185"><mml:mn>32</mml:mn></mml:math></inline-formula> convolutional filters of size <inline-formula id="ieqn-186"><mml:math id="mml-ieqn-186"><mml:mn>3</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>3</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>3</mml:mn></mml:math></inline-formula>. The discriminator network had eight convolutional layers of kernel size <inline-formula id="ieqn-187"><mml:math id="mml-ieqn-187"><mml:mn>3</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>3</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>3</mml:mn></mml:math></inline-formula>. <inline-formula id="ieqn-188"><mml:math id="mml-ieqn-188"><mml:mn>589</mml:mn><mml:mi>T</mml:mi><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo></mml:math></inline-formula>weighted images (<inline-formula id="ieqn-189"><mml:math id="mml-ieqn-189"><mml:mn>470</mml:mn></mml:math></inline-formula> for training and <inline-formula id="ieqn-190"><mml:math id="mml-ieqn-190"><mml:mn>119</mml:mn></mml:math></inline-formula> for testing) from the Alzheimers Disease Neuro imaging Initiative (<italic>ADNI</italic>) database was used and skull stripping was done for all images. Adam optimization was employed in both generator and discriminator. The generated super resolution images for upsampling factor of <inline-formula id="ieqn-191"><mml:math id="mml-ieqn-191"><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mn>4</mml:mn></mml:math></inline-formula> were compared against the ground truth in terms of of peak signal-to-noise ratio (<italic>PSNR</italic>) and structural similarity index measure (<italic>SSIM</italic>). Both metrics yielded a better value for the upsampling factor of <inline-formula id="ieqn-192"><mml:math id="mml-ieqn-192"><mml:mn>2</mml:mn></mml:math></inline-formula>.</p>
<p>Histopathology is the study of presence of disease in biopsy or surgical samples using microscopes [<xref ref-type="bibr" rid="ref-121">121</xref>]. Histopathological images provides a comprehensive view of cell tissues. Different segments in a tissue are visualized by pigmenting it using different dyes. Histopathological images analysis involves segmentation, detection, feature extraction, classification, etc. A unified <italic>GAN</italic> framework with a newly designed loss function was introduced for the same in [<xref ref-type="bibr" rid="ref-96">96</xref>]. The loss function is derived from both <italic>WGAN</italic> with gradient penalty (<italic>WGAN</italic>&#x2212;<italic>GP</italic>) and Information Maximizing Generative Adversarial Networks (<inline-formula id="ieqn-193"><mml:math id="mml-ieqn-193"><mml:mi>I</mml:mi><mml:mi>n</mml:mi><mml:mi>f</mml:mi><mml:mi>o</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>). During training, the generator (<italic>G</italic>) aimed to estimate the distribution <inline-formula id="ieqn-194"><mml:math id="mml-ieqn-194"><mml:mi>P</mml:mi><mml:mi>g</mml:mi></mml:math></inline-formula> such that it aligns with the real data distribution <inline-formula id="ieqn-195"><mml:math id="mml-ieqn-195"><mml:mi>P</mml:mi><mml:mi>r</mml:mi></mml:math></inline-formula>. This estimation was done by optimizing the Earth Mover (<italic>EM</italic>) distance approximation between generator (<italic>G</italic>) and discriminator (<italic>D</italic>). In the meantime the auxiliary network (<italic>Q</italic>) maximized the mutual information between the chosen random variables and the samples generated by <italic>Q</italic>. In the testing phase <italic>Q</italic> was capable of classifying as softmax function as the final layer in the <italic>Q</italic> network. Proposed unified architecture was modelled as a pipeline comprising of the following four stages-Nuclei segmentation, Clustering/classification of cell by <italic>Q</italic> network, calculating the cell proportions and image level prediction based on cell proportions, generating interpretable cell categories by generator for different noises. Four different bone marrow datasets were examined for testing and training the network. Images from these datasets were further sub divided into cell-level images and were classified into categories such as neutrophils, myeloblasts, monocytes and lymphocytes. While training, <inline-formula id="ieqn-196"><mml:math id="mml-ieqn-196"><mml:mn>32</mml:mn></mml:math></inline-formula> or <inline-formula id="ieqn-197"><mml:math id="mml-ieqn-197"><mml:mn>64</mml:mn></mml:math></inline-formula> Gaussian noise variables were used and <inline-formula id="ieqn-198"><mml:math id="mml-ieqn-198"><mml:mn>64</mml:mn></mml:math></inline-formula> categorical variable were used. While segmenting the cell images, to separate out the touched cells, morphological opening with kernel size <inline-formula id="ieqn-199"><mml:math id="mml-ieqn-199"><mml:mn>7</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>7</mml:mn></mml:math></inline-formula> was carried out. Feature extraction process could be disturbed by the color and feature contrast, this could be avoided by using non segmented histopathological images. By using such images, cells could overlap drastically. The problem of overlapping cells could be excluded by using a bounding box of size <inline-formula id="ieqn-200"><mml:math id="mml-ieqn-200"><mml:mn>32</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>32</mml:mn></mml:math></inline-formula> on non segmented image. For each cell class, precision, recall and F-Score were the segmentation evaluation parameters. Cell clustering evaluation metrics were purity, entropy and F-score. High purity value with low entropy indicated better clustering. Gradient penalty required to compute the second order derivative, so is really time consuming while training. The proposed architecture could be improvised by changing the segmentation method. It could be further improved by considering patient related data such as clinical trials, gene expression data. Anatomical segmentation of lesions and locating pathology becomes complex when anatomy or pathology is small like retinal images, cardiac <italic>MR</italic>, etc. or the image quality is too low due the acquisition process [<xref ref-type="bibr" rid="ref-92">92</xref>].</p>
<p>Diabetic retinopathy, glaucoma are the diseases which are diagnosed with the help of retinal fundus images. Retinal fundus image resolution are not good enough to detect microaneurysms, hemorrhages as they cover small image areas. Progressive <italic>GAN</italic> (<italic>P</italic>&#x2212;<italic>GAN</italic>) generated a high resolution image from a low resolution input image. The proposed <italic>P</italic>&#x2212;<italic>GAN</italic> architecture incorporated two stages-stage <inline-formula id="ieqn-201"><mml:math id="mml-ieqn-201"><mml:mn>1</mml:mn></mml:math></inline-formula> and stage <inline-formula id="ieqn-202"><mml:math id="mml-ieqn-202"><mml:mn>2</mml:mn></mml:math></inline-formula>, with output from stage <inline-formula id="ieqn-203"><mml:math id="mml-ieqn-203"><mml:mn>1</mml:mn></mml:math></inline-formula> being the input to the stage <inline-formula id="ieqn-204"><mml:math id="mml-ieqn-204"><mml:mn>2</mml:mn></mml:math></inline-formula>. Triplet loss function ensured quality improvement in images as they progressed from one stage to another. Low resolution version from high resolution image could be acquired by the application of Gaussian filter followed by down sampling. The proposed network could yield better quality images for high scaling factors, greater than <inline-formula id="ieqn-205"><mml:math id="mml-ieqn-205"><mml:mn>8</mml:mn></mml:math></inline-formula>. The proposed architecture with <inline-formula id="ieqn-206"><mml:math id="mml-ieqn-206"><mml:mn>2</mml:mn></mml:math></inline-formula> stages have a set of generator (<inline-formula id="ieqn-207"><mml:math id="mml-ieqn-207"><mml:mi>G</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula>) and discriminator (<inline-formula id="ieqn-208"><mml:math id="mml-ieqn-208"><mml:mi>D</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula>) in stage <inline-formula id="ieqn-209"><mml:math id="mml-ieqn-209"><mml:mn>1</mml:mn></mml:math></inline-formula> and another set of generator (<inline-formula id="ieqn-210"><mml:math id="mml-ieqn-210"><mml:mi>G</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula>) and discriminator (<inline-formula id="ieqn-211"><mml:math id="mml-ieqn-211"><mml:mi>D</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula>) in stage <inline-formula id="ieqn-212"><mml:math id="mml-ieqn-212"><mml:mn>2</mml:mn></mml:math></inline-formula>. The loss function for the stage <inline-formula id="ieqn-213"><mml:math id="mml-ieqn-213"><mml:mn>1</mml:mn></mml:math></inline-formula> include Mean square error (<italic>MSE</italic>) and <italic>CNN</italic> loss terms. In contrast to stage <inline-formula id="ieqn-214"><mml:math id="mml-ieqn-214"><mml:mn>1</mml:mn></mml:math></inline-formula>, the loss function for stage <inline-formula id="ieqn-215"><mml:math id="mml-ieqn-215"><mml:mn>2</mml:mn></mml:math></inline-formula> is the triplet loss function. Every generator block the network contains a up sampler with factor <inline-formula id="ieqn-216"><mml:math id="mml-ieqn-216"><mml:mn>2</mml:mn></mml:math></inline-formula>. Triplet loss function is a minmax function such that it minimizes the distance between output image from stage <inline-formula id="ieqn-217"><mml:math id="mml-ieqn-217"><mml:mn>2</mml:mn></mml:math></inline-formula> and ground truth and maximizes the distance between output image from stage <inline-formula id="ieqn-218"><mml:math id="mml-ieqn-218"><mml:mn>2</mml:mn></mml:math></inline-formula> and output image from stage <inline-formula id="ieqn-219"><mml:math id="mml-ieqn-219"><mml:mn>1</mml:mn></mml:math></inline-formula>. This dual constrains made sure that improvement was achieved both qualitatively and quantitatively. The network wass trained with <inline-formula id="ieqn-220"><mml:math id="mml-ieqn-220"><mml:mn>5000</mml:mn></mml:math></inline-formula> retinal fundus images and tested with another <inline-formula id="ieqn-221"><mml:math id="mml-ieqn-221"><mml:mn>1000</mml:mn></mml:math></inline-formula> images. The dataset was down sampled by factors <inline-formula id="ieqn-222"><mml:math id="mml-ieqn-222"><mml:mn>2</mml:mn></mml:math></inline-formula>, <inline-formula id="ieqn-223"><mml:math id="mml-ieqn-223"><mml:mn>4</mml:mn></mml:math></inline-formula>, <inline-formula id="ieqn-224"><mml:math id="mml-ieqn-224"><mml:mn>8</mml:mn></mml:math></inline-formula>, <inline-formula id="ieqn-225"><mml:math id="mml-ieqn-225"><mml:mn>16</mml:mn></mml:math></inline-formula>, <inline-formula id="ieqn-226"><mml:math id="mml-ieqn-226"><mml:mn>32</mml:mn></mml:math></inline-formula> and for each sampling the time taken to generate super resolved image was computed. As the up sampling factor increases, there was evident increase in the time to generate super resolved images. The proposed <italic>P</italic>&#x2212;<italic>GAN</italic> was compared with various <italic>SR</italic>&#x2212;<italic>GAN</italic> algorithms. <italic>SSIM</italic>, <inline-formula id="ieqn-227"><mml:math id="mml-ieqn-227"><mml:mi>P</mml:mi><mml:mi>S</mml:mi><mml:mi>N</mml:mi><mml:mi>R</mml:mi></mml:math></inline-formula> (dB), sharpness metric (<italic>S</italic>) [<xref ref-type="bibr" rid="ref-122">122</xref>] were the metrics for network performance evaluation and was computed by varying the scaling factor and for various noise types like Gaussian, salt and pepper noise and speckle noise. With the increase in upsampling factor, a decrease in <italic>SSIM</italic>, <italic>PSNR</italic>, <inline-formula id="ieqn-228"><mml:math id="mml-ieqn-228"><mml:mi>S</mml:mi><mml:mn>3</mml:mn></mml:math></inline-formula> metrics was estimated. Out of the three noises, low quality images with Gaussian noise, only yielded the expected results. Vessel segmentation with <italic>P</italic>&#x2212;<italic>GAN</italic> outperform all other <italic>GAN</italic> algorithms and was compared with respect to segmentation accuracy, Specificity, area under curve (<italic>AUC</italic>) and sensitivity. The proposed <italic>P</italic>&#x2212;<italic>GAN</italic> had been evaluated using cardiac <italic>MR</italic> dataset for improving the resolution and for cardiac <italic>LV</italic> segmentation.</p>
<p><italic>X</italic>-Ray Computed Tomography (<italic>X</italic>-Ray <italic>CT</italic>) uses <italic>X</italic>-Rays for screening, diagnosis, image guided surgery, etc. The quality of <italic>X</italic>-Ray <italic>CT</italic> images can be improved in two ways-either by using a good quality hardware or by computationally enhancing the images. The former is not economical and also are of high radiations [<xref ref-type="bibr" rid="ref-123">123</xref>]. High radiations cause gene damages and can cause cancer. Low Dose <italic>CT</italic> (<italic>LDCT</italic>) uses lower radiation, which results in low quality <italic>CT</italic> images. Computational techniques can be employed to obtain High Resolution <italic>CT</italic> (<italic>HRCT</italic>) from <italic>LDCT</italic>. A novel residual <italic>CNN</italic>-based network in the <inline-formula id="ieqn-229"><mml:math id="mml-ieqn-229"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> framework known as Deep Cycle-Consistent Adversarial <italic>SRCT</italic> network (<italic>GAN</italic>&#x2212;<italic>CIRCLE</italic>) [<xref ref-type="bibr" rid="ref-93">93</xref>] was one among the computational techniques which preserved the anatomical details in <italic>CT</italic> images. The proposed algorithm had the following advantages over the other conventional <italic>GAN</italic> architectures 
<list list-type="bullet">
<list-item>
<p>cycle consistency to ensures strong across domain consistency between <italic>LRCT</italic> and <italic>HRCT</italic>,</p></list-item>
<list-item>
<p>exclusion of Nash equilibrium [<xref ref-type="bibr" rid="ref-30">30</xref>] problem for training <italic>GAN</italic>,</p></list-item>
<list-item>
<p>omission of over fitting problem by optimizing the network,</p></list-item>
<list-item>
<p>Inclusion of multiple cascaded layers to extract hierarchical features,</p></list-item>
<list-item>
<p>to enhance deblurring <inline-formula id="ieqn-230"><mml:math id="mml-ieqn-230"><mml:mi>L</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> norm is employed instead of <inline-formula id="ieqn-231"><mml:math id="mml-ieqn-231"><mml:mi>L</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula> norm.</p></list-item>
</list></p>
<p>The main challenges in recovering <italic>HRCT</italic> images from noisy <italic>LRCT</italic> images are-complex spatial variations in the images, presence of unique noise patterns, sampling and degradation makes the image blur. To handle these limitations, non linear <italic>SR</italic> functional blocks with residual module is included in the proposed framework, which have the ability to learn high frequency details. Adversarial learning in a cyclic manner is followed, which ensures superior quality <italic>CT</italic> images. The proposed network has two generators&#x2014;<italic>G</italic> &#x0026; <italic>F</italic>, with feature extraction network and reconstruction network. Reconstruction network has a Parallelized <italic>CNN</italic> with multi layer perceptron (<italic>MLP</italic>) to carry out nonlinear projection in the spatial domain. Parallelized <italic>CNN</italic> are network within network which can perform dimensionality reduction with faster computation and less information loss, able to learn the complex mapping at finer levels with better accuracy. The framework has been trained and tested using two datasets-twenty five images from Tibia dataset and <inline-formula id="ieqn-232"><mml:math id="mml-ieqn-232"><mml:mn>5936</mml:mn></mml:math></inline-formula> images from Abdominal dataset. Performance of proposed <italic>GAN</italic>&#x2212;<italic>CIRCLE</italic> is compared with various <italic>GAN</italic> using the metrics <italic>PSNR</italic>, <italic>SSIM</italic> and Information fidelity criterion (<italic>IFC</italic>). Enhancing the image quality acquired by portable miniature devices is of great attention recently.</p>
<p>A deep learning approach named <inline-formula id="ieqn-233"><mml:math id="mml-ieqn-233"><mml:mi>M</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi><mml:mi>S</mml:mi><mml:mi>R</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> was proposed to enhance the resolution of medical images using Generative Adversarial Networks <inline-formula id="ieqn-234"><mml:math id="mml-ieqn-234"><mml:mo stretchy="false">(</mml:mo><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. <xref ref-type="fig" rid="fig-9">Fig. 9</xref> shows the architecture of MedSRGAN. At the core of this method is a specially designed generator known as the Residual Whole Map Attention Network <inline-formula id="ieqn-235"><mml:math id="mml-ieqn-235"><mml:mo stretchy="false">(</mml:mo><mml:mi>R</mml:mi><mml:mi>W</mml:mi><mml:mi>M</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, which effectively extracts important features across multiple channels while emphasizing critical regions within the images. To optimize performance, a composite loss function was utilized&#x2014;combining content loss, adversarial loss, and feature-based adversarial loss. The <inline-formula id="ieqn-236"><mml:math id="mml-ieqn-236"><mml:mi>M</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi><mml:mi>S</mml:mi><mml:mi>R</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> model was trained and evaluated using a dataset comprising <inline-formula id="ieqn-237"><mml:math id="mml-ieqn-237"><mml:mn>242</mml:mn></mml:math></inline-formula> thoracic <italic>CT</italic> scans and <inline-formula id="ieqn-238"><mml:math id="mml-ieqn-238"><mml:mn>110</mml:mn></mml:math></inline-formula> brain <italic>MR</italic> scans. Experimental results indicated that the method not only maintains fine texture details but also produces more lifelike and accurate high-resolution images. Furthermore, a Mean Opinion Score <inline-formula id="ieqn-239"><mml:math id="mml-ieqn-239"><mml:mo stretchy="false">(</mml:mo><mml:mi>M</mml:mi><mml:mi>O</mml:mi><mml:mi>S</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> assessment<xref ref-type="fig" rid="fig-9"> </xref> conducted by five experienced radiologists on <italic>CT</italic> image slices confirmed the model&#x2019;s effectiveness.</p>
<fig id="fig-9">
<label>Figure 9</label>
<caption>
<title>Architecture of <inline-formula id="ieqn-240"><mml:math id="mml-ieqn-240"><mml:mi>M</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi><mml:mi>S</mml:mi><mml:mi>R</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-124">124</xref>]</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-9.tif"/>
</fig>
<p>Degraded images are the major drawback of portable imaging devices. Spatial resolution, contrast and noise are the three issues related with ultrasound images. Portable devices produce low resolution, low contrast and high noise images, which makes the disease diagnosis inaccurate. <italic>GAN</italic> based method provides promising results for resolution enhancement with the following advantages-non linear multi level mapping between <italic>LR</italic> and <italic>HR</italic> images, adaptive feature extraction without human intervention, image quality enhancement using discriminator <italic>D</italic>, direct and efficient one step feed forward reconstruction procedure and easier implementation on hardware like <inline-formula id="ieqn-241"><mml:math id="mml-ieqn-241"><mml:mi>F</mml:mi><mml:mi>P</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>. The encoder-decoder based <italic>GAN</italic> has the problem of bottleneck which constrain the information sharing between <italic>LR</italic> image and <italic>HR</italic> image. To eliminate this problem, U-net based architecture is employed successfully. U-net model lacks in performance due to the presence of speckle noise in the <italic>HR</italic> image. In order to overcome the problems of encoder decoder model and U-net model, sparse skip connection UNet (<inline-formula id="ieqn-242"><mml:math id="mml-ieqn-242"><mml:mi>S</mml:mi><mml:mi>S</mml:mi><mml:mi>C</mml:mi><mml:mi>U</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>N</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula>) is proposed [<xref ref-type="bibr" rid="ref-125">125</xref>]. Local patches in the low resolution images are used in the discriminator. Assuming these patches are independent, discriminator is capable of modeling high frequency details. The input image patches makes the training easier and lower the memory constraints. The proposed network uses two losses-<inline-formula id="ieqn-243"><mml:math id="mml-ieqn-243"><mml:mi>L</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> loss for conserving the low frequency information and differential loss for high frequency information like edge sharpness. Dataset for testing and training the proposed network is done using <inline-formula id="ieqn-244"><mml:math id="mml-ieqn-244"><mml:mn>50</mml:mn></mml:math></inline-formula> simulation pairs, <inline-formula id="ieqn-245"><mml:math id="mml-ieqn-245"><mml:mn>72</mml:mn></mml:math></inline-formula> vivo image pairs and <inline-formula id="ieqn-246"><mml:math id="mml-ieqn-246"><mml:mn>40</mml:mn></mml:math></inline-formula> phantom pair. Each dataset is divided into five groups&#x2014;four groups for training and one group for testing. <italic>PSNR</italic>, <italic>SSIM</italic>, contrast resolution (<italic>CR</italic>) and mutual information (<italic>MI</italic>) are estimated for the performance evaluation. <italic>PSNR</italic>, <italic>SSIM</italic> measure similarity between LR and HR images, higher the <italic>PSNR</italic>, higher intensity similarity. <italic>CR</italic> estimates the ability to differentiate the intensity difference. Even though the <italic>SSIM</italic> for <italic>U</italic>-net is higher, the <italic>U</italic>-net images have a over smoothed appearance which implies the loss of some high frequency details. The proposed <italic>SSCU</italic>-Net out performed the <italic>U</italic>-Net and the encoder-decoder model and also the images have better resolution with preserving more edge details. Luan et al. in [<xref ref-type="bibr" rid="ref-126">126</xref>] have introduced a deep learning based adaptive matching network (<inline-formula id="ieqn-247"><mml:math id="mml-ieqn-247"><mml:mi>A</mml:mi><mml:mi>M</mml:mi></mml:math></inline-formula>-<inline-formula id="ieqn-248"><mml:math id="mml-ieqn-248"><mml:mi>N</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula>) alongwith a dataset generation method named Multi-mapping (<italic>MMP</italic>) for Ultrasound Localization Microscopy (<italic>ULM</italic>). A <italic>GAN</italic> based <italic>MR</italic> image super resolution technique was proposed in [<xref ref-type="bibr" rid="ref-127">127</xref>] that uses a generator mechanism incorporating multiple feature selection methods. Jia et al. in [<xref ref-type="bibr" rid="ref-128">128</xref>] have introduced a super resolution method for retinal fundus image based on <inline-formula id="ieqn-249"><mml:math id="mml-ieqn-249"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> which uses a vascular structure prior. This is shown to have overcome the shortcomings of its previous model <inline-formula id="ieqn-250"><mml:math id="mml-ieqn-250"><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>E</mml:mi><mml:mi>S</mml:mi><mml:mi>R</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-129">129</xref>].</p>
</sec>
<sec id="s2_4">
<label>2.4</label>
<title>Translation</title>
<p>Disease diagnosis and treatment with single image modality may not be adequate in most of the cases. The images acquired cannot outline the complete anatomical details or fails to acquire the details with the desired imaging modality [<xref ref-type="bibr" rid="ref-85">85</xref>,<xref ref-type="bibr" rid="ref-90">90</xref>,<xref ref-type="bibr" rid="ref-130">130</xref>&#x2013;<xref ref-type="bibr" rid="ref-132">132</xref>]. Image translation is an optimum solution, where required image is synthesized from a different image modality, without inducing much cost or risk. The challenge involved in translating one image modality to another is the presence of unrealistic data in the output image [<xref ref-type="bibr" rid="ref-133">133</xref>]. Fully supervised learning methods are among the most widely used deep learning approaches for this task [<xref ref-type="bibr" rid="ref-134">134</xref>]. However, these methods require paired low- and high-quality images for training, which is especially challenging in medical imaging, where obtaining such aligned image pairs is difficult in real-world scenarios. To address this limitation, several unsupervised learning frameworks have been developed [<xref ref-type="bibr" rid="ref-132">132</xref>,<xref ref-type="bibr" rid="ref-135">135</xref>]. Despite their potential, these approaches often face issues such as instability, noise amplification, and the occurrence of halo artifacts. A well-known solution for unpaired image-to-image translation is the Cycle-consistent Generative Adversarial Network <inline-formula id="ieqn-251"><mml:math id="mml-ieqn-251"><mml:mo stretchy="false">(</mml:mo><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-100">100</xref>,<xref ref-type="bibr" rid="ref-136">136</xref>]. The block diagram of <inline-formula id="ieqn-252"><mml:math id="mml-ieqn-252"><mml:mo stretchy="false">(</mml:mo><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> for image-to-image translation between domains A and B is shown in <xref ref-type="fig" rid="fig-10">Fig. 10</xref>. This architecture enables the model to learn domain-specific knowledge from representative images and transfer it to another domain without requiring paired training images. Nevertheless, most bidirectional <italic>GAN</italic>-based models are insufficiently constrained. For instance, while <inline-formula id="ieqn-253"><mml:math id="mml-ieqn-253"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> excels at capturing inter-domain cycle-consistency and global appearance within domains, it often struggles to preserve local details. This limitation is particularly significant in medical imaging, where precise local details are essential for accurate decision-making. High-quality medical images are expected to have uniform illumination and well-defined structural details to support effective diagnosis.</p>
<fig id="fig-10">
<label>Figure 10</label>
<caption>
<title>The <inline-formula id="ieqn-254"><mml:math id="mml-ieqn-254"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> architecture enables unpaired image-to-image translation between two domains, <italic>A</italic> and <italic>B</italic>. It employs two generators, <inline-formula id="ieqn-255"><mml:math id="mml-ieqn-255"><mml:msub><mml:mi>G</mml:mi><mml:mrow><mml:mi>A</mml:mi><mml:mi>B</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-256"><mml:math id="mml-ieqn-256"><mml:msub><mml:mi>G</mml:mi><mml:mrow><mml:mi>B</mml:mi><mml:mi>A</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, to convert images from one domain to the other, resulting in translated outputs <inline-formula id="ieqn-257"><mml:math id="mml-ieqn-257"><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>A</mml:mi><mml:mi>B</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-258"><mml:math id="mml-ieqn-258"><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>B</mml:mi><mml:mi>A</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. To enforce cycle consistency, the model reconstructs the original inputs as <inline-formula id="ieqn-259"><mml:math id="mml-ieqn-259"><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>A</mml:mi><mml:mi>B</mml:mi><mml:mi>A</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-260"><mml:math id="mml-ieqn-260"><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>B</mml:mi><mml:mi>A</mml:mi><mml:mi>B</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> from these translations. Reconstruction fidelity is encouraged through cycle consistency losses <inline-formula id="ieqn-261"><mml:math id="mml-ieqn-261"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-262"><mml:math id="mml-ieqn-262"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, which penalize discrepancies between the original and reconstructed images. Discriminators <inline-formula id="ieqn-263"><mml:math id="mml-ieqn-263"><mml:msub><mml:mi>D</mml:mi><mml:mi>A</mml:mi></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-264"><mml:math id="mml-ieqn-264"><mml:msub><mml:mi>D</mml:mi><mml:mi>B</mml:mi></mml:msub></mml:math></inline-formula> are tasked with distinguishing real images from synthetic ones, and are trained using adversarial losses <inline-formula id="ieqn-265"><mml:math id="mml-ieqn-265"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-266"><mml:math id="mml-ieqn-266"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>. Additionally, identity losses <inline-formula id="ieqn-267"><mml:math id="mml-ieqn-267"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-268"><mml:math id="mml-ieqn-268"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> are incorporated to ensure that if a generator processes an image already from its target domain, the output remains unchanged, thus helping to preserve the image&#x2019;s content [<xref ref-type="bibr" rid="ref-100">100</xref>]</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-10.tif"/>
</fig>
<p>A skin lesion synthesizer based on <italic>GAN</italic> was proposed in [<xref ref-type="bibr" rid="ref-98">98</xref>], having a coarse to fine generator, multi scale discriminator and a robust objective function for learning. This method synthesized images from a semantic label map and an instance map [<xref ref-type="bibr" rid="ref-60">60</xref>]. Images of resolution <inline-formula id="ieqn-269"><mml:math id="mml-ieqn-269"><mml:mn>1024</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>512</mml:mn></mml:math></inline-formula> were synthesized with proposed generator which incorporated convolutional layers, few residual blocks and deconvolutional layers. The multi scale discriminator consisted of three discriminators, with same input of different resolutions. The training was stabilized using the feature matching loss function, that compared the real and synthetic image features from all the discriminators. The loss function for the proposed network included conditional <italic>GAN</italic> loss and feature matching loss. Clinically meaningful <inline-formula id="ieqn-270"><mml:math id="mml-ieqn-270"><mml:mn>2594</mml:mn></mml:math></inline-formula> skin lesion images were split into two-2346 training images and <inline-formula id="ieqn-271"><mml:math id="mml-ieqn-271"><mml:mn>248</mml:mn></mml:math></inline-formula> testing images. Quantitative analysis was performed using <italic>AUC</italic> and <inline-formula id="ieqn-272"><mml:math id="mml-ieqn-272"><mml:mi>p</mml:mi></mml:math></inline-formula>-value. Multiple images of same anatomy with different contrast improves disease diagnosis easier. Multi contrast <italic>MR</italic> image synthesizer with conditional <inline-formula id="ieqn-273"><mml:math id="mml-ieqn-273"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> was proposed in [<xref ref-type="bibr" rid="ref-61">61</xref>] for spatially registered images. <inline-formula id="ieqn-274"><mml:math id="mml-ieqn-274"><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> was compared with cyclic <italic>GAN</italic> for registered and unregistered <inline-formula id="ieqn-275"><mml:math id="mml-ieqn-275"><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula>, <inline-formula id="ieqn-276"><mml:math id="mml-ieqn-276"><mml:mi>T</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula><italic>MRI</italic> images. <inline-formula id="ieqn-277"><mml:math id="mml-ieqn-277"><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> weighted <italic>MRI</italic> image described the gray and white matters, <inline-formula id="ieqn-278"><mml:math id="mml-ieqn-278"><mml:mi>T</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula> images described cortical tissue fluid. The method based on <inline-formula id="ieqn-279"><mml:math id="mml-ieqn-279"><mml:mi>p</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> was also evaluated in the same. <inline-formula id="ieqn-280"><mml:math id="mml-ieqn-280"><mml:mi>p</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> consists of a generator, pre-trained <italic>V GG</italic> network and a discriminator. The generator learned to synthesize <inline-formula id="ieqn-281"><mml:math id="mml-ieqn-281"><mml:mi>T</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula>-weighted image from the <inline-formula id="ieqn-282"><mml:math id="mml-ieqn-282"><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula>-weighted image. Simultaneously the discriminator differentiated between the real and synthetic image. Pixel by pixel, adversarial and perceptual losses were minimized by the generator and discriminator maximized the adversarial loss. <inline-formula id="ieqn-283"><mml:math id="mml-ieqn-283"><mml:mi>p</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> was trained with pixel and perceptual loss. Likewise <inline-formula id="ieqn-284"><mml:math id="mml-ieqn-284"><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula>-weighted image was synthesized from <inline-formula id="ieqn-285"><mml:math id="mml-ieqn-285"><mml:mi>T</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula>-weighted image. <inline-formula id="ieqn-286"><mml:math id="mml-ieqn-286"><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> implementation used cycle loss instead of pixel wise loss in the <inline-formula id="ieqn-287"><mml:math id="mml-ieqn-287"><mml:mi>p</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>. <inline-formula id="ieqn-288"><mml:math id="mml-ieqn-288"><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> used two generators-<inline-formula id="ieqn-289"><mml:math id="mml-ieqn-289"><mml:mi>G</mml:mi><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> and <inline-formula id="ieqn-290"><mml:math id="mml-ieqn-290"><mml:mi>G</mml:mi><mml:mi>T</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula> and two discriminators-<inline-formula id="ieqn-291"><mml:math id="mml-ieqn-291"><mml:mi>D</mml:mi><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> and <inline-formula id="ieqn-292"><mml:math id="mml-ieqn-292"><mml:mi>D</mml:mi><mml:mi>T</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula>; one pair of generator and discriminator for <inline-formula id="ieqn-293"><mml:math id="mml-ieqn-293"><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> image and other for <inline-formula id="ieqn-294"><mml:math id="mml-ieqn-294"><mml:mi>T</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula> image. <inline-formula id="ieqn-295"><mml:math id="mml-ieqn-295"><mml:mi>G</mml:mi><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> synthesized a <inline-formula id="ieqn-296"><mml:math id="mml-ieqn-296"><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> weighted image from the respective <inline-formula id="ieqn-297"><mml:math id="mml-ieqn-297"><mml:mi>T</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula> image, meanwhile <inline-formula id="ieqn-298"><mml:math id="mml-ieqn-298"><mml:mi>D</mml:mi><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> discriminated between real <inline-formula id="ieqn-299"><mml:math id="mml-ieqn-299"><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> image and synthetic <inline-formula id="ieqn-300"><mml:math id="mml-ieqn-300"><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> image. Three datasets were used-healthy images from <italic>MIDAS</italic>, <italic>IXI</italic> and abnormal images from <italic>BRATS</italic>. Out of the <inline-formula id="ieqn-301"><mml:math id="mml-ieqn-301"><mml:mn>66</mml:mn></mml:math></inline-formula> subjects analyzed from <italic>MIDAS</italic> dataset, <inline-formula id="ieqn-302"><mml:math id="mml-ieqn-302"><mml:mn>48</mml:mn></mml:math></inline-formula> were used for training, <inline-formula id="ieqn-303"><mml:math id="mml-ieqn-303"><mml:mn>5</mml:mn></mml:math></inline-formula> for validation and <inline-formula id="ieqn-304"><mml:math id="mml-ieqn-304"><mml:mn>13</mml:mn></mml:math></inline-formula> for testing. From each subject, <inline-formula id="ieqn-305"><mml:math id="mml-ieqn-305"><mml:mn>75</mml:mn></mml:math></inline-formula> axial cross sections of brain tissues without artifacts were selected manually. <italic>IXI</italic> dataset used <inline-formula id="ieqn-306"><mml:math id="mml-ieqn-306"><mml:mn>40</mml:mn></mml:math></inline-formula> subjects with <inline-formula id="ieqn-307"><mml:math id="mml-ieqn-307"><mml:mn>25</mml:mn></mml:math></inline-formula> subjects for training, <inline-formula id="ieqn-308"><mml:math id="mml-ieqn-308"><mml:mn>5</mml:mn></mml:math></inline-formula> for validation and <inline-formula id="ieqn-309"><mml:math id="mml-ieqn-309"><mml:mn>10</mml:mn></mml:math></inline-formula> for training. <inline-formula id="ieqn-310"><mml:math id="mml-ieqn-310"><mml:mn>41</mml:mn></mml:math></inline-formula> low grade giloma patients from <italic>BRATS</italic> dataset were evaluated where <inline-formula id="ieqn-311"><mml:math id="mml-ieqn-311"><mml:mn>24</mml:mn></mml:math></inline-formula> for training, <inline-formula id="ieqn-312"><mml:math id="mml-ieqn-312"><mml:mn>2</mml:mn></mml:math></inline-formula> for validation and <inline-formula id="ieqn-313"><mml:math id="mml-ieqn-313"><mml:mn>15</mml:mn></mml:math></inline-formula> for testing. To avoid biases on dataset, <inline-formula id="ieqn-314"><mml:math id="mml-ieqn-314"><mml:mn>40</mml:mn></mml:math></inline-formula> subjects with <inline-formula id="ieqn-315"><mml:math id="mml-ieqn-315"><mml:mn>4000</mml:mn><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mn>5000</mml:mn></mml:math></inline-formula> images from each dataset were analyzed. Dataset normalization ensured the absence of bias in quantitative evaluation. Armanious et al. in [<xref ref-type="bibr" rid="ref-90">90</xref>] presented <inline-formula id="ieqn-316"><mml:math id="mml-ieqn-316"><mml:mi>M</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> which is a framework designed for end to end medical image to image translation at the image level. It leverages recent progress in Generative Adversarial Networks <inline-formula id="ieqn-317"><mml:math id="mml-ieqn-317"><mml:mo stretchy="false">(</mml:mo><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> by integrating the adversarial approach with a novel blend of non-adversarial loss functions. A key component of the system is a discriminator network that functions as a trainable feature extractor, enforcing alignment between the generated medical images and their target modalities. To ensure accurate reproduction of textures and detailed structures, style transfer losses are incorporated. Furthermore, the architecture introduces a new generator design called <inline-formula id="ieqn-318"><mml:math id="mml-ieqn-318"><mml:mi>C</mml:mi><mml:mi>a</mml:mi><mml:mi>s</mml:mi><mml:mi>N</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula>, which progressively refines medical image outputs using a series of encoder-decoder modules to improve image clarity and detail. Also, five experienced radiologists evaluated and verified <inline-formula id="ieqn-319"><mml:math id="mml-ieqn-319"><mml:mi>M</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>&#x2019;<italic>s</italic> outputs quality based on subjective assessment.</p>
<p>Yang et al. in [<xref ref-type="bibr" rid="ref-9">9</xref>] introduced a method to perform Image Modality Translation <inline-formula id="ieqn-320"><mml:math id="mml-ieqn-320"><mml:mo stretchy="false">(</mml:mo><mml:mi>I</mml:mi><mml:mi>M</mml:mi><mml:mi>T</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> using a deep learning approach grounded in Conditional Generative Adversarial Networks <inline-formula id="ieqn-321"><mml:math id="mml-ieqn-321"><mml:mo stretchy="false">(</mml:mo><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. This framework captured low-level pixel information and high-level semantic features such as brain tumors and anatomical structures across different imaging modalities showcasing strong potential as a supportive tool in medical diagnostics. A cross-modality registration technique that integrated deformation fields was introduced, enabling the model to incorporate information from the translated imaging modalities. A Translated Multichannel Segmentation <inline-formula id="ieqn-322"><mml:math id="mml-ieqn-322"><mml:mo stretchy="false">(</mml:mo><mml:mi>T</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> approach for <italic>MR</italic> data. In this method, both original and translated modalities were processed together using Fully Convolutional Networks <inline-formula id="ieqn-323"><mml:math id="mml-ieqn-323"><mml:mo stretchy="false">(</mml:mo><mml:mi>F</mml:mi><mml:mi>C</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> to perform segmentation in a multichannel fashion. These two approaches effectively leverage cross-modality information to enhance performance without the need for additional data. <xref ref-type="fig" rid="fig-11">Fig. 11</xref> illustrates the above method.</p>
<fig id="fig-11">
<label>Figure 11</label>
<caption>
<title>The figure outlines the structure of an end-to-end <italic>IMT</italic> network designed for cross-modality image generation. The training dataset is defined as <inline-formula id="ieqn-324"><mml:math id="mml-ieqn-324"><mml:mi>S</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2223;</mml:mo><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x2026;</mml:mo><mml:mo>,</mml:mo><mml:mi>n</mml:mi></mml:mrow></mml:math></inline-formula>, where each <inline-formula id="ieqn-325"><mml:math id="mml-ieqn-325"><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:math></inline-formula> represents an image from the source (given) modality, and <inline-formula id="ieqn-326"><mml:math id="mml-ieqn-326"><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:math></inline-formula> is its corresponding image in the target modality. The training process comprises two main components. First, the generator <italic>G</italic> receives the input image <inline-formula id="ieqn-327"><mml:math id="mml-ieqn-327"><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:math></inline-formula> along with a random noise vector <inline-formula id="ieqn-328"><mml:math id="mml-ieqn-328"><mml:mi>z</mml:mi></mml:math></inline-formula> and is trained to generate an output <inline-formula id="ieqn-329"><mml:math id="mml-ieqn-329"><mml:msub><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mi>i</mml:mi></mml:msub></mml:math></inline-formula> that closely resembles the true image <inline-formula id="ieqn-330"><mml:math id="mml-ieqn-330"><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:math></inline-formula>. Second, the discriminator <italic>D</italic> is responsible for distinguishing between the real target images <inline-formula id="ieqn-331"><mml:math id="mml-ieqn-331"><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:math></inline-formula> and the generated ones <inline-formula id="ieqn-332"><mml:math id="mml-ieqn-332"><mml:msub><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mi>i</mml:mi></mml:msub></mml:math></inline-formula> produced by <italic>G</italic>. The discriminator outputs either <inline-formula id="ieqn-333"><mml:math id="mml-ieqn-333"><mml:mn>1</mml:mn></mml:math></inline-formula> or <inline-formula id="ieqn-334"><mml:math id="mml-ieqn-334"><mml:mn>0</mml:mn></mml:math></inline-formula>, where <inline-formula id="ieqn-335"><mml:math id="mml-ieqn-335"><mml:mn>1</mml:mn></mml:math></inline-formula> indicates a real image and <inline-formula id="ieqn-336"><mml:math id="mml-ieqn-336"><mml:mn>0</mml:mn></mml:math></inline-formula> indicates a synthetic one. During inference, the generator utilizes the learned parameters to produce translated-modality images from new input samples <italic>G</italic> [<xref ref-type="bibr" rid="ref-9">9</xref>]</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-11.tif"/>
</fig>
<p>A novel <italic>GAN</italic> architecture named <inline-formula id="ieqn-337"><mml:math id="mml-ieqn-337"><mml:mi>M</mml:mi><mml:mi>M</mml:mi><mml:mi>T</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, which leverages a Swin Transformer backbone for multi-modal medical image translation, is introduced in [<xref ref-type="bibr" rid="ref-137">137</xref>]. The proposed system is composed of three primary components: a generator, a registration module, and a discriminator. The registration module employs a Swin Transformer to estimate a deformable vector field <inline-formula id="ieqn-338"><mml:math id="mml-ieqn-338"><mml:mo stretchy="false">(</mml:mo><mml:mi>D</mml:mi><mml:mi>V</mml:mi><mml:mi>F</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> based on the <inline-formula id="ieqn-339"><mml:math id="mml-ieqn-339"><mml:mi>S</mml:mi><mml:mi>w</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>I</mml:mi><mml:mi>R</mml:mi></mml:math></inline-formula> framework, which aligns the generated image with the target [<xref ref-type="bibr" rid="ref-138">138</xref>]. For paired datasets, spatial inconsistencies between the input and target images are corrected using this registration network. In scenarios involving unpaired data, the generator <inline-formula id="ieqn-340"><mml:math id="mml-ieqn-340"><mml:mi>G</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> produces outputs that preserve the anatomical structure of <inline-formula id="ieqn-341"><mml:math id="mml-ieqn-341"><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> images and adopt the visual style of <inline-formula id="ieqn-342"><mml:math id="mml-ieqn-342"><mml:mi>T</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula>. The registered output <inline-formula id="ieqn-343"><mml:math id="mml-ieqn-343"><mml:mi>R</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>G</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo><mml:mi>y</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> further the generated image to match both the morphology and appearance of <inline-formula id="ieqn-344"><mml:math id="mml-ieqn-344"><mml:mi>T</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula>. A convolutional neural network-based discriminator evaluates whether the generated images are indistinguishable from actual target modality images. Through extensive testing on both publicly available and clinical datasets&#x2014;covering both paired and unpaired cases&#x2014;<inline-formula id="ieqn-345"><mml:math id="mml-ieqn-345"><mml:mi>M</mml:mi><mml:mi>M</mml:mi><mml:mi>T</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> demonstrated superior performance compared to existing methods and showed strong potential for clinical adoption. Nonetheless, there are some limitations to this approach. Although processing <inline-formula id="ieqn-346"><mml:math id="mml-ieqn-346"><mml:mn>2</mml:mn><mml:mi>D</mml:mi></mml:math></inline-formula> slices offers computational efficiency, the spatial context provided by <inline-formula id="ieqn-347"><mml:math id="mml-ieqn-347"><mml:mn>3</mml:mn><mml:mi>D</mml:mi></mml:math></inline-formula> medical volumes is crucial for many diagnostic tasks. Consequently, future work should focus on extending MMTrans to operate effectively on <inline-formula id="ieqn-348"><mml:math id="mml-ieqn-348"><mml:mn>3</mml:mn><mml:mi>D</mml:mi></mml:math></inline-formula> data to fully leverage its clinical applicability.</p>
<p>Chen et al. in [<xref ref-type="bibr" rid="ref-139">139</xref>] introduced <italic>MI</italic>&#x2212;<italic>GAN</italic>, a novel multi-domain medical image translation algorithm that incorporates a key transfer branch. By analyzing the imbalance present in medical imaging datasets, the approach identified critical target domain images and constructed a specialized transfer branch. Utilizing a single generator, the method facilitates multi-domain image translation in the medical context. This structure enhanced both the model&#x2019;s attention mechanism and the quality of the generated images. Additionally, a lung image classification model was presented, leveraging synthetic image data for augmentation. The training dataset combines both synthetic lung <italic>CT</italic> scans and original real-world images to evaluate the effectiveness of the model in diagnosing normal individuals, as well as patients with mild and severe cases of <inline-formula id="ieqn-349"><mml:math id="mml-ieqn-349"><mml:mi>C</mml:mi><mml:mi>O</mml:mi><mml:mi>V</mml:mi><mml:mi>I</mml:mi><mml:mi>D</mml:mi></mml:math></inline-formula>-19. The method was seen to out perform the state of art techniques. Ozbey et al. have introduced a method in which adversarial diffusion modelling is used for obtaining improved results in image translation [<xref ref-type="bibr" rid="ref-140">140</xref>].</p>
</sec>
<sec id="s2_5">
<label>2.5</label>
<title>Reconstruction</title>
<p>Medical image reconstruction is a crucial process for generating high-quality images needed for accurate analysis. However, the quality of these images is often affected by noise and artifacts [<xref ref-type="bibr" rid="ref-141">141</xref>&#x2013;<xref ref-type="bibr" rid="ref-143">143</xref>]. To address these limitations, there has been a paradigm shift from traditional analytical and iterative reconstruction methods to data-driven machine learning approaches [<xref ref-type="bibr" rid="ref-144">144</xref>,<xref ref-type="bibr" rid="ref-145">145</xref>].</p>
<p>A review of the literature reveals that frameworks like <inline-formula id="ieqn-350"><mml:math id="mml-ieqn-350"><mml:mi>p</mml:mi><mml:mi>i</mml:mi><mml:mi>x</mml:mi><mml:mn>2</mml:mn><mml:mi>p</mml:mi><mml:mi>i</mml:mi><mml:mi>x</mml:mi></mml:math></inline-formula> and <inline-formula id="ieqn-351"><mml:math id="mml-ieqn-351"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> are widely employed for <italic>MR</italic> image reconstruction [<xref ref-type="bibr" rid="ref-120">120</xref>]. In this method <inline-formula id="ieqn-352"><mml:math id="mml-ieqn-352"><mml:mn>3</mml:mn><mml:mi>D</mml:mi></mml:math></inline-formula> neural network architecture called the multi-level densely connected super-resolution network <inline-formula id="ieqn-353"><mml:math id="mml-ieqn-353"><mml:mo stretchy="false">(</mml:mo><mml:mi>m</mml:mi><mml:mi>D</mml:mi><mml:mi>C</mml:mi><mml:mi>S</mml:mi><mml:mi>R</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, which incorporates training guidance from a <italic>GAN</italic> is proposed. The proposed <inline-formula id="ieqn-354"><mml:math id="mml-ieqn-354"><mml:mi>m</mml:mi><mml:mi>D</mml:mi><mml:mi>C</mml:mi><mml:mi>S</mml:mi><mml:mi>R</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> is designed for efficient training and inference, while the <italic>GAN</italic> component enhances the realism of the super-resolved images, making them nearly indistinguishable from the original high-resolution counterparts. An illustrative representation of this method is given in <xref ref-type="fig" rid="fig-12">Fig. 12</xref>. Liao et al. [<xref ref-type="bibr" rid="ref-146">146</xref>] explored sparse-view <italic>CBCT</italic> reconstruction to reduce artifacts, proposing a feature pyramid network for the discriminator and computing a modulated focus map to preserve anatomical structures during reconstruction. Alongside reconstruction from undersampled data, maintaining domain data accuracy is essential. In <italic>MR</italic> reconstruction, undersampled <inline-formula id="ieqn-355"><mml:math id="mml-ieqn-355"><mml:mi>k</mml:mi></mml:math></inline-formula>-space data in the frequency domain has also been addressed [<xref ref-type="bibr" rid="ref-147">147</xref>,<xref ref-type="bibr" rid="ref-148">148</xref>]. Various types of loss functions have been applied in image reconstruction to capture local image structures effectively. For example, cycle-consistency and identity loss have been utilized together for denoising cardiac <italic>CT</italic> [<xref ref-type="bibr" rid="ref-149">149</xref>]. Wolterink et al. proposed a method for low-dose <italic>CT</italic> denoising by excluding some domain loss, but this approach resulted in compromised local image structure [<xref ref-type="bibr" rid="ref-106">106</xref>]. Bhadra et al. in [<xref ref-type="bibr" rid="ref-150">150</xref>] introduced an image-adaptive <italic>GAN</italic> based reconstruction approach <inline-formula id="ieqn-356"><mml:math id="mml-ieqn-356"><mml:mo stretchy="false">(</mml:mo><mml:mi>I</mml:mi><mml:mi>A</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, designed to enhance data fidelity by adjusting the pretrained generative model parameters based on the acquired measurement data. The <italic>IAGAN</italic> framework is applied to reconstruct images from undersampled <italic>MR</italic> data. A cutting-edge generative adversarial model, Progressive Growing of <inline-formula id="ieqn-357"><mml:math id="mml-ieqn-357"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula><inline-formula id="ieqn-358"><mml:math id="mml-ieqn-358"><mml:mo stretchy="false">(</mml:mo><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>o</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, was trained using a large dataset of high-quality images from the <inline-formula id="ieqn-359"><mml:math id="mml-ieqn-359"><mml:mi>N</mml:mi><mml:mi>Y</mml:mi><mml:mi>U</mml:mi><mml:mi>f</mml:mi><mml:mi>a</mml:mi><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>M</mml:mi><mml:mi>R</mml:mi><mml:mi>I</mml:mi></mml:math></inline-formula> repository. The trained generator was then integrated into the <italic>IAGAN</italic> architecture to reconstruct high-resolution images from retrospectively undersampled <inline-formula id="ieqn-360"><mml:math id="mml-ieqn-360"><mml:mi>k</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>s</mml:mi><mml:mi>p</mml:mi><mml:mi>a</mml:mi><mml:mi>c</mml:mi><mml:mi>e</mml:mi></mml:math></inline-formula> data in the validation set. The results demonstrate that this <italic>GAN</italic> driven reconstruction method can recover intricate anatomical details from noisy or incomplete measurements, offering a level of detail that conventional reconstruction techniques&#x2014;typically dependent on sparsity-based regularization&#x2014;may struggle to achieve. This highlights the potential of <italic>IAGAN</italic> in improving the diagnostic value of <italic>MR</italic> scans.</p>
<fig id="fig-12">
<label>Figure 12</label>
<caption>
<title>The architecture consists of (<bold>A</bold>) a <inline-formula id="ieqn-361"><mml:math id="mml-ieqn-361"><mml:mi>D</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi><mml:mi>B</mml:mi><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>c</mml:mi><mml:mi>k</mml:mi></mml:math></inline-formula> utilizing <inline-formula id="ieqn-362"><mml:math id="mml-ieqn-362"><mml:mn>3</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>3</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>3</mml:mn></mml:math></inline-formula> convolutions, and (<bold>B, C</bold>) the <inline-formula id="ieqn-363"><mml:math id="mml-ieqn-363"><mml:mi>m</mml:mi><mml:mi>D</mml:mi><mml:mi>C</mml:mi><mml:mi>S</mml:mi><mml:mi>R</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>-<italic>GAN</italic> framework. The generator (<bold><italic>G</italic></bold>) follows a <inline-formula id="ieqn-364"><mml:math id="mml-ieqn-364"><mml:mi>b</mml:mi><mml:mn>4</mml:mn><mml:mi>u</mml:mi><mml:mn>4</mml:mn></mml:math></inline-formula> configuration, meaning it comprises 4 blocks with 4 units each. The initial convolutional layer produces <inline-formula id="ieqn-365"><mml:math id="mml-ieqn-365"><mml:mn>2</mml:mn><mml:mi>k</mml:mi></mml:math></inline-formula> feature maps, where <inline-formula id="ieqn-366"><mml:math id="mml-ieqn-366"><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>16</mml:mn></mml:math></inline-formula>. Each compressor within the network reduces the number of feature maps to <inline-formula id="ieqn-367"><mml:math id="mml-ieqn-367"><mml:mn>2</mml:mn><mml:mi>k</mml:mi></mml:math></inline-formula> using a <inline-formula id="ieqn-368"><mml:math id="mml-ieqn-368"><mml:mn>1</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula> convolution. The final image reconstruction is performed through an additional <inline-formula id="ieqn-369"><mml:math id="mml-ieqn-369"><mml:mn>1</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula> convolution layer. The discriminator (<bold><italic>D</italic></bold>) mirrors the architecture of <italic>SRGAN</italic>, with the exception that <inline-formula id="ieqn-370"><mml:math id="mml-ieqn-370"><mml:mi>B</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>c</mml:mi><mml:mi>h</mml:mi><mml:mi>N</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>m</mml:mi></mml:math></inline-formula> layers are replaced by <inline-formula id="ieqn-371"><mml:math id="mml-ieqn-371"><mml:mi>L</mml:mi><mml:mi>a</mml:mi><mml:mi>y</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mi>N</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>m</mml:mi></mml:math></inline-formula>, as recommended in the <italic>WGAN</italic>-<italic>GP</italic> framework [<xref ref-type="bibr" rid="ref-120">120</xref>]</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-12a.tif"/>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-12b.tif"/>
</fig>
<p>Additionally, conditional <inline-formula id="ieqn-372"><mml:math id="mml-ieqn-372"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> with skip connections in the generator have been used to synthesize full-dose equivalent <italic>PET</italic> scans from low-dose data was explored by in Rashid et al. in [<xref ref-type="bibr" rid="ref-151">151</xref>]. The primary aim of this study was to assess the effectiveness of a <inline-formula id="ieqn-373"><mml:math id="mml-ieqn-373"><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>-based approach in enhancing image quality, minimizing noise, and accelerating reconstruction time, in comparison to conventional methods such as Maximum Likelihood Expectation maximization <inline-formula id="ieqn-374"><mml:math id="mml-ieqn-374"><mml:mo stretchy="false">(</mml:mo><mml:mi>M</mml:mi><mml:mi>L</mml:mi><mml:mi>E</mml:mi><mml:mi>M</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> and Total Variation <inline-formula id="ieqn-375"><mml:math id="mml-ieqn-375"><mml:mo stretchy="false">(</mml:mo><mml:mi>T</mml:mi><mml:mi>V</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. The approach involved iterative training of a <inline-formula id="ieqn-376"><mml:math id="mml-ieqn-376"><mml:mi>U</mml:mi></mml:math></inline-formula>-<italic>Net</italic> based generator with a full-image discriminator. Results demonstrated that the proposed <inline-formula id="ieqn-377"><mml:math id="mml-ieqn-377"><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> framework significantly improved image sharpness, reduced noise, and offered faster reconstruction, outperforming traditional techniques.</p>
<p>Ahn et al. in [<xref ref-type="bibr" rid="ref-152">152</xref>] utilized <inline-formula id="ieqn-378"><mml:math id="mml-ieqn-378"><mml:mn>10</mml:mn><mml:mo>,</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mn>000</mml:mn></mml:math></inline-formula> anteroposterior <inline-formula id="ieqn-379"><mml:math id="mml-ieqn-379"><mml:mo stretchy="false">(</mml:mo><mml:mi>A</mml:mi><mml:mi>P</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> knee radiographs to develop a more cost-effective and balanced medical imaging dataset. Two convolutional neural network models were implemented: Deep Convolutional <inline-formula id="ieqn-380"><mml:math id="mml-ieqn-380"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>D</mml:mi><mml:mi>C</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> and <inline-formula id="ieqn-381"><mml:math id="mml-ieqn-381"><mml:mi>S</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mn>2</mml:mn></mml:math></inline-formula> with Adaptive Discriminator Augmentation <inline-formula id="ieqn-382"><mml:math id="mml-ieqn-382"><mml:mo stretchy="false">(</mml:mo><mml:mi>S</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mn>2</mml:mn><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>A</mml:mi><mml:mi>D</mml:mi><mml:mi>A</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. To assess the images generated by <inline-formula id="ieqn-383"><mml:math id="mml-ieqn-383"><mml:mi>S</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mn>2</mml:mn><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>A</mml:mi><mml:mi>D</mml:mi><mml:mi>A</mml:mi></mml:math></inline-formula> compared to actual radiographs, a Visual Turing Test was conducted involving two computer vision specialists, two orthopedic surgeons, and a musculoskeletal radiologist. For evaluation, the Fr&#x00E9;chet Inception Distance <inline-formula id="ieqn-384"><mml:math id="mml-ieqn-384"><mml:mo stretchy="false">(</mml:mo><mml:mi>F</mml:mi><mml:mi>I</mml:mi><mml:mi>D</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> and Principal Component Analysis <inline-formula id="ieqn-385"><mml:math id="mml-ieqn-385"><mml:mo stretchy="false">(</mml:mo><mml:mi>P</mml:mi><mml:mi>C</mml:mi><mml:mi>A</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> were employed. The synthetic images successfully replicated key pathological features such as osteophyte formation, joint space narrowing, and subchondral sclerosis. Expert classification accuracy when distinguishing real from generated images varied, with scores of 34%, 43%, 44%, 57%, and 50%. The <italic>FID</italic> score between the generated and authentic images was <inline-formula id="ieqn-386"><mml:math id="mml-ieqn-386"><mml:mn>2.96</mml:mn></mml:math></inline-formula>, significantly lower than that of another medical dataset <inline-formula id="ieqn-387"><mml:math id="mml-ieqn-387"><mml:mo stretchy="false">(</mml:mo><mml:mi>B</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>C</mml:mi><mml:mi>a</mml:mi><mml:mi>H</mml:mi><mml:mi>A</mml:mi><mml:mi>D</mml:mi><mml:mo>=</mml:mo><mml:mn>15.1</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, indicating higher image fidelity. <italic>PCA</italic> results revealed no statistically significant differences between principal components of the real and generated images <inline-formula id="ieqn-388"><mml:math id="mml-ieqn-388"><mml:mo stretchy="false">(</mml:mo><mml:mi>p</mml:mi><mml:mo>&#x003E;</mml:mo><mml:mn>0.05</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. Overall, this research highlights the potential of <inline-formula id="ieqn-389"><mml:math id="mml-ieqn-389"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in generating realistic images.</p>
<p>A novel framework named <inline-formula id="ieqn-390"><mml:math id="mml-ieqn-390"><mml:mi>D</mml:mi><mml:mi>u</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>M</mml:mi><mml:mi>M</mml:mi><mml:mi>P</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> was introduced in [<xref ref-type="bibr" rid="ref-153">153</xref>] for generating high fidelity medical images from a given source modality. This method enhances the traditional <inline-formula id="ieqn-391"><mml:math id="mml-ieqn-391"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> by incorporating dilated residual blocks, a dual-scale patch based discriminator, and a perceptual consistency loss to improve generation quality, particularly in regions containing lesions. <inline-formula id="ieqn-392"><mml:math id="mml-ieqn-392"><mml:mi>D</mml:mi><mml:mi>u</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>M</mml:mi><mml:mi>M</mml:mi><mml:mi>P</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> excels in preserving both contextual information and fine structural details, making it more effective in reconstructing lesion areas. Instead of relying on the standard single-scale discriminator used in <inline-formula id="ieqn-393"><mml:math id="mml-ieqn-393"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>, <inline-formula id="ieqn-394"><mml:math id="mml-ieqn-394"><mml:mi>D</mml:mi><mml:mi>u</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>M</mml:mi><mml:mi>M</mml:mi><mml:mi>P</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> adopts a dual-scale discriminator that operates on image patches. This design allows the network to learn features at both fine and coarse levels. The synergy between these two scales improves the model&#x2019;s adaptability to lesions of varying sizes, shapes, and locations. To enhance the network&#x2019;s ability to capture contextual features without significantly increasing computational cost, dilated residual blocks are introduced in place of standard residual blocks. These dilated blocks widen the receptive field, enabling extraction of spatial and structural information from <italic>MR</italic> data. This contributes to more accurate preservation of lesion boundaries, morphology, and overall image continuity, resulting in more detailed and high-resolution outputs. Furthermore, the model leverages perceptual consistency loss, an improvement over the traditional cycle consistency loss employed by <inline-formula id="ieqn-395"><mml:math id="mml-ieqn-395"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>. By comparing feature representations across different layers, this loss function ensures that the generated images closely resemble the target modality at multiple levels of abstraction, thereby enhancing visual clarity and detail in the synthesized <italic>MR</italic> scans.</p>
<p>Synthetic pterygium images were produced using the default configuration of the <inline-formula id="ieqn-396"><mml:math id="mml-ieqn-396"><mml:mi>S</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mn>3</mml:mn></mml:math></inline-formula> architecture in [<xref ref-type="bibr" rid="ref-154">154</xref>], chosen for its strong performance and advanced generative capabilities, which are well-suited for producing photorealistic outputs. The generator employed a <inline-formula id="ieqn-397"><mml:math id="mml-ieqn-397"><mml:mn>512</mml:mn></mml:math></inline-formula>-dimensional latent and intermediate space, with a mapping network consisting of two layers specifically adjusted for image synthesis. To maximize the model&#x2019;s representational power, the channel base was set to <inline-formula id="ieqn-398"><mml:math id="mml-ieqn-398"><mml:mn>32</mml:mn></mml:math></inline-formula>, <inline-formula id="ieqn-399"><mml:math id="mml-ieqn-399"><mml:mn>768</mml:mn></mml:math></inline-formula>, capped at <inline-formula id="ieqn-400"><mml:math id="mml-ieqn-400"><mml:mn>512</mml:mn></mml:math></inline-formula> channels. An Exponential Moving Average <inline-formula id="ieqn-401"><mml:math id="mml-ieqn-401"><mml:mo stretchy="false">(</mml:mo><mml:mi>E</mml:mi><mml:mi>M</mml:mi><mml:mi>A</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> value of 0.99 was applied to control parameter updates and promote training stability.</p>
<p>The discriminator also followed the <inline-formula id="ieqn-402"><mml:math id="mml-ieqn-402"><mml:mi>S</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mn>3</mml:mn></mml:math></inline-formula> design without freezing any layers. A minibatch standard deviation group size of four was used to encourage variation across samples during training. It mirrored the generator in terms of channel base and maximum. Both models were optimized using the Adam optimizer, configured with beta parameters <inline-formula id="ieqn-403"><mml:math id="mml-ieqn-403"><mml:mo stretchy="false">[</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mn>0.99</mml:mn><mml:mo stretchy="false">]</mml:mo></mml:math></inline-formula> and an epsilon of <inline-formula id="ieqn-404"><mml:math id="mml-ieqn-404"><mml:mn>1</mml:mn><mml:mrow><mml:mtext>e</mml:mtext></mml:mrow><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mn>08</mml:mn></mml:math></inline-formula>. Learning rates were assigned as <inline-formula id="ieqn-405"><mml:math id="mml-ieqn-405"><mml:mn>0.0025</mml:mn></mml:math></inline-formula> for the generator and <inline-formula id="ieqn-406"><mml:math id="mml-ieqn-406"><mml:mn>0.002</mml:mn></mml:math></inline-formula> for the discriminator to maintain stable learning. Loss was utilized alongside an <inline-formula id="ieqn-407"><mml:math id="mml-ieqn-407"><mml:mi>R</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> regularization term with a weight of <inline-formula id="ieqn-408"><mml:math id="mml-ieqn-408"><mml:mn>8.0</mml:mn></mml:math></inline-formula> to ensure controlled gradient updates within the discriminator.</p>
<p>Despite the architectural strengths of <inline-formula id="ieqn-409"><mml:math id="mml-ieqn-409"><mml:mi>S</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mn>3</mml:mn></mml:math></inline-formula>, generating highly realistic medical images presented certain limitations. Initially, a lack of diversity in the training set contributed to elevated Fr&#x00E9;chet Inception Distance <inline-formula id="ieqn-410"><mml:math id="mml-ieqn-410"><mml:mo stretchy="false">(</mml:mo><mml:mi>F</mml:mi><mml:mi>I</mml:mi><mml:mi>D</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> scores, reflecting a degree of synthetic bias. Additionally, the complexity and variability found in medical imagery required meticulous tuning of model parameters and implementation of sophisticated augmentation techniques to improve visual authenticity. Evaluation through confusion matrices demonstrated that the synthetic images reached a high level of realism, though clinician performance varied&#x2014;highlighting the subjective nature of visual interpretation and the difficulty in achieving universally indistinguishable synthetic outputs.</p>
<p>These outcomes suggest that <inline-formula id="ieqn-411"><mml:math id="mml-ieqn-411"><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> based reconstruction holds strong potential to enhance diagnostic precision and streamline clinical imaging workflows. Although <inline-formula id="ieqn-412"><mml:math id="mml-ieqn-412"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> have been widely applied in various medical imaging tasks, their direct integration into clinical diagnostics and decision-making remains challenging. A significant portion of current research in image reconstruction relies on conventional quantitative metrics to evaluate performance. However, when <inline-formula id="ieqn-413"><mml:math id="mml-ieqn-413"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> are trained using additional loss functions, optimizing for visual quality becomes difficult, particularly in the absence of a specialized reference metric tailored to assess perceptual image quality. This presents a key obstacle to aligning <italic>GAN</italic> generated outputs with clinical standards [<xref ref-type="bibr" rid="ref-65">65</xref>].</p>
<p>Understanding how diseases evolve over time is essential for early detection and effective treatment planning. This is particularly important for severe conditions like Idiopathic Pulmonary Fibrosis <inline-formula id="ieqn-414"><mml:math id="mml-ieqn-414"><mml:mo stretchy="false">(</mml:mo><mml:mi>I</mml:mi><mml:mi>P</mml:mi><mml:mi>F</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, a chronic and progressive lung disease with a survival rate similar to that of various cancers. <italic>CT</italic> scans are widely recognized as a dependable method for diagnosing <italic>IPF</italic>. Predicting future <italic>CT</italic> images for patients in the early stages of <italic>IPF</italic> can play a significant role in enhancing treatment strategies and improving patient outcomes. Zhao et al. introduced a novel model in [<xref ref-type="bibr" rid="ref-155">155</xref>] named <inline-formula id="ieqn-415"><mml:math id="mml-ieqn-415"><mml:mn>4</mml:mn><mml:mi>D</mml:mi></mml:math></inline-formula> Vector Quantised Generative Adversarial Networks <inline-formula id="ieqn-416"><mml:math id="mml-ieqn-416"><mml:mo stretchy="false">(</mml:mo><mml:mn>4</mml:mn><mml:mi>D</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>V</mml:mi><mml:mi>Q</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> designed to synthesize realistic <italic>CT</italic> volumes of <italic>IPF</italic> patients across different time points. The model training process involved two main stages. Initially, a <inline-formula id="ieqn-417"><mml:math id="mml-ieqn-417"><mml:mn>3</mml:mn><mml:mi>D</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>V</mml:mi><mml:mi>Q</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> was trained to reconstruct <italic>CT</italic> volumes. Subsequently, a temporal model based on Neural Ordinary Differential Equations <inline-formula id="ieqn-418"><mml:math id="mml-ieqn-418"><mml:mo stretchy="false">(</mml:mo><mml:mi>O</mml:mi><mml:mi>D</mml:mi><mml:mi>E</mml:mi><mml:mi>s</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> was trained to learn the temporal progression of the quantized latent embeddings produced by the first stage&#x2019;s encoder. Various configurations of the models were tested to generate sequential <italic>CT</italic> scans and compare to actual data using both quantitative metrics and qualitative assessments. Survival analysis based on imaging biomarkers extracted from the synthetic <italic>CT</italic> volumes were performed to validate the clinical relevance of our generated scans. The resulting concordance index <inline-formula id="ieqn-419"><mml:math id="mml-ieqn-419"><mml:mo stretchy="false">(</mml:mo><mml:mi>C</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> was on par with that of biomarkers obtained from real <italic>CT</italic> scans, indicating the potential of generated <italic>CT</italic> data for accurate survival prediction and real-world clinical application.</p>
</sec>
<sec id="s2_6">
<label>2.6</label>
<title>Data Augmentation</title>
<p>Deep learning models typically require large datasets, making it difficult to apply them in situations where limited data is available. One common solution is data augmentation, which involves creating training examples by generating new data. This approach involves basic techniques like random rotations, flipping, cropping, and adding noise. However, such transformations often fall short when applied to complex datasets like medical images. To address this, researchers have developed more strategies for medical imaging. The primary aim is to produce synthetic data that closely mirrors the original distribution. The emergence of Generative Adversarial Networks <inline-formula id="ieqn-420"><mml:math id="mml-ieqn-420"><mml:mo stretchy="false">(</mml:mo><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> has significantly enhanced data augmentation capabilities [<xref ref-type="bibr" rid="ref-70">70</xref>].</p>
<p><italic>GAN</italic> based data augmentation involves training a generator network to produce synthetic images from a latent space, thereby increasing the dataset&#x2019;s diversity and variability beyond simple transformations. This method is especially advantageous in situations with limited labeled data or class imbalance, where creating additional samples of underrepresented classes can greatly enhance classifier performance [<xref ref-type="bibr" rid="ref-156">156</xref>]. For instance, in medical image analysis, <inline-formula id="ieqn-421"><mml:math id="mml-ieqn-421"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> have been employed to generate rare pathological cases, facilitating better diagnostic model training without the need for extensive manual data collection [<xref ref-type="bibr" rid="ref-148">148</xref>].</p>
<p>A data augmentation technique was presented by Frid et al. in [<xref ref-type="bibr" rid="ref-157">157</xref>] data augmentation approach that integrates traditional image perturbation techniques with the generation of synthetic liver lesions using Generative Adversarial Networks <inline-formula id="ieqn-422"><mml:math id="mml-ieqn-422"><mml:mo stretchy="false">(</mml:mo><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> to enhance liver lesion classification accuracy. The main contributions of this work include: (1) the generation of high-quality synthetic focal liver lesions from <italic>CT</italic> scans using <inline-formula id="ieqn-423"><mml:math id="mml-ieqn-423"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, (2) the development of a convolutional neural network <inline-formula id="ieqn-424"><mml:math id="mml-ieqn-424"><mml:mo stretchy="false">(</mml:mo><mml:mi>C</mml:mi><mml:mi>N</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>-based model for liver lesion classification, and (3) the enrichment of the <italic>CNN</italic> training dataset with synthetic samples to achieve improved classification performance. The work compares the results obtained with classical augmentation (not involving the usage og <italic>GAN</italic>). It could be seen that the classification performance improved progressively with the increase in training data, reaching a plateau at approximately 78.6%, beyond which the inclusion of additional augmented samples did not lead to further enhancement in accuracy.</p>
<p>Gan et al. introduced a generative adversarial network <inline-formula id="ieqn-425"><mml:math id="mml-ieqn-425"><mml:mo stretchy="false">(</mml:mo><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> model, named <inline-formula id="ieqn-426"><mml:math id="mml-ieqn-426"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> in [<xref ref-type="bibr" rid="ref-158">158</xref>], which employs a hierarchical structure to produce high-quality synthetic knee images. This model is intended to enhance data augmentation strategies for deep learning tasks. During the training phase, <inline-formula id="ieqn-427"><mml:math id="mml-ieqn-427"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> incorporates an attention mechanism within both the generator and discriminator networks, specifically before the <inline-formula id="ieqn-428"><mml:math id="mml-ieqn-428"><mml:mn>256</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>256</mml:mn></mml:math></inline-formula> image scale, to better extract critical features from knee images. To ensure stable training, a novel approach combining pixelwise and spectral normalization was applied. <inline-formula id="ieqn-429"><mml:math id="mml-ieqn-429"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> was assessed using a large-scale dataset of knee images, with performance measured by <inline-formula id="ieqn-430"><mml:math id="mml-ieqn-430"><mml:mi>A</mml:mi><mml:mi>m</mml:mi><mml:mi>S</mml:mi><mml:mi>c</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi></mml:math></inline-formula> and <inline-formula id="ieqn-431"><mml:math id="mml-ieqn-431"><mml:mi>M</mml:mi><mml:mi>o</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>S</mml:mi><mml:mi>c</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi></mml:math></inline-formula>. Experimental results demonstrated that <inline-formula id="ieqn-432"><mml:math id="mml-ieqn-432"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> surpassed existing state-of-the-art methods. Consequently, <inline-formula id="ieqn-433"><mml:math id="mml-ieqn-433"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> represents a promising advancement toward developing more reliable deep learning models for knee image segmentation. Future research should explore clinical validation through Visual Turing Tests.</p>
<p>Furthermore, datasets augmented with <inline-formula id="ieqn-434"><mml:math id="mml-ieqn-434"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> have demonstrated improved performance of deep learning models by reducing overfitting and increasing robustness to noise and variations in data. The adversarial training process pushes the generator to create samples that challenge the discriminator, resulting in realistic synthetic data that enriches the training set [<xref ref-type="bibr" rid="ref-159">159</xref>]. However, challenges persist in ensuring the quality and diversity of <italic>GAN</italic> generated samples, as mode collapse and training instability can hinder their effectiveness.</p>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Results and Discussion</title>
<p>In medical imaging applications of <inline-formula id="ieqn-435"><mml:math id="mml-ieqn-435"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, result analysis is carried out by assessing the generated images through various performance metrics such as <italic>SSIM</italic>, <italic>PSNR</italic>, accuracy, <italic>AUC</italic>, Dice coefficient, <italic>MAE</italic>, <italic>ROC</italic> curve, <inline-formula id="ieqn-436"><mml:math id="mml-ieqn-436"><mml:mi>I</mml:mi><mml:mi>o</mml:mi><mml:mi>U</mml:mi></mml:math></inline-formula>, entropy, and normalization. These metrics are used to evaluate aspects like image quality, classification performance, segmentation accuracy, and overall model efficiency. Depending on the specific objective&#x2014;whether synthesis, denoising, translation, or segmentation&#x2014;researchers choose appropriate metrics to ensure a thorough evaluation of the <italic>GAN</italic> model&#x2019;s effectiveness. The consolidated findings of applications of <inline-formula id="ieqn-437"><mml:math id="mml-ieqn-437"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical images using standardized metrics is as given in <xref ref-type="table" rid="table-8">Tables 8</xref> and <xref ref-type="table" rid="table-9">9</xref>. In examining the table outlining <italic>GAN</italic> based applications in medical image processing, it becomes clear that different algorithms perform optimally in specific tasks. For instance, <inline-formula id="ieqn-438"><mml:math id="mml-ieqn-438"><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> demonstrates strong performance in image denoising, achieving a high Peak to Signal Ratio <inline-formula id="ieqn-439"><mml:math id="mml-ieqn-439"><mml:mo stretchy="false">(</mml:mo><mml:mi>P</mml:mi><mml:mi>S</mml:mi><mml:mi>N</mml:mi><mml:mi>R</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> and Structural Similarity Index <inline-formula id="ieqn-440"><mml:math id="mml-ieqn-440"><mml:mo stretchy="false">(</mml:mo><mml:mi>S</mml:mi><mml:mi>S</mml:mi><mml:mi>I</mml:mi><mml:mi>M</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. In the domain of image super resolution, Deep <inline-formula id="ieqn-441"><mml:math id="mml-ieqn-441"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> achieves a notable <italic>PSNR</italic> of <inline-formula id="ieqn-442"><mml:math id="mml-ieqn-442"><mml:mn>30.72</mml:mn></mml:math></inline-formula> dB and <italic>SSIM</italic> of <inline-formula id="ieqn-443"><mml:math id="mml-ieqn-443"><mml:mn>0.924</mml:mn></mml:math></inline-formula>. <italic>TGAN</italic>, which focuses on image reconstruction, records a <italic>PSNR</italic> of <inline-formula id="ieqn-444"><mml:math id="mml-ieqn-444"><mml:mn>34.69</mml:mn></mml:math></inline-formula> dB and <italic>SSIM</italic> of <inline-formula id="ieqn-445"><mml:math id="mml-ieqn-445"><mml:mn>0.953</mml:mn></mml:math></inline-formula>.</p>
<table-wrap id="table-8">
<label>Table 8</label>
<caption>
<title>Consolidated findings of applications of <inline-formula id="ieqn-446"><mml:math id="mml-ieqn-446"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical images using standardized evaluation metrics</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th>Application</th>
<th>Study/Model</th>
<th>Dataset used</th>
<th><italic>GAN</italic> architecture</th>
<th>Results</th>
</tr>
</thead>
<tbody>
<tr>
<td>Denoising</td>
<td>Li et al. [<xref ref-type="bibr" rid="ref-53">53</xref>]</td>
<td><italic>JSRT</italic> and <italic>LIDC</italic> datasets</td>
<td><inline-formula id="ieqn-447"><mml:math id="mml-ieqn-447"><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula></td>
<td><inline-formula id="ieqn-448"><mml:math id="mml-ieqn-448"><mml:mrow><mml:mtext mathvariant="bold">JSRT</mml:mtext></mml:mrow></mml:math></inline-formula><break/>PSNR: 33.26 dB,<break/>SSIM:0.9206<break/><inline-formula id="ieqn-449"><mml:math id="mml-ieqn-449"><mml:mrow><mml:mtext mathvariant="bold">LIDC</mml:mtext></mml:mrow></mml:math></inline-formula><break/>PSNR: 35.11 dB<break/>SSIM 0.9328</td>
</tr>
<tr>
<td>Denoising</td>
<td>Yang et al. [<xref ref-type="bibr" rid="ref-94">94</xref>]</td>
<td>Clinical low-dose <italic>CT</italic> scans</td>
<td><italic>WGAN</italic></td>
<td>PSNR: 22.01 dB<break/>SSIM: 0.7745</td>
</tr>
<tr>
<td>Denoising</td>
<td>Huang et al. [<xref ref-type="bibr" rid="ref-107">107</xref>]</td>
<td>Simulated and real-world low-dose <italic>CT</italic> datasets</td>
<td><italic>DU</italic>-<italic>GAN</italic> with dual <inline-formula id="ieqn-450"><mml:math id="mml-ieqn-450"><mml:mi>U</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>N</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula> discriminators</td>
<td>PSNR: 22.12 dB<break/>SSIM: 0.7454</td>
</tr>
<tr>
<td>Segmentation</td>
<td>Li et al. [<xref ref-type="bibr" rid="ref-95">95</xref>]</td>
<td><inline-formula id="ieqn-451"><mml:math id="mml-ieqn-451"><mml:mi>I</mml:mi><mml:mn>3</mml:mn><mml:mi>A</mml:mi></mml:math></inline-formula> datasets</td>
<td><inline-formula id="ieqn-452"><mml:math id="mml-ieqn-452"><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula></td>
<td>SEG: 87.29%</td>
</tr>
<tr>
<td>Segmentation</td>
<td>Yang et al. [<xref ref-type="bibr" rid="ref-94">94</xref>]</td>
<td>Clinical low-dose <italic>CT</italic> scans</td>
<td><italic>WGAN</italic></td>
<td>PSNR: 22.01 dB<break/>SSIM: 0.7745</td>
</tr>
<tr>
<td>Segmentation</td>
<td>Han et al. [<xref ref-type="bibr" rid="ref-57">57</xref>]</td>
<td>Spinal <italic>MR</italic> datasets</td>
<td><inline-formula id="ieqn-453"><mml:math id="mml-ieqn-453"><mml:mi>S</mml:mi><mml:mi>p</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>e</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula></td>
<td>Accuracy: 96.2%<break/>Dice Coefficient: 87.1%<break/>Sensitivity: 89.1%<break/>Specificity: 86.0%</td>
</tr>
<tr>
<td>Super resolution</td>
<td>Gu et al. [<xref ref-type="bibr" rid="ref-124">124</xref>]</td>
<td><italic>ADNI</italic> (589 <inline-formula id="ieqn-454"><mml:math id="mml-ieqn-454"><mml:mi>T</mml:mi><mml:mn>1</mml:mn></mml:math></inline-formula> weighted images)</td>
<td><italic>SRGAN</italic> with <inline-formula id="ieqn-455"><mml:math id="mml-ieqn-455"><mml:mn>3</mml:mn><mml:mi>D</mml:mi><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi></mml:math></inline-formula> layers</td>
<td>PSNR: 29.59 dB<break/>SSIM: 0.61</td>
</tr>
<tr>
<td>Super resolution</td>
<td>You et al. [<xref ref-type="bibr" rid="ref-93">93</xref>]</td>
<td>Tibia (25) Abdominal (5936)</td>
<td>Deep <inline-formula id="ieqn-456"><mml:math id="mml-ieqn-456"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> <inline-formula id="ieqn-457"><mml:math id="mml-ieqn-457"><mml:mo stretchy="false">(</mml:mo><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>C</mml:mi><mml:mi>I</mml:mi><mml:mi>R</mml:mi><mml:mi>C</mml:mi><mml:mi>L</mml:mi><mml:mi>E</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula></td>
<td>PSNR: 27.74 dB<break/>SSIM: 0.895<break/>PSNR: 30.72 dB<break/>SSIM: 0.924</td>
</tr>
<tr>
<td>Super resolution</td>
<td>Wang et al. [<xref ref-type="bibr" rid="ref-125">125</xref>]</td>
<td>50 simulated pairs, 40 phantom pairs, 72 <italic>in vivo</italic> pairs</td>
<td><inline-formula id="ieqn-458"><mml:math id="mml-ieqn-458"><mml:mi>S</mml:mi><mml:mi>S</mml:mi><mml:mi>C</mml:mi><mml:mi>U</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>N</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula> (Sparse Skip U-Net)</td>
<td>PSNR: 25.64 db<break/>SSIM:0.38</td>
</tr>
<tr>
<td>Translation</td>
<td>Kong et al. [<xref ref-type="bibr" rid="ref-160">160</xref>]</td>
<td><inline-formula id="ieqn-459"><mml:math id="mml-ieqn-459"><mml:mi>B</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>T</mml:mi><mml:mi>S</mml:mi><mml:mn>2018</mml:mn></mml:math></inline-formula></td>
<td>RegGAN</td>
<td>PSNR: 26.0 dB<break/>SSIM: 0.86</td>
</tr>
<tr>
<td>Translation</td>
<td>Ozbey et al. [<xref ref-type="bibr" rid="ref-140">140</xref>]</td>
<td><italic>MR</italic> brain datasets</td>
<td>Adversarial Diffusion Models <inline-formula id="ieqn-460"><mml:math id="mml-ieqn-460"><mml:mo stretchy="false">(</mml:mo><mml:mi>S</mml:mi><mml:mi>y</mml:mi><mml:mi>n</mml:mi><mml:mi>D</mml:mi><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mi>f</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula></td>
<td>PSNR: 30.09 dB<break/>SSIM: 0.94</td>
</tr>
<tr>
<td>Translation</td>
<td>Chen et al. [<xref ref-type="bibr" rid="ref-139">139</xref>]</td>
<td>Chest CT images</td>
<td>Multi domain <italic>GAN</italic> with key transfer</td>
<td><italic>GAN</italic> test: 92.19%<break/><italic>GAN</italic> train: 85.07%<break/>Accuracy: 93.85%<break/>Sensitivity: 96.69%<break/>Specificity: 89.70%<break/>AUC: 96.17%</td>
</tr>
<tr>
<td>Translation</td>
<td>Chen et al. [<xref ref-type="bibr" rid="ref-120">120</xref>]</td>
<td>Brain structural <italic>MR</italic> dataset</td>
<td><inline-formula id="ieqn-461"><mml:math id="mml-ieqn-461"><mml:mi>m</mml:mi><mml:mi>D</mml:mi><mml:mi>C</mml:mi><mml:mi>S</mml:mi><mml:mi>R</mml:mi><mml:mi>N</mml:mi><mml:mo>+</mml:mo><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula></td>
<td>PSNR: 35.88 dB<break/>SSIM: 0.9424</td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-9">
<label>Table 9</label>
<caption>
<title>Consolidated findings of applications of <inline-formula id="ieqn-462"><mml:math id="mml-ieqn-462"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical images using standardized evaluation metrics</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Application</th>
<th>Study/Model</th>
<th>Dataset used</th>
<th><italic>GAN</italic> architecture</th>
<th>Results</th>
</tr>
</thead>
<tbody>
<tr>
<td>Reconstruction</td>
<td>Ran et al. [<xref ref-type="bibr" rid="ref-102">102</xref>]</td>
<td>Clinical data and Simulated data</td>
<td>Residual encoder decoder wasserstein <inline-formula id="ieqn-463"><mml:math id="mml-ieqn-463"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>R</mml:mi><mml:mi>E</mml:mi><mml:mi>D</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>W</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula></td>
<td><bold>Clinical result</bold><break/>PSNR: 31.15 dB<break/>SSIM: 0.8624<break/><bold>Simulated result</bold><break/>PSNR: 30.22 dB<break/>SSIM: 0.8323</td>
</tr>
<tr>
<td>Reconstruction</td>
<td>Seitzer et al. [<xref ref-type="bibr" rid="ref-143">143</xref>]</td>
<td>Cardiac <italic>MR</italic> (8 fold undersampling)</td>
<td>Two stage <italic>CNN</italic> with <italic>GAN</italic> &#x002B; perceptual loss</td>
<td>PSNR: 31.89 dB<break/><italic>MOS</italic>: 3.24 <inline-formula id="ieqn-464"><mml:math id="mml-ieqn-464"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 0.63<break/><italic>SIS</italic>: 0.941</td>
</tr>
<tr>
<td>Reconstruction</td>
<td>Guo et al. [<xref ref-type="bibr" rid="ref-67">67</xref>]</td>
<td>Blister, Demodicosis, Parakeratosis, Molluscum</td>
<td><italic>MEDGAN</italic></td>
<td>MAP &#x003D; 0.96<break/>Accuracy &#x003D; 62%</td>
</tr>
<tr>
<td>Reconstriuction</td>
<td>Du et al. [<xref ref-type="bibr" rid="ref-161">161</xref>]</td>
<td>Abdominal <italic>MR</italic> image</td>
<td><italic>T</italic>&#x2212;<italic>GAN</italic></td>
<td>PSNR: 34.69 dB<break/>SSIM: 0.9353</td>
</tr>
<tr>
<td>Reconstruction</td>
<td>Zuo et al. [<xref ref-type="bibr" rid="ref-162">162</xref>]</td>
<td>ADNI Dataset</td>
<td><italic>VAE</italic>&#x2212;<italic>GAN</italic></td>
<td>ACC: 85.18<break/>SEN: 84.44<break/>SPE: 85.92<break/>AUC: 86.16</td>
</tr>
<tr>
<td>Data augmentation</td>
<td>Frid-Adar et al. [<xref ref-type="bibr" rid="ref-157">157</xref>]</td>
<td><italic>CT</italic> images of v 182 liver lesions</td>
<td><italic>GAN</italic> based Synthetic Medical Augmentation</td>
<td>Accuracy: 85.7%</td>
</tr>
<tr>
<td>Data augmentation</td>
<td>Antoniou et al. [<xref ref-type="bibr" rid="ref-156">156</xref>]</td>
<td>Omniglot dataset, <italic>EMNIST</italic> dataset, <inline-formula id="ieqn-465"><mml:math id="mml-ieqn-465"><mml:mi>V</mml:mi><mml:mi>G</mml:mi><mml:mi>G</mml:mi><mml:mi>F</mml:mi><mml:mi>a</mml:mi><mml:mi>c</mml:mi><mml:mi>e</mml:mi></mml:math></inline-formula> dataset</td>
<td>Data Augmentation <italic>GAN</italic> <inline-formula id="ieqn-466"><mml:math id="mml-ieqn-466"><mml:mo stretchy="false">(</mml:mo><mml:mi>D</mml:mi><mml:mi>A</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula></td>
<td><bold>Omniglot</bold> Accuracy: 82%<break/><bold>EMNIST</bold> Accuracy: 76%<break/><bold>VGG-Face</bold> Accuracy: 12%</td>
</tr>
<tr>
<td>Data augmentation</td>
<td>Zhang et al. [<xref ref-type="bibr" rid="ref-163">163</xref>]</td>
<td>Thyroid ultrasound image</td>
<td><italic>RFI</italic>&#x2212;<italic>GAN</italic></td>
<td>PSNR: 32.78 dB<break/>SSIM: 0.9598</td>
</tr>
<tr>
<td>Data augmentation</td>
<td>Zhang et al. [<xref ref-type="bibr" rid="ref-164">164</xref>]</td>
<td>Heart disease cleveland</td>
<td><italic>WGAN</italic>&#x2212;<italic>GP</italic></td>
<td>AUC: 0.902<break/>SPE: 0.82</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Key datasets such as <italic>OASIS</italic>, <italic>SRPBS</italic>, and <italic>ABIDE</italic> are frequently employed in these studies, highlighting their critical role in advancing research within the field. Their extensive use demonstrates their value in bench marking and evaluating algorithm performance across a range of medical imaging applications. Ultimately, the diverse nature of medical imaging tasks requires a strategic approach to selecting algorithms. While certain <italic>GAN</italic> models excel in specific areas, the variation in results underscores the need for task-specific algorithm selection. This tailored approach is essential for advancing the capabilities and accuracy of medical image analysis.</p>
</sec>
<sec id="s4">
<label>4</label>
<title>Challenges, Ethics and Future Research Directions of GAN for Medical Images</title>
<p>Although <inline-formula id="ieqn-467"><mml:math id="mml-ieqn-467"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> have demonstrated significant potential in creating lifelike medical images and enhancing diagnostic tools, they face notable challenges. The following sub-section discusses some of the major challenges while using <inline-formula id="ieqn-468"><mml:math id="mml-ieqn-468"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> for clinical use under regulatory frameworks [<xref ref-type="bibr" rid="ref-165">165</xref>].</p>
<sec id="s4_1">
<label>4.1</label>
<title>The Non Convergence Problem</title>
<p>In <inline-formula id="ieqn-469"><mml:math id="mml-ieqn-469"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, achieving convergence between the generator and discriminator at a global optimum&#x2014;known as the Nash equilibrium&#x2014;is essential. The training process follows a minimax game framework aimed at reaching this equilibrium. Effective training strategies for both networks are crucial for optimal performance. As the generator becomes more proficient, it generates synthetic images that closely resemble real ones, making it increasingly difficult for the discriminator to tell them apart. When the generator reaches peak performance and produces highly realistic images, the discriminator&#x2019;s accuracy drops to around <inline-formula id="ieqn-470"><mml:math id="mml-ieqn-470"><mml:mn>50</mml:mn><mml:mi mathvariant="normal">&#x0025;</mml:mi></mml:math></inline-formula>, indicating it can no longer differentiate between real and fake data. At this stage, the feedback provided to the generator becomes uninformative, hindering further improvement in image quality. This imbalance can ultimately result in non-convergence during <italic>GAN</italic> training [<xref ref-type="bibr" rid="ref-166">166</xref>]. The issue of non-convergence significantly impacts the quality of synthetic image generation. This problem becomes evident when examining the characteristics of the generated outputs. In many cases, non-convergence causes the generator to fail, resulting in the production of flat, single-color images&#x2014;such as entirely black or white&#x2014;particularly in gray scale image synthesis.</p>
<p>The issue of non-convergence is a significant challenge encountered during the training of <inline-formula id="ieqn-471"><mml:math id="mml-ieqn-471"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>. To address the persistent issue of non-convergence in biomedical imaging applications of <inline-formula id="ieqn-472"><mml:math id="mml-ieqn-472"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, various technical studies have been examined. One proposed solution involves guiding the training process toward achieving a Nash equilibrium, which can help stabilize <italic>GAN</italic> training. However, maintaining this equilibrium during training is notably challenging. With this concept as a foundation, the surveyed literature is organized into three principal categories [<xref ref-type="bibr" rid="ref-166">166</xref>]:
<list list-type="bullet">
<list-item>
<p>optimization of update algorithms [<xref ref-type="bibr" rid="ref-167">167</xref>]</p></list-item>
<list-item>
<p>adversarial learning [<xref ref-type="bibr" rid="ref-168">168</xref>]</p></list-item>
<list-item>
<p>tuning of hyperparameters [<xref ref-type="bibr" rid="ref-169">169</xref>].</p></list-item>
</list></p>
<sec id="s4_1_1">
<label>4.1.1</label>
<title>Optimization of Update Algorithms</title>
<p>The evolution of updating algorithms has been examined across different <italic>GAN</italic> architectures, including the original Vanilla <italic>GAN</italic> [<xref ref-type="bibr" rid="ref-42">42</xref>], the Wasserstein <italic>GAN</italic> (<italic>WGAN</italic>) introduced by [<xref ref-type="bibr" rid="ref-40">40</xref>], and the more recent <inline-formula id="ieqn-473"><mml:math id="mml-ieqn-473"><mml:mi>u</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> model designed by [<xref ref-type="bibr" rid="ref-167">167</xref>]. The update algorithms introduced in the original Vanilla <inline-formula id="ieqn-474"><mml:math id="mml-ieqn-474"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> are mostly restricted to their initial experimental contexts. While the update mechanism in <inline-formula id="ieqn-475"><mml:math id="mml-ieqn-475"><mml:mi>W</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-170">170</xref>] data has shown some success in achieving Nash equilibrium for specific applications, its applicability remains limited. While the update mechanisms in Vanilla <italic>GAN</italic> and <italic>WGAN</italic> were developed for general image generation tasks, <inline-formula id="ieqn-476"><mml:math id="mml-ieqn-476"><mml:mi>u</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> specifically targets applications in biomedical image synthesis. Each of these approaches introduces methods for adjusting the frequency of discriminator updates relative to generator updates during training. These strategies have demonstrated improved training stability and a greater ability to reach equilibrium in <italic>GAN</italic> learning.</p>
</sec>
<sec id="s4_1_2">
<label>4.1.2</label>
<title>Adversarial Learning</title>
<p>Achieving balance in <italic>GAN</italic> training is closely linked to adjusting the learning rates of the generator and discriminator. This approach was adopted by [<xref ref-type="bibr" rid="ref-168">168</xref>] to mitigate non-convergence issues in biomedical image synthesis. The underlying concept of stabilizing <italic>GAN</italic> training through learning rate control was originally proposed by [<xref ref-type="bibr" rid="ref-171">171</xref>], who introduced the Two Time-scale Update Rule (<italic>TTUR</italic>). <italic>TTUR</italic> employs separate learning rates for the generator and discriminator, enabling the model to approach a local Nash equilibrium without relying on multiple update steps. In their study, Abdelhalim et al. incorporated both <italic>TTUR</italic> and a custom discriminator update strategy into the <italic>SPGGAN</italic> framework for synthesizing skin lesion images. Specifically, they updated the discriminator five times for each generator update, promoting greater training stability. This adjustment aimed to slow down discriminator learning just enough to allow the generator to keep pace and improve image quality without mode collapse.</p>
</sec>
<sec id="s4_1_3">
<label>4.1.3</label>
<title>Tuning of Hyperparameters</title>
<p>Selecting suitable hyperparameters for controlling the generator and discriminator in <inline-formula id="ieqn-477"><mml:math id="mml-ieqn-477"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> remains a significant challenge. To tackle this issue, optimization techniques have been explored to derive adaptive loss functions that effectively guide the generator&#x2019;s weight updates. Goel et al. [<xref ref-type="bibr" rid="ref-169">169</xref>] introduced an optimized <italic>GAN</italic> framework designed to generate synthetic chest <italic>CT</italic> images for <inline-formula id="ieqn-478"><mml:math id="mml-ieqn-478"><mml:mi>C</mml:mi><mml:mi>O</mml:mi><mml:mi>V</mml:mi><mml:mi>I</mml:mi><mml:mi>D</mml:mi></mml:math></inline-formula>-19 cases. Their approach integrates a Conditional <italic>GAN</italic> (<italic>CGAN</italic>) with the Whale Optimization Algorithm (<italic>WOA</italic>), a nature-inspired metaheuristic based on the bubble-net hunting behavior of humpback whales [<xref ref-type="bibr" rid="ref-172">172</xref>]. Within this framework, the behavior of whales in locating prey is modeled to guide the generator&#x2019;s hyperparameter search. The optimization process is governed by three main rules:
<list list-type="bullet">
<list-item>
<p>Encircling strategy: The lead whale locates the prey and simulates encircling it. Analogously, the generator&#x2019;s candidate solutions (search agents) evaluate a fitness function during each iteration and refine their positions accordingly.</p></list-item>
<list-item>
<p>Distance-based updating: The proximity between the prey (optimal solution) and each search agent is calculated, and agent positions are adjusted based on this measure.</p></list-item>
<list-item>
<p>Exploration through random search: Unlike the first rule which focuses on the best-known position, this rule updates the agents&#x2019; positions based on a randomized strategy to encourage exploration of the solution space.</p></list-item>
</list></p>
<p>The use of this optimized strategy enhances both the generator&#x2019;s performance and the discriminator&#x2019;s ability to distinguish between real and synthetic images. As a result, the model achieves adaptive loss tuning, leading to the generation of higher-quality and more diverse images. In terms of performance, the optimized <italic>GAN</italic> outperformed the baseline <italic>CGAN</italic> in classification tasks using the synthesized and original images. Specifically, it achieved an <inline-formula id="ieqn-479"><mml:math id="mml-ieqn-479"><mml:mi>F</mml:mi><mml:mn>1</mml:mn><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>s</mml:mi><mml:mi>c</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi></mml:math></inline-formula> of 98.79% and accuracy of 98.78%, compared to 90.99% <inline-formula id="ieqn-480"><mml:math id="mml-ieqn-480"><mml:mi>F</mml:mi><mml:mn>1</mml:mn><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>s</mml:mi><mml:mi>c</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi></mml:math></inline-formula> and 91.60% accuracy from the baseline <italic>CGAN</italic>. These results indicate that the optimized <italic>GAN</italic> effectively balances <italic>GAN</italic> training through hyperparameter tuning.</p>
<p>Researchers have explored the use of Jensen-Shannon <inline-formula id="ieqn-481"><mml:math id="mml-ieqn-481"><mml:mo stretchy="false">(</mml:mo><mml:mi>J</mml:mi><mml:mi>S</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> divergence [<xref ref-type="bibr" rid="ref-173">173</xref>] to maintain a training balance. Alternative strategies, such as utilizing <italic>f</italic>-divergence and refined Wasserstein loss functions, have been proposed, but they still require further refinement. Research should aim to enhance the stability of <italic>GAN</italic> training by refining <italic>JS</italic> divergence and utilizing strategies like stochastic gradient descent and Pareto optimization. Also, new game-theoretic frameworks combined with divergence measures may provide promising solutions to address the non-convergence issue in <inline-formula id="ieqn-482"><mml:math id="mml-ieqn-482"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>. A summary of existing solutions to address the non-convergence problem in <inline-formula id="ieqn-483"><mml:math id="mml-ieqn-483"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> for medical images is given in <xref ref-type="table" rid="table-10">Table 10</xref>.</p>
<table-wrap id="table-10">
<label>Table 10</label>
<caption>
<title>summary of existing solutions to address the non-convergence problem in <inline-formula id="ieqn-484"><mml:math id="mml-ieqn-484"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> for medical images</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th><italic>GAN</italic> variant</th>
<th>Proposed solution</th>
<th>Summary of solution</th>
</tr>
</thead>
<tbody>
<tr>
<td><italic>SPGGANTTUR</italic> [<xref ref-type="bibr" rid="ref-168">168</xref>]</td>
<td>Adversarial training</td>
<td>The Two Time-Scale Update Rule (<italic>TTUR</italic>) helps the achievement of a local Nash equilibrium by using different learning rates to the generator and discriminator. <italic>TTUR</italic> ensures stable training dynamics and the creation of synthetic skin lesion images through this strategy.</td>
</tr>
<tr>
<td>Optimized <italic>GAN</italic> [<xref ref-type="bibr" rid="ref-169">169</xref>]</td>
<td>Adversarial training</td>
<td>The Whale Optimization Algorithm (<italic>WOA</italic>) is utilized to stabilize the training of a Conditional <italic>GAN</italic> (<italic>CGAN</italic>) to synthesize <italic>CT</italic> images. It adjusts the hyperparameters by mimicking the unique hunting strategy of humpback whales using which they locate the optimal location of the prey. This is modeled to locate the generator&#x2019;s effective search agents in coordination with a fixed discriminator, hence ensuring balanced and efficient <italic>GAN</italic> training.</td>
</tr>
<tr>
<td><inline-formula id="ieqn-485"><mml:math id="mml-ieqn-485"><mml:mi>u</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-167">167</xref>]</td>
<td>Adversarial training</td>
<td>The generator and discriminator are updated with the same number of training iterations. This enables the production of high-quality synthetic retinal images.</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Mode Collapse &#x0026; Hallucinated Features</title>
<p>Mode collapse represents a significant challenge during the training of <inline-formula id="ieqn-486"><mml:math id="mml-ieqn-486"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>. This issue often results in outputs that lack the diversity observed in real medical images. When mode collapse occurs, the generator overlooks important features and generates identical patterns, leading to a loss of meaningful variability in the synthetic data. This repetition reduces the utility of the generated images. Training a <italic>GAN</italic> to completely eliminate mode collapse remains a difficult task. For instance, when generating segmented chest radiographs using ground truth data and corresponding segmentation masks, mode collapse can lead to repetitive or incomplete outputs. The high complexity and dimensionality of <inline-formula id="ieqn-487"><mml:math id="mml-ieqn-487"><mml:mn>3</mml:mn><mml:mi>D</mml:mi></mml:math></inline-formula> brain <italic>MR</italic> data further contribute to the occurrence of mode collapse during image synthesis, making it a persistent challenge in this domain.</p>
<p><inline-formula id="ieqn-488"><mml:math id="mml-ieqn-488"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> may also introduce feature hallucinations while generating synthetic data [<xref ref-type="bibr" rid="ref-174">174</xref>]. Feature hallucination refers to the creation of artificial elements or the omission of critical features in generated images, which can lead to diagnostic errors [<xref ref-type="bibr" rid="ref-175">175</xref>]. These hallucinations often arise as a result of mode collapse during training. This issue is prominent in image to image translation tasks, where synthetic outputs may include inaccurate or misleading features. Addressing mode collapse effectively can help minimize the occurrence of hallucinated features in <italic>GAN</italic> generated images. Various strategies have been employed in <inline-formula id="ieqn-489"><mml:math id="mml-ieqn-489"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> to tackle the issue of mode collapse. Ensuring training of both the generator and discriminator is necessary. The generator must be able to capture the complete range of feature distributions and anatomical structures present, while the discriminator must provide feedback so that the generator can produce diverse outputs. Exploring advanced optimization methods to improve the stability of the training process by resolving challenges like mode collapse can significantly boost the performance of <inline-formula id="ieqn-490"><mml:math id="mml-ieqn-490"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>.</p>
<p>The mode collapse problem can be alleviated by using different methods such as
<list list-type="bullet">
<list-item>
<p>regularization</p></list-item>
<list-item>
<p>modified architectures</p></list-item>
<list-item>
<p>adversarial training.</p></list-item>
</list></p>
<sec id="s4_2_1">
<label>4.2.1</label>
<title>Regularization</title>
<p>In deep learning, minimizing the loss function is a primary objective; however, achieving this becomes difficult when the model contains excessively large weight values. Large weights can lead to overfitting, where the model performs well on training data but generalizes poorly to new, unseen data. To counteract this, regularization techniques are employed to constrain the size of the weights or limit the overall capacity of the model [<xref ref-type="bibr" rid="ref-42">42</xref>].</p>
<p>In the context of <inline-formula id="ieqn-491"><mml:math id="mml-ieqn-491"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, both the generator and discriminator are neural networks and are therefore susceptible to similar challenges. Mode collapse often occurs when the discriminator provides inconsistent or vague gradient feedback. To mitigate this issue, weight normalization is applied as a form of regularization. Unlike traditional regularization methods that introduce additional loss terms, Weight Normalization (<italic>WN</italic>) directly modifies the training process by adjusting how weights are updated. During training, gradients are backpropagated based on normalized weight matrices, improving training stability without adding to the loss function [<xref ref-type="bibr" rid="ref-176">176</xref>]. Several normalization strategies have been proposed for <inline-formula id="ieqn-492"><mml:math id="mml-ieqn-492"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, including:
<list list-type="simple">
<list-item><label>1.</label><p>Spectral Normalization [<xref ref-type="bibr" rid="ref-177">177</xref>]</p></list-item>
<list-item><label>2.</label><p>Batch Normalization [<xref ref-type="bibr" rid="ref-38">38</xref>]</p></list-item>
<list-item><label>3.</label><p>Self-Normalization [<xref ref-type="bibr" rid="ref-178">178</xref>]</p></list-item>
</list></p>
<p>Among these, spectral normalization has proven particularly effective in stabilizing <italic>GAN</italic> training by controlling the Lipschitz constant of the discriminator network.</p>
<p>Xu et al. [<xref ref-type="bibr" rid="ref-179">179</xref>] addressed the problem of mode collapse in <inline-formula id="ieqn-493"><mml:math id="mml-ieqn-493"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> applied to low-dose <italic>X</italic>-ray image super-resolution. They introduced a model called Spectral Normalization Super-Resolution <italic>GAN</italic> (<italic>SNSRGAN</italic>), which incorporated spectral normalization in the discriminator to constrain the Lipschitz constant to a value of 1. This was accomplished by applying the spectral norm&#x2014;the largest singular value of a weight matrix, equivalent to its L2 norm&#x2014;during training. To assess the performance of their model, the authors employed the Inception Score (<italic>IS</italic>) and Multi-Scale Structural Similarity Index (<italic>MS</italic>&#x2212;<italic>SSIM</italic>), both of which measure the quality and diversity of the generated images. <italic>SNSRGAN</italic> achieved an <italic>IS</italic> of <inline-formula id="ieqn-494"><mml:math id="mml-ieqn-494"><mml:mn>6.56</mml:mn></mml:math></inline-formula> and an <italic>MS</italic>&#x2212;<italic>SSIM</italic> of <inline-formula id="ieqn-495"><mml:math id="mml-ieqn-495"><mml:mn>0.986</mml:mn></mml:math></inline-formula>, outperforming the baseline <italic>SRGAN</italic> and demonstrating enhanced image diversity and resolution quality.</p>
</sec>
<sec id="s4_2_2">
<label>4.2.2</label>
<title>Modified Architectures</title>
<p>In the context of <inline-formula id="ieqn-496"><mml:math id="mml-ieqn-496"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, architectural changes involving the generator, discriminator, or both&#x2014;relative to the original (vanilla) <italic>GAN</italic>&#x2014;are referred to as modified architectures. A common strategy to mitigate the mode collapse issue is to employ multiple generators rather than a single one, as used in the vanilla <italic>GAN</italic>. This method has been shown to improve the diversity of generated outputs [<xref ref-type="bibr" rid="ref-180">180</xref>]. However, managing and training multiple generators introduces considerable complexity and demands significant computational resources. To overcome this challenge, Wu et al. [<xref ref-type="bibr" rid="ref-181">181</xref>] introduced a novel approach that utilizes multiple data distributions rather than multiple generators for synthesizing human cell images. Their model incorporates a generator based on a Gaussian Mixture Model (<italic>GMM</italic>), allowing it to capture various data distributions within the latent space. This structure, called <italic>MDGAN</italic>, enables the generation of a wide variety of image samples by drawing from a mixture of distributions. It was observed that while increasing the number of distributions can enhance sample diversity, it also substantially increases computational requirements. The synthetic human cell images generated by <italic>MDGAN</italic> were used to augment training data for classification purposes. Although the paper does not provide quantitative evaluation metrics for the generated images, it highlights that incorporating these synthetic samples led to a 4.6% improvement in <italic>CNN</italic> classification precision.</p>
</sec>
<sec id="s4_2_3">
<label>4.2.3</label>
<title>Adversarial Training</title>
<p>Creating segmentation masks and corresponding ground truth images separately using <inline-formula id="ieqn-497"><mml:math id="mml-ieqn-497"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> can be a resource-intensive and time-consuming process. To streamline this, Neff et al. [<xref ref-type="bibr" rid="ref-182">182</xref>] introduced a modified version of <italic>DCGAN</italic> designed to simultaneously generate both chest <italic>X</italic>-ray images and their associated segmentation masks within a single generation step. During adversarial training, the generator may begin to produce repetitive image-segmentation pairs with minor variations, leading to a mode collapse issue. To tackle this, the researchers employed a perceptual image hashing technique to filter out duplicate synthetic image-segmentation pairs. This approach involves computing hash values based on distinct visual features of both real and generated images. By comparing values, the similarity between two images can be determined. The effectiveness of the generated images were verified by using them to augment training data for a segmentation task. Specifically, a <inline-formula id="ieqn-498"><mml:math id="mml-ieqn-498"><mml:mi>U</mml:mi></mml:math></inline-formula>-<inline-formula id="ieqn-499"><mml:math id="mml-ieqn-499"><mml:mi>N</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula> model was trained using a dataset composed of <inline-formula id="ieqn-500"><mml:math id="mml-ieqn-500"><mml:mn>30</mml:mn></mml:math></inline-formula> real images and <inline-formula id="ieqn-501"><mml:math id="mml-ieqn-501"><mml:mn>120</mml:mn></mml:math></inline-formula> synthetic images. The evaluation pointed a Hausdorff distance of <inline-formula id="ieqn-502"><mml:math id="mml-ieqn-502"><mml:mn>7.2885</mml:mn></mml:math></inline-formula>, which was lower than when the model was trained exclusively on real or synthetic data. Despite these improvements, the authors acknowledged the presence of a mild form of mode collapse, indicating limited diversity among the generated samples. A summary of existing solutions to address the mode collapse problem in <inline-formula id="ieqn-503"><mml:math id="mml-ieqn-503"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> for medical images is shown in <xref ref-type="table" rid="table-11">Tables 11</xref> and <xref ref-type="table" rid="table-12">12</xref>.</p>
<table-wrap id="table-11">
<label>Table 11</label>
<caption>
<title>summary of existing solutions to address the mode collapse problem in <inline-formula id="ieqn-504"><mml:math id="mml-ieqn-504"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> for medical images</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th align="center"><italic>GAN</italic> variant</th>
<th align="center">Proposed solution</th>
<th align="center">Summary of solution</th>
</tr>
</thead>
<tbody>
<tr>
<td><italic>SPGGAN</italic> [<xref ref-type="bibr" rid="ref-168">168</xref>]</td>
<td>Modified architecture</td>
<td>A self-attention mechanism is presented. This module enables attention maps to highlight lesion characteristics of the dermoscopic images, enabling the generator to produce a more diverse set of synthetic images.</td>
</tr>
<tr>
<td><italic>DCGAN</italic> [<xref ref-type="bibr" rid="ref-182">182</xref>]</td>
<td>Adversarial training</td>
<td>A perceptual image hash function is employed which eliminates duplicate pairs generated during training. The differences between the real and generated images are compared by computing hash values.</td>
</tr>
<tr>
<td><italic>SNSRGAN</italic> [<xref ref-type="bibr" rid="ref-179">179</xref>]</td>
<td>Modified architecture</td>
<td>Spectral normalization is applied in <italic>SNSRGAN</italic> which constrains the spectral norm of the discriminator&#x2019;s weight matrices, thus limiting the Lipschitz constant to 1 which inturn stabizes the training.</td>
</tr>
<tr>
<td><italic>AIIN DCGAN</italic> [<xref ref-type="bibr" rid="ref-183">183</xref>]</td>
<td>Pre-processing</td>
<td><italic>AIIN</italic> is utilized in <italic>DCGAN</italic> adjusting image contrast and emphasizing key features within chest <italic>X</italic>-ray images.</td>
</tr>
<tr>
<td><italic>MSG SAGAN</italic> [<xref ref-type="bibr" rid="ref-184">184</xref>]</td>
<td>Modified architecture</td>
<td>Self attention mechanism is incorporated in generator nd discriminator <italic>MSG</italic>&#x2212;<italic>GAN</italic> to enhance the diversity of synthetic <italic>X</italic>-ray images. This mechanism utilizes attention feature maps to enable the model to capture long-range dependencies among key features the images.</td>
</tr>
<tr>
<td>Modified <italic>CGAN</italic> [<xref ref-type="bibr" rid="ref-185">185</xref>]</td>
<td>Modified architecture</td>
<td>Minibatch discrimination is incorporated into the discriminator of <italic>CGAN</italic>. The discriminator aligns its gradients with the training data, fostering more stable learning. This technique penalizes the generator when it defaults to a single mode.</td>
</tr>
<tr>
<td><italic>CGAN</italic> [<xref ref-type="bibr" rid="ref-186">186</xref>]</td>
<td>Modified architecture</td>
<td>A discriminator incorporating conditional information is introduced. This discriminator utilizes specific details from the <italic>CT</italic> scans to direct the generator. By using this information, the generator is guided to produce a diverse set of images.</td>
</tr>
<tr>
<td><italic>DCR AEGAN</italic> [<xref ref-type="bibr" rid="ref-187">187</xref>]</td>
<td>Modified architecture</td>
<td><italic>AEGAN</italic> employs a deep convolutional refiner to produce diverse <italic>MR</italic> images. This refiner, built on a <inline-formula id="ieqn-505"><mml:math id="mml-ieqn-505"><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>s</mml:mi><mml:mi>N</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula> architecture, incorporates skip connections alongside deep convolutional layers. They help in regulating the bypassing of certain training layers, which contributes to refining the shapes of generated images and enhances the overall diversity in image synthesis.</td>
</tr>
<tr>
<td><italic>AEGAN</italic> [<xref ref-type="bibr" rid="ref-188">188</xref>]</td>
<td>Modified architecture</td>
<td>A Variational Autoencoder (<italic>VAE</italic>) into <italic>AEGAN</italic> which models the probability distributions of the training data.</td>
</tr>
<tr>
<td><inline-formula id="ieqn-506"><mml:math id="mml-ieqn-506"><mml:mi>S</mml:mi><mml:mi>L</mml:mi><mml:mi>S</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-189">189</xref>]</td>
<td>Modified architecture</td>
<td>The generator of <inline-formula id="ieqn-507"><mml:math id="mml-ieqn-507"><mml:mi>S</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>&#x2019;<italic>s</italic> is fine-tuned by adjusting the number of fully connected layers. A configuration with four fully connected layers is identified as the most effective, resulting in improved diversity.</td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-12">
<label>Table 12</label>
<caption>
<title>summary of existing solutions to address the mode collapse problem in <inline-formula id="ieqn-508"><mml:math id="mml-ieqn-508"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> for medical images</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th align="center"><italic>GAN</italic> variant</th>
<th align="center">Proposed solution</th>
<th align="center">Summary of solution</th>
</tr>
</thead>
<tbody>
<tr>
<td><inline-formula id="ieqn-509"><mml:math id="mml-ieqn-509"><mml:mi>S</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-189">189</xref>]</td>
<td>Adversarial training</td>
<td>An experience replay buffer strategy is implemented. A portion of the generated masks is saved in a buffer for future reuse. The discriminator selects half of its training batches from this buffer, allowing it to revisit previously generated scar tissue samples. This helps to discourage the generator from repeatedly producing scar tissue masks with similar shape.</td>
</tr>
<tr>
<td><italic>MDGAN</italic> [<xref ref-type="bibr" rid="ref-181">181</xref>]</td>
<td>Modified architecture</td>
<td>A generator based on a Gaussian Mixture Model (<italic>GMM</italic>) is employed in <italic>MDGAN</italic>. This enables the <italic>GAN</italic> to represent multiple data distributions within the latent space.</td>
</tr>
<tr>
<td><inline-formula id="ieqn-510"><mml:math id="mml-ieqn-510"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-190">190</xref>]</td>
<td>Modified architecture</td>
<td>A modified architecture featuring a <inline-formula id="ieqn-511"><mml:math id="mml-ieqn-511"><mml:mn>34</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>34</mml:mn></mml:math></inline-formula> patch discriminator is implemented in <inline-formula id="ieqn-512"><mml:math id="mml-ieqn-512"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> to address mode collapse in <italic>MR</italic> to <italic>MR</italic> image translation. This limited receptive field enables the discriminator to focus on fine grained details, enabling the generator to preserve sharpness and high-frequency features.</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s4_3">
<label>4.3</label>
<title>Metrics for Quantitative Evaluation</title>
<p>In addressing the training challenges of <inline-formula id="ieqn-513"><mml:math id="mml-ieqn-513"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, evaluation metrics play a vital role in assessing the diversity and quality of the generated images. To measure image diversity, particularly in the context of mode collapse, metrics such as Inception Score <inline-formula id="ieqn-514"><mml:math id="mml-ieqn-514"><mml:mo stretchy="false">(</mml:mo><mml:mi>I</mml:mi><mml:mi>S</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, Multi-Scale Structural Similarity Index <inline-formula id="ieqn-515"><mml:math id="mml-ieqn-515"><mml:mo stretchy="false">(</mml:mo><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>S</mml:mi><mml:mi>S</mml:mi><mml:mi>I</mml:mi><mml:mi>M</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, and Maximum Mean Discrepancy <inline-formula id="ieqn-516"><mml:math id="mml-ieqn-516"><mml:mo stretchy="false">(</mml:mo><mml:mi>M</mml:mi><mml:mi>M</mml:mi><mml:mi>D</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> are commonly used. For issues like non-convergence and instability, Peak Signal-to-Noise Ratio <inline-formula id="ieqn-517"><mml:math id="mml-ieqn-517"><mml:mo stretchy="false">(</mml:mo><mml:mi>P</mml:mi><mml:mi>S</mml:mi><mml:mi>N</mml:mi><mml:mi>R</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> and Fr&#x00E9;chet Inception Distance <inline-formula id="ieqn-518"><mml:math id="mml-ieqn-518"><mml:mo stretchy="false">(</mml:mo><mml:mi>F</mml:mi><mml:mi>I</mml:mi><mml:mi>D</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> are frequently applied. Among these, <italic>IS</italic> and <italic>FID</italic> are standard metrics for evaluating image quality; however, both are based on models pretrained on the ImageNet dataset [<xref ref-type="bibr" rid="ref-191">191</xref>], which does not include biomedical image classes. As a result, these metrics are not ideally suited for applications in the biomedical imaging field. Additionally, <italic>MS</italic>&#x2212;<italic>SSIM</italic> is a perceptual metric focused primarily on luminance and contrast, offering a limited view of image similarity. <italic>PSNR</italic>, while widely adopted for assessing image quality, is primarily effective for grayscale images and lacks robustness in more complex scenarios. In the field of biomedical image analysis, researchers often rely on conventional pixel-wise evaluation metrics to assess the performance of <inline-formula id="ieqn-519"><mml:math id="mml-ieqn-519"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>. These metrics are typically designed for supervised learning scenarios that depend on the presence of reference images. However, in biomedical imaging, acquiring reference data is challenging due to privacy concerns and the errors associated with manual annotations. Hence, unsupervised learning methods are commonly used. Also, assessing the performance of <inline-formula id="ieqn-520"><mml:math id="mml-ieqn-520"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> is crucial due to factors such as random initialization, optimization variability, and technical complexities. Comparing generated images to real ones remains a difficult task, highlighting the need for further investigation into reliable evaluation techniques. Consequently, developing evaluation metrics that capture both subjective perceptions and objective measures presents a significant challenge for the field.</p>
</sec>
<sec id="s4_4">
<label>4.4</label>
<title>Privacy Concerns and Ethics</title>
<p>Medical imaging concerns handling highly sensitive patient data, which raises important privacy concerns during both data collection and application. While <inline-formula id="ieqn-521"><mml:math id="mml-ieqn-521"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> are capable of generating visually convincing medical images, a major issue is the introduction of artificial elements known as hallucinated features. These artifacts may mimic real pathological signs, leading to incorrect diagnostic algorithms. These can compromise clinical reliability. Therefore, it is essential to validate synthetic outputs through both expert evaluation and robust quantitative testing to ensure they reflect true features. To improve the integrity of these images, it is essential to integrate physics-informed simulations and conduct thorough experimental evaluations aimed at understanding the convergence behavior of <inline-formula id="ieqn-522"><mml:math id="mml-ieqn-522"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> within the medical imaging domain. Using <italic>GAN</italic> generated data in clinical settings requires adherence to strict regulatory standards. Health authorities such as the Food and Drug Administration <inline-formula id="ieqn-523"><mml:math id="mml-ieqn-523"><mml:mo stretchy="false">(</mml:mo><mml:mi>F</mml:mi><mml:mi>D</mml:mi><mml:mi>A</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, USA and European Medicines Agency <inline-formula id="ieqn-524"><mml:math id="mml-ieqn-524"><mml:mo stretchy="false">(</mml:mo><mml:mi>E</mml:mi><mml:mi>M</mml:mi><mml:mi>A</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> demand clear evidence of performance, safety, and reliability to show that the synthetic images are clinically valid, reproducible, and free from bias. Documentation of the development process, ethical oversight, and compliance with data protection laws (e.g., <italic>HIPAA</italic>, <italic>GDPR</italic>) are also crucial. Additionally, synthetic images must be clearly labeled, and their use justified with comprehensive risk assessments to ensure transparency.</p>
<p>Federated learning presents a promising approach to reduce privacy concerns while using <inline-formula id="ieqn-525"><mml:math id="mml-ieqn-525"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in medical imaging. However, it encounters several challenges. Achieving effective model convergence under the constraints of restricted communication bandwidth in decentralized environments remains a major difficulty. Also, combining federated learning with differential privacy requires a careful balance between safeguarding sensitive data and maintaining high model performance. The complexity of medical data because of its heterogeneity and the variety of imaging modalities complicates the collaborative model training. Future research should lay priority on overcoming communication inefficiencies, improving the resilience of models, and customizing federated learning techniques to accommodate the characteristics of medical datasets. Furthermore, exploring advanced privacy preserving methods and striving to create a fully decentralized, privacy centric <italic>GAN</italic> capable of generating high-quality medical images without compromising patient confidentiality continues to be a challenging yet vital endeavor in the healthcare sector.</p>
<p><xref ref-type="fig" rid="fig-13">Fig. 13</xref> gives an illustration of the steps in federated learning. In the first phase, each participant independently calculates the model gradients on their local data. To ensure data confidentiality, cryptographic methods such as homomorphic encryption are applied before transmitting the encrypted gradients to the central server. In the second phase, the central (master) server performs secure aggregation of the encrypted gradients. In the third phase, the aggregated results are sent back to the participants. During the fourth phase, each participant decrypts the aggregated gradients and updates their local model accordingly. This cycle is repeated iteratively until either the loss function reaches convergence or a predefined number of iterations is completed. Throughout this process, the participants&#x2019; data remains stored locally, maintaining privacy and offering an advantage over centralized approaches such as those based on Hadoop. Federated learning facilitates collaborative model training across multiple databases without the need to centralize the data, enabling scalability with growing datasets while minimizing communication overhead, as only gradients&#x2014;not raw data&#x2014;are shared.</p>
<fig id="fig-13">
<label>Figure 13</label>
<caption>
<title>Steps in federated learning [<xref ref-type="bibr" rid="ref-165">165</xref>]</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-13.tif"/>
</fig>
</sec>
<sec id="s4_5">
<label>4.5</label>
<title>Need for Human in the Loop Studies</title>
<p><italic>GAN</italic> based image generation models can produce photorealistic images; however, in medical applications, caution is essential as these images play a critical role in disease detection and diagnosis. It is imperative for medical professionals, as the end users, to thoroughly evaluate and validate <italic>GAN</italic> generated images to ensure their reliability and clinical utility. Trust and acceptance from doctors are crucial for integrating such models into healthcare workflows. While <italic>PSNR</italic> and <italic>SSIM</italic> are standard for evaluating image processing techniques [<xref ref-type="bibr" rid="ref-192">192</xref>], they often fail to reflect perceptual quality or clinical relevance, particularly in the presence of subtle distortions. This underscores the importance of expert-driven evaluation, as automated metrics alone may not capture diagnostic integrity. In medical imaging, radiologists and clinicians are best positioned to perform this qualitative assessment, as their expertise ensures that reconstructed images are not only visually plausible but also diagnostically accurate.</p>
<p>Realistic full-field digital mammograms were generated using a progressive <italic>GAN</italic> architecture [<xref ref-type="bibr" rid="ref-193">193</xref>], achieving high resolution and realism indistinguishable from real images, even by domain experts. Despite the specialized nature of medical imaging, both experts and non-experts in a reader study showed random success probabilities, emphasizing the critical role of human validation in ensuring clinical applicability. Similarly, the progressive growing <italic>GAN</italic> (<italic>PGGAN</italic>) [<xref ref-type="bibr" rid="ref-194">194</xref>] was employed to generate high-resolution chest radiographs (<inline-formula id="ieqn-526"><mml:math id="mml-ieqn-526"><mml:mi>C</mml:mi><mml:mi>X</mml:mi><mml:mi>R</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>), with two models trained separately on normal and abnormal <inline-formula id="ieqn-527"><mml:math id="mml-ieqn-527"><mml:mi>C</mml:mi><mml:mi>X</mml:mi><mml:mi>R</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, producing synthetic images of <inline-formula id="ieqn-528"><mml:math id="mml-ieqn-528"><mml:mn>1000</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>1000</mml:mn></mml:math></inline-formula> pixels. Six radiologists performed binary Turing tests on two validation sets, attaining mean accuracies of 67.42% and 69.92% in the first and second trials, respectively. Majority voting and Cohen&#x2019;s score were used to further assess diagnostic agreement and reliability. Another notable contribution is <inline-formula id="ieqn-529"><mml:math id="mml-ieqn-529"><mml:mi>M</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> [<xref ref-type="bibr" rid="ref-90">90</xref>], an end-to-end image translation framework tailored for medical applications. By integrating a conditional adversarial setup with multiple non-adversarial losses and a <inline-formula id="ieqn-530"><mml:math id="mml-ieqn-530"><mml:mi>C</mml:mi><mml:mi>a</mml:mi><mml:mi>s</mml:mi><mml:mi>N</mml:mi><mml:mi>E</mml:mi><mml:mi>T</mml:mi></mml:math></inline-formula> generator, <inline-formula id="ieqn-531"><mml:math id="mml-ieqn-531"><mml:mi>M</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> enhances global consistency and preserves high-frequency details. Its effectiveness was confirmed through expert radiologist evaluations, indicating strong clinical fidelity.</p>
<p>Further advancing <italic>GAN</italic> applications, <italic>GANCS</italic> [<xref ref-type="bibr" rid="ref-195">195</xref>] presents a compressed sensing framework that models the low-dimensional manifold of high-quality MR images using a least-squares <italic>GAN</italic> (<italic>LSGAN</italic>) to capture fine textures, combined with <inline-formula id="ieqn-532"><mml:math id="mml-ieqn-532"><mml:msub><mml:mi>&#x2113;</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:math></inline-formula>/<inline-formula id="ieqn-533"><mml:math id="mml-ieqn-533"><mml:msub><mml:mi>&#x2113;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:math></inline-formula> loss to suppress high-frequency noise. Expert radiologist ratings on a large paediatric contrast-enhanced <italic>MR</italic> dataset consistently preferred <italic>GANCS</italic> over traditional wavelet, dictionary learning, and pixel-wise deep learning methods. Diagnostic reliability was supported by normalized Radiologist Opinion Scores (<italic>ROS</italic>) for image quality, artifact presence, and sharpness, aligning with <italic>SSIM</italic> and <italic>SNR</italic> metrics. Additionally, a hybrid deep learning reconstruction strategy was proposed by integrating <inline-formula id="ieqn-534"><mml:math id="mml-ieqn-534"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> with projection onto convex sets (<italic>POCS</italic>) [<xref ref-type="bibr" rid="ref-196">196</xref>]. The initial <inline-formula id="ieqn-535"><mml:math id="mml-ieqn-535"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> output was refined using <italic>POCS</italic> and reused in a second training iteration to enhance performance. Radiologists evaluated five reconstruction methods (<inline-formula id="ieqn-536"><mml:math id="mml-ieqn-536"><mml:mi>U</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>N</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi></mml:math></inline-formula>, <italic>GAN</italic>, <inline-formula id="ieqn-537"><mml:math id="mml-ieqn-537"><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>, <inline-formula id="ieqn-538"><mml:math id="mml-ieqn-538"><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>f</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>, and <inline-formula id="ieqn-539"><mml:math id="mml-ieqn-539"><mml:mi>P</mml:mi><mml:mi>O</mml:mi><mml:mi>C</mml:mi><mml:mi>S</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula>) across brain and knee MRI datasets. <inline-formula id="ieqn-540"><mml:math id="mml-ieqn-540"><mml:mi>P</mml:mi><mml:mi>O</mml:mi><mml:mi>C</mml:mi><mml:mi>S</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mo>&#x2212;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mi>C</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:math></inline-formula> achieved the highest average scores for both knee (<inline-formula id="ieqn-541"><mml:math id="mml-ieqn-541"><mml:mn>3.3</mml:mn></mml:math></inline-formula> and <inline-formula id="ieqn-542"><mml:math id="mml-ieqn-542"><mml:mn>2.7</mml:mn></mml:math></inline-formula>) and brain <italic>MR</italic> images (<inline-formula id="ieqn-543"><mml:math id="mml-ieqn-543"><mml:mn>3.7</mml:mn></mml:math></inline-formula> and <inline-formula id="ieqn-544"><mml:math id="mml-ieqn-544"><mml:mn>4.05</mml:mn></mml:math></inline-formula>), consistently demonstrating superior image quality across different modalities and readers.</p>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Other Generative Models</title>
<p>This section provides an insight into various generative models, including auto-encoders and diffusion models, each tailored for specific tasks in data generation and representation learning. The auto-encoder [<xref ref-type="bibr" rid="ref-197">197</xref>] has encoder&#x2013;decoder network maps input data to a low-dimensional latent space, enabling the decoder to accurately reconstruct the input. This latent space also facilitates systematic analysis and manipulation of input properties, making the architecture essential for biomedical tasks like image reconstruction, data augmentation, and modality transfer. Diffusion models, a class of deep generative models [<xref ref-type="bibr" rid="ref-198">198</xref>], learn the prior probability distribution of images (e.g., brain <italic>PET</italic> or cardiac <italic>MRI</italic>) from training data and generate new samples by sampling from this distribution. Recently, they have emerged as state-of-the-art in generative modelling, producing higher-fidelity samples compared to auto-encoders and normalizing flows. The comparison between the generative models are shown in <xref ref-type="table" rid="table-13">Table 13</xref>. Diffusion models are extensively applied in medical image processing tasks such as reconstruction, registration, classification, image-to-image translation, segmentation, denoising, and image generation. A detailed explanation regarding this is discussed in the following sub-section.</p>
<table-wrap id="table-13">
<label>Table 13</label>
<caption>
<title>Comparison of generative models</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th>Feature</th>
<th>Auto-encoder</th>
<th><italic>GAN</italic></th>
<th>Diffusion model</th>
</tr>
</thead>
<tbody>
<tr>
<td><bold>Training stability</bold></td>
<td>Good</td>
<td>Poor</td>
<td>Good</td>
</tr>
<tr>
<td><bold>Image generation quality</bold></td>
<td>Medium</td>
<td>Good</td>
<td>Good</td>
</tr>
<tr>
<td><bold>Generation speed</bold></td>
<td>Fast</td>
<td>Fast</td>
<td>Slow</td>
</tr>
<tr>
<td><bold>Generation diversity</bold></td>
<td>Good at approximating diversity</td>
<td>Limited due to mode collapse</td>
<td>High, due to variability in input samples</td>
</tr>
<tr>
<td><bold>Generation technique</bold></td>
<td>One-shot</td>
<td>One-shot</td>
<td>Iterative</td>
</tr>
<tr>
<td><bold>Computational complexity</bold></td>
<td>Medium</td>
<td>High, due to training both generator and discriminator networks.</td>
<td>Very high due to iterative noise removal process</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><bold><italic>Diffusion Models</italic></bold></p>
<p>Emerging as promising alternatives to <inline-formula id="ieqn-545"><mml:math id="mml-ieqn-545"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, diffusion models are gaining attention in image generation tasks. These models operate on score-based methodologies, where training involves progressively adding Gaussian noise to images and learning the data distribution that underlies this transformation. Rather than directly reversing the noise, the goal is to accurately capture and reproduce the complex structure of the data distribution. Diffusion models are capable of producing highly realistic images, offering advantages such as stable training dynamics and comprehensive mode coverage. However, one of the primary limitations lies in the prolonged sampling time required during image generation. Although this limitation may be acceptable in fields like medical imaging, where real-time performance is not always essential, reducing this latency remains an active area of research. Current advancements aim to optimize diffusion models for faster inference without compromising image fidelity. This includes the development of new model architectures, improvements in computational efficiency, and enhancements in their suitability for time-sensitive applications. In the context of medical imaging, such innovations could significantly improve both the practicality and impact of diffusion-based approaches. <xref ref-type="fig" rid="fig-14">Fig. 14</xref> illustrates the process of diffusion models. Diffusion models operate through two distinct phases: the forward and reverse diffusion processes. In the forward phase, a Markov chain progressively corrupts the input data by adding Gaussian noise over a series of steps&#x2014;typically around <inline-formula id="ieqn-546"><mml:math id="mml-ieqn-546"><mml:mn>1000</mml:mn></mml:math></inline-formula> until the data resembles pure white noise. This phase is fixed and non-trainable. Conversely, the reverse phase is designed to gradually remove the added noise, effectively reconstructing the original data. This denoising process is guided by a neural network that is trained specifically to approximate the reverse of the forward diffusion.</p>
<fig id="fig-14">
<label>Figure 14</label>
<caption>
<title>Diffusion model [<xref ref-type="bibr" rid="ref-165">165</xref>]</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_67108-fig-14.tif"/>
</fig>
</sec>
<sec id="s6">
<label>6</label>
<title>Conclusion</title>
<p>Generative Adversarial Networks <inline-formula id="ieqn-547"><mml:math id="mml-ieqn-547"><mml:mo stretchy="false">(</mml:mo><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> have shown significant promise in transforming medical image analysis by enabling high-quality image synthesis, cross-modality translation, data augmentation, and aiding diagnostic tasks. Their capability to learn complex data distributions makes them valuable in scenarios where annotated medical data is limited or imbalanced. This study has explored the diverse applications of <inline-formula id="ieqn-548"><mml:math id="mml-ieqn-548"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> in the medical imaging domain, highlighting their role in generating synthetic data to overcome data limitations and improve the performance of diagnostic models. Furthermore, it outlines how <inline-formula id="ieqn-549"><mml:math id="mml-ieqn-549"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> contribute to modality translation, enabling the transformation of images between different imaging techniques. By identifying the advancements and the inherent limitations of <inline-formula id="ieqn-550"><mml:math id="mml-ieqn-550"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, this study provides a balanced view on the current landscape and future potential and existing research gaps of <italic>GAN</italic> in healthcare.</p>
<p>Despite the advancements in <inline-formula id="ieqn-551"><mml:math id="mml-ieqn-551"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula>, several challenges still hinder their full integration into clinical practice. Key difficulties include unstable training processes, lack of standard evaluation protocols for image generation, and the potential for generated images to introduce clinically misleading artifacts. Although <inline-formula id="ieqn-552"><mml:math id="mml-ieqn-552"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> have shown notable advancements in image quality metrics like <italic>SSIM</italic>, <italic>PSNR</italic>, and <italic>FID</italic>, it is crucial to separate these technical achievements from their actual clinical usefulness. Assertions about improved diagnostics or better treatment planning need to be backed by empirical research, including clinical trials, expert assessments, or regulatory benchmarks. Without such validation, enhancements in image realism or reconstruction quality should not be presumed to equate to clinical readiness. Ensuring that synthetic images retain diagnostic relevance and do not compromise patient safety remains a critical concern. Additionally, the opaque nature of <italic>GAN</italic> models limits interpretability, which is essential for medical decision making and clinician acceptance. Ethical issues, such as data misuse and patient privacy, further complicate the deployment of <italic>GAN</italic> generated data.</p>
<p>Looking to the future, research should focus on building more stable and explainable <italic>GAN</italic> architectures tailored specifically to medical applications. Developing robust validation frameworks that incorporate expert clinical feedback will be vital to ensuring safety and effectiveness. Future studies should focus on interpretability, reliability, and adherence to clinical and regulatory standards to confirm the practical applicability of <italic>GAN</italic> based tools in healthcare. There is also potential for <inline-formula id="ieqn-553"><mml:math id="mml-ieqn-553"><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi><mml:mi>s</mml:mi></mml:math></inline-formula> to work synergistically with other <italic>AI</italic> techniques, including self-supervised learning and multimodal models, to further improve diagnostic support systems.</p>
</sec>
<sec sec-type="supplementary-material" id="s7">
<title>Supplementary Materials</title>
<supplementary-material id="SD1">
<media xlink:href="CMES_67108-s001.docx"/>
</supplementary-material>
<supplementary-material id="SD2">
<media xlink:href="CMES_67108-s002.docx"/>
</supplementary-material></sec>
</body>
<back>
<ack>
<p>The authors extend their appreciation to the Deanship of Research and Graduate Studies at King Khalid University for funding this work through Large Research Project under grant number RGP2/540/46.</p>
</ack>
<sec>
<title>Funding Statement</title>
<p>The research was supported by Deanship of Research and Graduate Studies at King Khalid University for funding this work through Large Research Project under grant number RGP2/540/46.</p>
</sec>
<sec>
<title>Author Contributions</title>
<p>The authors confirm contribution to the paper as follows: Conceptualization, Sameera V. Mohd Sagheer and U. Nimitha; methodology, P. M. Ameer; software, Sameera V. Mohd Sagheer; validation, Sameera V. Mohd Sagheer, U. Nimitha and P. M. Ameer; formal analysis, P. M. Ameer; investigation, Muneer Parayangat; resources, Mohamed Abbas; writing&#x2014;original draft preparation, Sameera V. Mohd Sagheer; writing&#x2014;review and editing, U. Nimitha; visualization, P. M. Ameer; supervision, Muneer Parayangat; project administration, Mohamed Abbas; funding acquisition, Mohamed Abbas and Krishna Prakash Arunachalam. All authors reviewed the results and approved the final version of the manuscript.</p>
</sec>
<sec sec-type="data-availability">
<title>Availability of Data and Materials</title>
<p>Data openly available in a public repository.</p>
</sec>
<sec>
<title>Ethics Approval</title>
<p>Not applicable.</p>
</sec>
<sec sec-type="COI-statement">
<title>Conflicts of Interest</title>
<p>The authors declare no conflicts of interest to report regarding the present study.</p>
</sec>
<sec>
<title>Supplementary Materials</title>
<p>The supplementary material is available online at <ext-link ext-link-type="uri" xlink:href="https://www.techscience.com/doi/10.32604/cmes.2025.067108/s1">https://www.techscience.com/doi/10.32604/cmes.2025.067108/s1</ext-link>.</p>
</sec>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Yu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Feng</surname> <given-names>T</given-names></string-name>, <string-name><surname>Qiu</surname> <given-names>H</given-names></string-name>, <string-name><surname>Gu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Zuo</surname> <given-names>C</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Simultaneous photoacoustic and ultrasound imaging: a review</article-title>. <source>Ultrasonics</source>. <year>2024</year>;<volume>139</volume>:<fpage>107277</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.ultras.2024.107277</pub-id>; <pub-id pub-id-type="pmid">38460216</pub-id></mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Dangoury</surname> <given-names>S</given-names></string-name>, <string-name><surname>marghichi El</surname> <given-names>M</given-names></string-name>, <string-name><surname>Sadik</surname> <given-names>M</given-names></string-name>, <string-name><surname>Fail</surname> <given-names>A</given-names></string-name></person-group>. <article-title>The design of an efficient low-cost FPGA-based unit for generation ultrasound beamforming</article-title>. <source>Pertanika J Sci Technol</source>. <year>2023</year>;<volume>31</volume>(<issue>6</issue>):<fpage>3077</fpage>&#x2013;<lpage>92</lpage>. doi:<pub-id pub-id-type="doi">10.47836/pjst.31.6.24</pub-id>.</mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Ali</surname> <given-names>M</given-names></string-name>, <string-name><surname>Magee</surname> <given-names>D</given-names></string-name>, <string-name><surname>Dasgupta</surname> <given-names>U</given-names></string-name></person-group>. <chapter-title>Signal processing overview of ultrasound systems for medical imaging</chapter-title>. In: <source>White paper</source>. Vol. <volume>SPRAB12</volume>. <publisher-loc>Dallas, TX, USA</publisher-loc>: <publisher-name>Texas Instruments</publisher-name>; <year>2008</year>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Tao</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Tagare</surname> <given-names>HD</given-names></string-name>, <string-name><surname>Beaty</surname> <given-names>JD</given-names></string-name></person-group>. <article-title>Evaluation of four probability distribution models for speckle in clinical cardiac ultrasound images</article-title>. <source>IEEE Transact Med Imag</source>. <year>2006</year>;<volume>25</volume>(<issue>11</issue>):<fpage>1483</fpage>&#x2013;<lpage>91</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2006.881376</pub-id>; <pub-id pub-id-type="pmid">17117777</pub-id></mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Weng</surname> <given-names>L</given-names></string-name>, <string-name><surname>Reid</surname> <given-names>JM</given-names></string-name>, <string-name><surname>Shankar</surname> <given-names>PM</given-names></string-name>, <string-name><surname>Soetanto</surname> <given-names>K</given-names></string-name></person-group>. <article-title>Ultrasound speckle analysis based on the K distribution</article-title>. <source>J Acoust Soc America</source>. <year>1991</year>;<volume>89</volume>(<issue>6</issue>):<fpage>2992</fpage>&#x2013;<lpage>5</lpage>. doi:<pub-id pub-id-type="doi">10.1121/1.400818</pub-id>; <pub-id pub-id-type="pmid">1918635</pub-id></mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Sagheer</surname> <given-names>SVM</given-names></string-name>, <string-name><surname>George</surname> <given-names>SN</given-names></string-name></person-group>. <article-title>An approach for despeckling a sequence of ultrasound images based on statistical analysis</article-title>. <source>Sens Imag</source>. <year>2017</year>;<volume>18</volume>(<issue>1</issue>):<fpage>29</fpage>. doi:<pub-id pub-id-type="doi">10.1007/s11220-017-0181-8</pub-id>.</mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Henkelman</surname> <given-names>RM</given-names></string-name></person-group>. <article-title>Measurement of signal intensities in the presence of noise in MR images</article-title>. <source>Medical Physics</source>. <year>1985</year>;<volume>12</volume>(<issue>2</issue>):<fpage>232</fpage>&#x2013;<lpage>3</lpage>. doi:<pub-id pub-id-type="doi">10.1118/1.595711</pub-id>; <pub-id pub-id-type="pmid">4000083</pub-id></mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zeng</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhu</surname> <given-names>J</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>J</given-names></string-name>, <string-name><surname>Parasuraman</surname> <given-names>P</given-names></string-name>, <string-name><surname>Busi</surname> <given-names>S</given-names></string-name>, <string-name><surname>Nauli</surname> <given-names>SM</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Functional probes for cardiovascular molecular imaging</article-title>. <source>Quantit Ima Med Surg</source>. <year>2018</year>;<volume>8</volume>(<issue>8</issue>):<fpage>838</fpage>. doi:<pub-id pub-id-type="doi">10.21037/qims.2018.09.19</pub-id>; <pub-id pub-id-type="pmid">30306063</pub-id></mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Yang</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Li</surname> <given-names>N</given-names></string-name>, <string-name><surname>Zhao</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Fan</surname> <given-names>X</given-names></string-name>, <string-name><surname>Chang</surname> <given-names>EIC</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>Y</given-names></string-name></person-group>. <article-title>MRI cross-modality image-to-image translation</article-title>. <source>Scient Rep</source>. <year>2020</year>;<volume>10</volume>(<issue>1</issue>):<fpage>3753</fpage>. doi:<pub-id pub-id-type="doi">10.1038/s41598-020-60520-6</pub-id>; <pub-id pub-id-type="pmid">32111966</pub-id></mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Mredhula</surname> <given-names>L</given-names></string-name>, <string-name><surname>Dorairangasamy</surname> <given-names>M</given-names></string-name></person-group>. <article-title>An extensive review of significant researches on medical image denoising techniques</article-title>. <source>Int J Comput Applicat</source>. <year>2013</year>;<volume>64</volume>(<issue>14</issue>):<fpage>1</fpage>&#x2013;<lpage>12</lpage>. doi:<pub-id pub-id-type="doi">10.5120/10699-1551</pub-id>.</mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Ding</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Long</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>X</given-names></string-name>, <string-name><surname>Fessler</surname> <given-names>JA</given-names></string-name></person-group>. <article-title>Statistical image reconstruction using mixed Poisson-Gaussian noise model for X-ray CT</article-title>. <comment>arXiv:1801.09533. 2018</comment>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Borsdorf</surname> <given-names>A</given-names></string-name>, <string-name><surname>Raupach</surname> <given-names>R</given-names></string-name>, <string-name><surname>Flohr</surname> <given-names>T</given-names></string-name>, <string-name><surname>Hornegger</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Wavelet based noise reduction in CT-images using correlation analysis</article-title>. <source>IEEE Transact Med Imag</source>. <year>2008</year>;<volume>27</volume>(<issue>12</issue>):<fpage>1685</fpage>&#x2013;<lpage>703</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2008.923983</pub-id>; <pub-id pub-id-type="pmid">19033085</pub-id></mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Rahiman</surname> <given-names>MF</given-names></string-name>, <string-name><surname>Rahim</surname> <given-names>RA</given-names></string-name>, <string-name><surname>Zakaria</surname> <given-names>Z</given-names></string-name></person-group>. <article-title>Design and modelling of ultrasonic tomography for two-component high-acoustic impedance mixture</article-title>. <source>Sens Actuat A: Phys</source>. <year>2008</year>;<volume>147</volume>(<issue>2</issue>):<fpage>409</fpage>&#x2013;<lpage>14</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.sna.2008.05.024</pub-id>.</mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Chen</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Shi</surname> <given-names>L</given-names></string-name>, <string-name><surname>Feng</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>J</given-names></string-name>, <string-name><surname>Shu</surname> <given-names>H</given-names></string-name>, <string-name><surname>Luo</surname> <given-names>L</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Artifact suppressed dictionary learning for low-dose CT image processing</article-title>. <source>IEEE Transact Med Imag</source>. <year>2014</year>;<volume>33</volume>(<issue>12</issue>):<fpage>2271</fpage>&#x2013;<lpage>92</lpage>. doi:<pub-id pub-id-type="doi">10.1109/isbi.2014.6868073</pub-id>.</mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ollinger</surname> <given-names>J</given-names></string-name>, <string-name><surname>Fessler</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Positron-emission tomography</article-title>. <source>IEEE Signal Process Mag</source>. <year>1997</year>;<volume>14</volume>(<issue>1</issue>):<fpage>43</fpage>&#x2013;<lpage>55</lpage>. doi:<pub-id pub-id-type="doi">10.1109/79.560323</pub-id>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Coxson</surname> <given-names>PG</given-names></string-name></person-group>. <article-title>Consequences of using a simplified kinetic model for dynamic PET data</article-title>. <source>J Nucl Med</source>. <year>1995</year>;<volume>38</volume>(<issue>4</issue>):<fpage>660</fpage>&#x2013;<lpage>7</lpage>.</mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Abdallah</surname> <given-names>NG</given-names></string-name>, <string-name><surname>Rashdan</surname> <given-names>M</given-names></string-name>, <string-name><surname>Khalaf</surname> <given-names>AA</given-names></string-name></person-group>. <article-title>High resolution time-to-digital converter for pet imaging</article-title>. In: <conf-name>International Conference on Innovative Trends in Communication and Computer Engineering (ITCE); 2020 Feb 8&#x2013;9</conf-name>; <publisher-loc>Aswan, Egypt</publisher-loc>. p. <fpage>295</fpage>&#x2013;<lpage>8</lpage>.</mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Shabbir</surname> <given-names>J</given-names></string-name>, <string-name><surname>Anwer</surname> <given-names>T</given-names></string-name></person-group>. <article-title>Artificial intelligence and its role in near future</article-title>. <comment>arXiv:1804.01396. 2018</comment>.</mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Bengio</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Courville</surname> <given-names>A</given-names></string-name>, <string-name><surname>Vincent</surname> <given-names>P</given-names></string-name></person-group>. <article-title>Representation learning: a review and new perspectives</article-title>. <source>IEEE Transact Patt Analy Mach Intell</source>. <year>2013</year>;<volume>35</volume>(<issue>8</issue>):<fpage>1798</fpage>&#x2013;<lpage>828</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tpami.2013.50</pub-id>; <pub-id pub-id-type="pmid">23787338</pub-id></mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Zhang</surname> <given-names>C</given-names></string-name>, <string-name><surname>Bengio</surname> <given-names>S</given-names></string-name>, <string-name><surname>Hardt</surname> <given-names>M</given-names></string-name>, <string-name><surname>Recht</surname> <given-names>B</given-names></string-name>, <string-name><surname>Vinyals</surname> <given-names>O</given-names></string-name></person-group>. <article-title>Understanding deep learning requires rethinking generalization</article-title>. <comment>arXiv:1611.03530. 2016</comment>.</mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Schmidhuber</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Deep learning in neural networks: an overview</article-title>. <source>Neur Netw</source>. <year>2015</year>;<volume>61</volume>(<issue>3</issue>):<fpage>85</fpage>&#x2013;<lpage>117</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.neunet.2014.09.003</pub-id>; <pub-id pub-id-type="pmid">25462637</pub-id></mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Talaei Khoei</surname> <given-names>T</given-names></string-name>, <string-name><surname>Ould Slimane</surname> <given-names>H</given-names></string-name>, <string-name><surname>Kaabouch</surname> <given-names>N</given-names></string-name></person-group>. <article-title>Deep learning: systematic review, models, challenges, and research directions</article-title>. <source>Neu Comput Applicat</source>. <year>2023</year>;<volume>35</volume>(<issue>31</issue>):<fpage>23103</fpage>&#x2013;<lpage>24</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s00521-023-08957-4</pub-id>.</mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Pan</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Yu</surname> <given-names>W</given-names></string-name>, <string-name><surname>Yi</surname> <given-names>X</given-names></string-name>, <string-name><surname>Khan</surname> <given-names>A</given-names></string-name>, <string-name><surname>Yuan</surname> <given-names>F</given-names></string-name>, <string-name><surname>Zheng</surname> <given-names>Y</given-names></string-name></person-group>. <article-title>Recent progress on generative adversarial networks (GANs): a survey</article-title>. <source>IEEE Access</source>. <year>2019</year>;<volume>7</volume>:<fpage>36322</fpage>&#x2013;<lpage>33</lpage>. doi:<pub-id pub-id-type="doi">10.1109/access.2019.2905015</pub-id>.</mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Aljuaid</surname> <given-names>A</given-names></string-name>, <string-name><surname>Anwar</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Survey of supervised learning for medical image processing</article-title>. <source>SN Comput Sci</source>. <year>2022</year>;<volume>3</volume>(<issue>4</issue>):<fpage>292</fpage>. doi:<pub-id pub-id-type="doi">10.1007/s42979-022-01166-1</pub-id>; <pub-id pub-id-type="pmid">35602289</pub-id></mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Rani</surname> <given-names>V</given-names></string-name>, <string-name><surname>Kumar</surname> <given-names>M</given-names></string-name>, <string-name><surname>Gupta</surname> <given-names>A</given-names></string-name>, <string-name><surname>Sachdeva</surname> <given-names>M</given-names></string-name>, <string-name><surname>Mittal</surname> <given-names>A</given-names></string-name>, <string-name><surname>Kumar</surname> <given-names>K</given-names></string-name></person-group>. <article-title>Self-supervised learning for medical image analysis: a comprehensive review</article-title>. <source>Evolv Syst</source>. <year>2024</year>;<volume>15</volume>(<issue>4</issue>):<fpage>1607</fpage>&#x2013;<lpage>33</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s12530-024-09581-w</pub-id>.</mixed-citation></ref>
<ref id="ref-26"><label>[26]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Halevy</surname> <given-names>A</given-names></string-name>, <string-name><surname>Norvig</surname> <given-names>P</given-names></string-name>, <string-name><surname>Pereira</surname> <given-names>F</given-names></string-name></person-group>. <article-title>The unreasonable effectiveness of data</article-title>. <source>IEEE Intell Syst</source>. <year>2009</year>;<volume>24</volume>(<issue>2</issue>):<fpage>8</fpage>&#x2013;<lpage>12</lpage>. doi:<pub-id pub-id-type="doi">10.1109/mis.2009.36</pub-id>.</mixed-citation></ref>
<ref id="ref-27"><label>[27]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Raza</surname> <given-names>K</given-names></string-name>, <string-name><surname>Singh</surname> <given-names>NK</given-names></string-name></person-group>. <article-title>A tour of unsupervised deep learning for medical image analysis</article-title>. <source>Current Med Imag Rev</source>. <year>2021</year>;<volume>17</volume>(<issue>9</issue>):<fpage>1059</fpage>&#x2013;<lpage>77</lpage>. doi:<pub-id pub-id-type="doi">10.2174/1573405617666210127154257</pub-id>; <pub-id pub-id-type="pmid">33504314</pub-id></mixed-citation></ref>
<ref id="ref-28"><label>[28]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Ganeshkumar</surname> <given-names>M</given-names></string-name>, <string-name><surname>Sowmya</surname> <given-names>V</given-names></string-name>, <string-name><surname>Gopalakrishnan</surname> <given-names>E</given-names></string-name>, <string-name><surname>Soman</surname> <given-names>K</given-names></string-name></person-group>. <chapter-title>Unsupervised deep learning-based disease diagnosis using medical images</chapter-title>. In: <source>Cognitive and soft computing techniques for the analysis of healthcare data</source>. <publisher-loc>Amsterdam, The Netherlands</publisher-loc>: <publisher-name>Elsevier</publisher-name>; <year>2022</year>. p. <fpage>203</fpage>&#x2013;<lpage>20</lpage>. doi:<pub-id pub-id-type="doi">10.1016/b978-0-323-85751-2.00011-6</pub-id>.</mixed-citation></ref>
<ref id="ref-29"><label>[29]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Croitoru</surname> <given-names>FA</given-names></string-name>, <string-name><surname>Hondru</surname> <given-names>V</given-names></string-name>, <string-name><surname>Ionescu</surname> <given-names>RT</given-names></string-name>, <string-name><surname>Shah</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Diffusion models in vision: a survey</article-title>. <source>IEEE Transact Pattern Anal Mach Intell</source>. <year>2023</year>;<volume>45</volume>(<issue>9</issue>):<fpage>10850</fpage>&#x2013;<lpage>69</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tpami.2023.3261988</pub-id>; <pub-id pub-id-type="pmid">37030794</pub-id></mixed-citation></ref>
<ref id="ref-30"><label>[30]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Goodfellow</surname> <given-names>I</given-names></string-name>, <string-name><surname>Pouget-Abadie</surname> <given-names>J</given-names></string-name>, <string-name><surname>Mirza</surname> <given-names>M</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>B</given-names></string-name>, <string-name><surname>Warde-Farley</surname> <given-names>D</given-names></string-name>, <string-name><surname>Ozair</surname> <given-names>S</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Generative adversarial networks</article-title>. <source>Communicat ACM</source>. <year>2020</year>;<volume>63</volume>(<issue>11</issue>):<fpage>139</fpage>&#x2013;<lpage>44</lpage>. doi:<pub-id pub-id-type="doi">10.1145/3422622</pub-id>.</mixed-citation></ref>
<ref id="ref-31"><label>[31]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>De Souza</surname> <given-names>VLT</given-names></string-name>, <string-name><surname>Marques</surname> <given-names>BAD</given-names></string-name>, <string-name><surname>Batagelo</surname> <given-names>HC</given-names></string-name>, <string-name><surname>Gois</surname> <given-names>JP</given-names></string-name></person-group>. <article-title>A review on generative adversarial networks for image generation</article-title>. <source>Comput Graph</source>. <year>2023</year>;<volume>114</volume>:<fpage>13</fpage>&#x2013;<lpage>25</lpage>.</mixed-citation></ref>
<ref id="ref-32"><label>[32]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Gui</surname> <given-names>J</given-names></string-name>, <string-name><surname>Sun</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Wen</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Tao</surname> <given-names>D</given-names></string-name>, <string-name><surname>Ye</surname> <given-names>J</given-names></string-name></person-group>. <article-title>A review on generative adversarial networks: algorithms, theory, and applications</article-title>. <source>IEEE Transact Know Data Eng</source>. <year>2023</year>;<volume>35</volume>(<issue>4</issue>):<fpage>3313</fpage>&#x2013;<lpage>32</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tkde.2021.3130191</pub-id>.</mixed-citation></ref>
<ref id="ref-33"><label>[33]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Krichen</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Generative adversarial networks</article-title>. In: <conf-name>2023 14th International Conference on Computing Communication and Networking Technologies (ICCCNT); 2023 Jul 6&#x2013;8</conf-name>; <publisher-loc>Delhi, India</publisher-loc>. p. <fpage>1</fpage>&#x2013;<lpage>7</lpage>.</mixed-citation></ref>
<ref id="ref-34"><label>[34]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Saxena</surname> <given-names>D</given-names></string-name>, <string-name><surname>Cao</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Generative adversarial networks (GANs) challenges, solutions, and future directions</article-title>. <source>ACM Comput Surv (CSUR)</source>. <year>2021</year>;<volume>54</volume>(<issue>3</issue>):<fpage>1</fpage>&#x2013;<lpage>42</lpage>. doi:<pub-id pub-id-type="doi">10.1145/3446374</pub-id>.</mixed-citation></ref>
<ref id="ref-35"><label>[35]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Aggarwal</surname> <given-names>A</given-names></string-name>, <string-name><surname>Mittal</surname> <given-names>M</given-names></string-name>, <string-name><surname>Battineni</surname> <given-names>G</given-names></string-name></person-group>. <article-title>Generative adversarial network: an overview of theory and applications</article-title>. <source>Int J Inform Manag Data Insights</source>. <year>2021</year>;<volume>1</volume>(<issue>1</issue>):<fpage>100004</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.jjimei.2020.100004</pub-id>.</mixed-citation></ref>
<ref id="ref-36"><label>[36]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Creswell</surname> <given-names>A</given-names></string-name>, <string-name><surname>White</surname> <given-names>T</given-names></string-name>, <string-name><surname>Dumoulin</surname> <given-names>V</given-names></string-name>, <string-name><surname>Arulkumaran</surname> <given-names>K</given-names></string-name>, <string-name><surname>Sengupta</surname> <given-names>B</given-names></string-name>, <string-name><surname>Bharath</surname> <given-names>AA</given-names></string-name></person-group>. <article-title>Generative adversarial networks: an overview</article-title>. <source>IEEE Signal Process Magaz</source>. <year>2018</year>;<volume>35</volume>(<issue>1</issue>):<fpage>53</fpage>&#x2013;<lpage>65</lpage>. doi:<pub-id pub-id-type="doi">10.1109/msp.2017.2765202</pub-id>.</mixed-citation></ref>
<ref id="ref-37"><label>[37]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Bisht</surname> <given-names>A</given-names></string-name>, <string-name><surname>Rawat</surname> <given-names>K</given-names></string-name>, <string-name><surname>Bhati</surname> <given-names>JP</given-names></string-name>, <string-name><surname>Vats</surname> <given-names>S</given-names></string-name>, <string-name><surname>Sharma</surname> <given-names>V</given-names></string-name>, <string-name><surname>Singh</surname> <given-names>S</given-names></string-name></person-group>. <article-title>Generating images using vanilla generative adversarial networks</article-title>. In: <conf-name>2024 4th International Conference on Technological Advancements in Computational Sciences (ICTACS). IEEE; 2024 Nov 13&#x2013;15</conf-name>; <publisher-loc>Tashkent, Uzbekistan</publisher-loc>: <publisher-name>IEEE</publisher-name>. p. <fpage>463</fpage>&#x2013;<lpage>8</lpage>.</mixed-citation></ref>
<ref id="ref-38"><label>[38]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Radford</surname> <given-names>A</given-names></string-name>, <string-name><surname>Metz</surname> <given-names>L</given-names></string-name>, <string-name><surname>Chintala</surname> <given-names>S</given-names></string-name></person-group>. <article-title>Unsupervised representation learning with deep convolutional generative adversarial networks</article-title>. <comment>arXiv:1511.06434. 2015</comment>.</mixed-citation></ref>
<ref id="ref-39"><label>[39]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kim</surname> <given-names>HJ</given-names></string-name>, <string-name><surname>Lee</surname> <given-names>D</given-names></string-name></person-group>. <article-title>Image denoising with conditional generative adversarial networks (CGAN) in low dose chest images</article-title>. <source>Nuclear Instrum Meth Phy Res Sec A Accel Spectrom Detect Assoc Equip</source>. <year>2020</year>;<volume>954</volume>(<issue>8</issue>):<fpage>161914</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.nima.2019.02.041</pub-id>.</mixed-citation></ref>
<ref id="ref-40"><label>[40]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Gulrajani</surname> <given-names>I</given-names></string-name>, <string-name><surname>Ahmed</surname> <given-names>F</given-names></string-name>, <string-name><surname>Arjovsky</surname> <given-names>M</given-names></string-name>, <string-name><surname>Dumoulin</surname> <given-names>V</given-names></string-name>, <string-name><surname>Courville</surname> <given-names>AC</given-names></string-name></person-group>. <chapter-title>Improved training of wasserstein gans</chapter-title>. In: <source>Advances in neural information processing systems</source>. <publisher-loc>Cambridge, MA, USA</publisher-loc>: <publisher-name>The MIT Press</publisher-name>; <year>2017</year>.</mixed-citation></ref>
<ref id="ref-41"><label>[41]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Chu</surname> <given-names>C</given-names></string-name>, <string-name><surname>Zhmoginov</surname> <given-names>A</given-names></string-name>, <string-name><surname>Sandler</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Cyclegan, a master of steganography</article-title>. <comment>arXiv:1712.02950. 2017</comment>.</mixed-citation></ref>
<ref id="ref-42"><label>[42]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Goodfellow</surname> <given-names>I</given-names></string-name>, <string-name><surname>Pouget-Abadie</surname> <given-names>J</given-names></string-name>, <string-name><surname>Mirza</surname> <given-names>M</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>B</given-names></string-name>, <string-name><surname>Warde-Farley</surname> <given-names>D</given-names></string-name>, <string-name><surname>Ozair</surname> <given-names>S</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Generative adversarial</article-title>. <comment>arXiv:1406.2661. 2014</comment>.</mixed-citation></ref>
<ref id="ref-43"><label>[43]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Karras</surname> <given-names>T</given-names></string-name>, <string-name><surname>Aila</surname> <given-names>T</given-names></string-name>, <string-name><surname>Laine</surname> <given-names>S</given-names></string-name>, <string-name><surname>Lehtinen</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Progressive growing of gans for improved quality, stability, and variation</article-title>. <comment>arXiv:1710.10196. 2017</comment>.</mixed-citation></ref>
<ref id="ref-44"><label>[44]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Motamed</surname> <given-names>S</given-names></string-name>, <string-name><surname>Rogalla</surname> <given-names>P</given-names></string-name>, <string-name><surname>Khalvati</surname> <given-names>F</given-names></string-name></person-group>. <article-title>RANDGAN: randomized generative adversarial network for detection of COVID-19 in chest X-ray</article-title>. <source>Scient Rep</source>. <year>2021</year>;<volume>11</volume>(<issue>1</issue>):<fpage>8602</fpage>. doi:<pub-id pub-id-type="doi">10.1038/s41598-021-87994-2</pub-id>; <pub-id pub-id-type="pmid">33883609</pub-id></mixed-citation></ref>
<ref id="ref-45"><label>[45]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kossen</surname> <given-names>T</given-names></string-name>, <string-name><surname>Subramaniam</surname> <given-names>P</given-names></string-name>, <string-name><surname>Madai</surname> <given-names>VI</given-names></string-name>, <string-name><surname>Hennemuth</surname> <given-names>A</given-names></string-name>, <string-name><surname>Hildebrand</surname> <given-names>K</given-names></string-name>, <string-name><surname>Hilbert</surname> <given-names>A</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Synthesizing anonymized and labeled TOF-MRA patches for brain vessel segmentation using generative adversarial networks</article-title>. <source>Comput Biol Med</source>. <year>2021</year>;<volume>131</volume>:<fpage>104254</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.compbiomed.2021.104254</pub-id>; <pub-id pub-id-type="pmid">33618105</pub-id></mixed-citation></ref>
<ref id="ref-46"><label>[46]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ahmad</surname> <given-names>B</given-names></string-name>, <string-name><surname>Sun</surname> <given-names>J</given-names></string-name>, <string-name><surname>You</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Palade</surname> <given-names>V</given-names></string-name>, <string-name><surname>Mao</surname> <given-names>Z</given-names></string-name></person-group>. <article-title>Brain tumor classification using a combination of variational autoencoders and generative adversarial networks</article-title>. <source>Biomedicines</source>. <year>2022</year>;<volume>10</volume>(<issue>2</issue>):<fpage>223</fpage>. doi:<pub-id pub-id-type="doi">10.3390/biomedicines10020223</pub-id>; <pub-id pub-id-type="pmid">35203433</pub-id></mixed-citation></ref>
<ref id="ref-47"><label>[47]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Krithika alias Anbu Devi</surname> <given-names>M</given-names></string-name>, <string-name><surname>Suganthi</surname> <given-names>K</given-names></string-name></person-group>. <article-title>Review of medical image synthesis using GAN techniques</article-title>. In: <conf-name>ITM Web of Conferences</conf-name>. Vol. 37. <publisher-loc>Les Ulis, France</publisher-loc>: <publisher-name>EDP Sciences</publisher-name>; <year>2021</year>. <fpage>01005</fpage> p.</mixed-citation></ref>
<ref id="ref-48"><label>[48]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zhou</surname> <given-names>T</given-names></string-name>, <string-name><surname>Li</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Lu</surname> <given-names>H</given-names></string-name>, <string-name><surname>Cheng</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>X</given-names></string-name></person-group>. <article-title>GAN review: models and medical image fusion applications</article-title>. <source>Inform Fus</source>. <year>2023</year>;<volume>91</volume>(<issue>11</issue>):<fpage>134</fpage>&#x2013;<lpage>48</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.inffus.2022.10.017</pub-id>.</mixed-citation></ref>
<ref id="ref-49"><label>[49]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ma</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>J</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Fu</surname> <given-names>H</given-names></string-name>, <string-name><surname>Hu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Cheng</surname> <given-names>J</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Structure and illumination constrained GAN for medical image enhancement</article-title>. <source>IEEE Transact Med Imag</source>. <year>2021</year>;<volume>40</volume>(<issue>12</issue>):<fpage>3955</fpage>&#x2013;<lpage>67</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2021.3101937</pub-id>; <pub-id pub-id-type="pmid">34339369</pub-id></mixed-citation></ref>
<ref id="ref-50"><label>[50]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Tightiz</surname> <given-names>L</given-names></string-name>, <string-name><surname>Jung</surname> <given-names>MH</given-names></string-name>, <string-name><surname>Song</surname> <given-names>I</given-names></string-name>, <string-name><surname>Lee</surname> <given-names>K</given-names></string-name></person-group>. <article-title>Trustworthy TAVR navigator system, I: a generative adversarial network-driven medical twin approach for Post-TAVR pacemaker implantation prediction</article-title>. <source>Expert Syst Applicat</source>. <year>2025</year>;<volume>275</volume>(<issue>1</issue>):<fpage>126973</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.eswa.2025.126973</pub-id>.</mixed-citation></ref>
<ref id="ref-51"><label>[51]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kazeminia</surname> <given-names>S</given-names></string-name>, <string-name><surname>Baur</surname> <given-names>C</given-names></string-name>, <string-name><surname>Kuijper</surname> <given-names>A</given-names></string-name>, <string-name><surname>van Ginneken</surname> <given-names>B</given-names></string-name>, <string-name><surname>Navab</surname> <given-names>N</given-names></string-name>, <string-name><surname>Albarqouni</surname> <given-names>S</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>GANs for medical image analysis</article-title>. <source>Artif Intelli Med</source>. <year>2020</year>;<volume>109</volume>(<issue>12</issue>):<fpage>101938</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.artmed.2020.101938</pub-id>; <pub-id pub-id-type="pmid">34756215</pub-id></mixed-citation></ref>
<ref id="ref-52"><label>[52]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Fu</surname> <given-names>B</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>X</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>L</given-names></string-name>, <string-name><surname>Ren</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Thanh</surname> <given-names>DN</given-names></string-name></person-group>. <article-title>A blind medical image denoising method with noise generation network</article-title>. <source>J X-Ray Sci Technol</source>. <year>2022</year>;<volume>30</volume>(<issue>3</issue>):<fpage>531</fpage>&#x2013;<lpage>47</lpage>. doi:<pub-id pub-id-type="doi">10.3233/xst-211098</pub-id>; <pub-id pub-id-type="pmid">35253724</pub-id></mixed-citation></ref>
<ref id="ref-53"><label>[53]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Li</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>K</given-names></string-name>, <string-name><surname>Shi</surname> <given-names>W</given-names></string-name>, <string-name><surname>Miao</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Jiang</surname> <given-names>Z</given-names></string-name></person-group>. <article-title>A novel medical image denoising method based on conditional generative adversarial network</article-title>. <source>Computat Mathem Meth Med</source>. <year>2021</year>;<volume>2021</volume>(<issue>2</issue>):<fpage>9974017</fpage>. doi:<pub-id pub-id-type="doi">10.1155/2021/9974017</pub-id>; <pub-id pub-id-type="pmid">34621329</pub-id></mixed-citation></ref>
<ref id="ref-54"><label>[54]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Geng</surname> <given-names>M</given-names></string-name>, <string-name><surname>Meng</surname> <given-names>X</given-names></string-name>, <string-name><surname>Yu</surname> <given-names>J</given-names></string-name>, <string-name><surname>Zhu</surname> <given-names>L</given-names></string-name>, <string-name><surname>Jin</surname> <given-names>L</given-names></string-name>, <string-name><surname>Jiang</surname> <given-names>Z</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Content-noise complementary learning for medical image denoising</article-title>. <source>IEEE Transact Med Imag</source>. <year>2021</year>;<volume>41</volume>(<issue>2</issue>):<fpage>407</fpage>&#x2013;<lpage>19</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2021.3113365</pub-id>; <pub-id pub-id-type="pmid">34529565</pub-id></mixed-citation></ref>
<ref id="ref-55"><label>[55]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Gondara</surname> <given-names>L</given-names></string-name></person-group>. <article-title>Medical image denoising using convolutional denoising autoencoders</article-title>. In: <conf-name>2016 IEEE 16th International Conference on Data Mining Workshops (ICDMW); 2016 Dec 12&#x2013;15</conf-name>; <publisher-loc>Barcelona, Spain</publisher-loc>. p. <fpage>241</fpage>&#x2013;<lpage>6</lpage>.</mixed-citation></ref>
<ref id="ref-56"><label>[56]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Xue</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>T</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>H</given-names></string-name>, <string-name><surname>Long</surname> <given-names>LR</given-names></string-name>, <string-name><surname>Huang</surname> <given-names>X</given-names></string-name></person-group>. <article-title>SegAN: adversarial network with multi-scale L1 loss for medical image segmentation</article-title>. <source>Neuroinformatics</source>. <year>2018</year>;<volume>16</volume>(<issue>3&#x2013;4</issue>):<fpage>383</fpage>&#x2013;<lpage>92</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s12021-018-9377-x</pub-id>; <pub-id pub-id-type="pmid">29725916</pub-id></mixed-citation></ref>
<ref id="ref-57"><label>[57]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Han</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Wei</surname> <given-names>B</given-names></string-name>, <string-name><surname>Mercado</surname> <given-names>A</given-names></string-name>, <string-name><surname>Leung</surname> <given-names>S</given-names></string-name>, <string-name><surname>Li</surname> <given-names>S</given-names></string-name></person-group>. <article-title>Spine-GAN: semantic segmentation of multiple spinal structures</article-title>. <source>Med Image Anal</source>. <year>2018</year>;<volume>50</volume>(<issue>1</issue>):<fpage>23</fpage>&#x2013;<lpage>35</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.media.2018.08.005</pub-id>; <pub-id pub-id-type="pmid">30176546</pub-id></mixed-citation></ref>
<ref id="ref-58"><label>[58]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Trinh</surname> <given-names>DH</given-names></string-name>, <string-name><surname>Luong</surname> <given-names>M</given-names></string-name>, <string-name><surname>Dibos</surname> <given-names>F</given-names></string-name>, <string-name><surname>Rocchisani</surname> <given-names>JM</given-names></string-name>, <string-name><surname>Pham</surname> <given-names>CD</given-names></string-name>, <string-name><surname>Nguyen</surname> <given-names>TQ</given-names></string-name></person-group>. <article-title>Novel example-based method for super-resolution and denoising of medical images</article-title>. <source>IEEE Transact Image Process</source>. <year>2014</year>;<volume>23</volume>(<issue>4</issue>):<fpage>1882</fpage>&#x2013;<lpage>95</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tip.2014.2308422</pub-id>; <pub-id pub-id-type="pmid">24808354</pub-id></mixed-citation></ref>
<ref id="ref-59"><label>[59]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>S&#x00E1;nchez</surname> <given-names>I</given-names></string-name>, <string-name><surname>Vilaplana</surname> <given-names>V</given-names></string-name></person-group>. <article-title>Brain MRI super-resolution using 3D generative adversarial networks</article-title>. <comment>arXiv:1812.11440. 2018</comment>.</mixed-citation></ref>
<ref id="ref-60"><label>[60]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>TC</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>MY</given-names></string-name>, <string-name><surname>Zhu</surname> <given-names>JY</given-names></string-name>, <string-name><surname>Tao</surname> <given-names>A</given-names></string-name>, <string-name><surname>Kautz</surname> <given-names>J</given-names></string-name>, <string-name><surname>Catanzaro</surname> <given-names>B</given-names></string-name></person-group>. <article-title>High-resolution image synthesis and semantic manipulation with conditional gans</article-title>. In: <conf-name>Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition; 2018 Jun 18&#x2013;23</conf-name>; <publisher-loc>Salt Lake City, UT, USA</publisher-loc>. p. <fpage>8798</fpage>&#x2013;<lpage>807</lpage>.</mixed-citation></ref>
<ref id="ref-61"><label>[61]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Dar</surname> <given-names>SU</given-names></string-name>, <string-name><surname>Yurt</surname> <given-names>M</given-names></string-name>, <string-name><surname>Karacan</surname> <given-names>L</given-names></string-name>, <string-name><surname>Erdem</surname> <given-names>A</given-names></string-name>, <string-name><surname>Erdem</surname> <given-names>E</given-names></string-name>, <string-name><surname>Cukur</surname> <given-names>T</given-names></string-name></person-group>. <article-title>Image synthesis in multi-contrast MRI with conditional generative adversarial networks</article-title>. <source>IEEE Transact Med Imag</source>. <year>2019</year>;<volume>38</volume>(<issue>10</issue>):<fpage>2375</fpage>&#x2013;<lpage>88</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2019.2901750</pub-id>; <pub-id pub-id-type="pmid">30835216</pub-id></mixed-citation></ref>
<ref id="ref-62"><label>[62]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Platscher</surname> <given-names>M</given-names></string-name>, <string-name><surname>Zopes</surname> <given-names>J</given-names></string-name>, <string-name><surname>Federau</surname> <given-names>C</given-names></string-name></person-group>. <article-title>Image translation for medical image generation: ischemic stroke lesion segmentation</article-title>. <source>Biomed Signal Process Cont</source>. <year>2022</year>;<volume>72</volume>(<issue>4</issue>):<fpage>103283</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.bspc.2021.103283</pub-id>.</mixed-citation></ref>
<ref id="ref-63"><label>[63]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Upadhyay</surname> <given-names>U</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Hepp</surname> <given-names>T</given-names></string-name>, <string-name><surname>Gatidis</surname> <given-names>S</given-names></string-name>, <string-name><surname>Akata</surname> <given-names>Z</given-names></string-name></person-group>. <article-title>Uncertainty-guided progressive GANs for medical image translation</article-title>. In: <conf-name>Medical Image Computing and Computer Assisted Intervention-MICCAI 2021: 24th International Conference</conf-name>. <publisher-loc>Cham, Swizterland</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>2021</year>. p. <fpage>614</fpage>&#x2013;<lpage>24</lpage>.</mixed-citation></ref>
<ref id="ref-64"><label>[64]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Han</surname> <given-names>C</given-names></string-name>, <string-name><surname>Hayashi</surname> <given-names>H</given-names></string-name>, <string-name><surname>Rundo</surname> <given-names>L</given-names></string-name>, <string-name><surname>Araki</surname> <given-names>R</given-names></string-name>, <string-name><surname>Shimoda</surname> <given-names>W</given-names></string-name>, <string-name><surname>Muramatsu</surname> <given-names>S</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>GAN-based synthetic brain MR image generation</article-title>. In: <conf-name>IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018); 2018 Apr 4&#x2013;7</conf-name>; <publisher-loc>Washington, DC, USA</publisher-loc>; <year>2018</year>. p. <fpage>734</fpage>&#x2013;<lpage>8</lpage>.</mixed-citation></ref>
<ref id="ref-65"><label>[65]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Singh</surname> <given-names>NK</given-names></string-name>, <string-name><surname>Raza</surname> <given-names>K</given-names></string-name></person-group>. <source>Medical image generation using generative adversarial networks: a review. In: Health informatics: a computational perspective in healthcare</source>. <publisher-loc>Singapore</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>2021</year>. p. <fpage>77</fpage>&#x2013;<lpage>96</lpage>.</mixed-citation></ref>
<ref id="ref-66"><label>[66]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Fan</surname> <given-names>C</given-names></string-name>, <string-name><surname>Lin</surname> <given-names>H</given-names></string-name>, <string-name><surname>Qiu</surname> <given-names>Y</given-names></string-name></person-group>. <article-title>U-Patch GAN: a medical image fusion method based on GAN</article-title>. <source>J Digit Imag</source>. <year>2023</year>;<volume>36</volume>(<issue>1</issue>):<fpage>339</fpage>&#x2013;<lpage>55</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s10278-022-00696-7</pub-id>; <pub-id pub-id-type="pmid">36038702</pub-id></mixed-citation></ref>
<ref id="ref-67"><label>[67]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Guo</surname> <given-names>K</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>J</given-names></string-name>, <string-name><surname>Qiu</surname> <given-names>T</given-names></string-name>, <string-name><surname>Guo</surname> <given-names>S</given-names></string-name>, <string-name><surname>Luo</surname> <given-names>T</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>T</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>MedGAN: an adaptive GAN approach for medical image generation</article-title>. <source>Comput Biol Med</source>. <year>2023</year>;<volume>163</volume>(<issue>7</issue>):<fpage>107119</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.compbiomed.2023.107119</pub-id>; <pub-id pub-id-type="pmid">37364533</pub-id></mixed-citation></ref>
<ref id="ref-68"><label>[68]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Chlap</surname> <given-names>P</given-names></string-name>, <string-name><surname>Min</surname> <given-names>H</given-names></string-name>, <string-name><surname>Vandenberg</surname> <given-names>N</given-names></string-name>, <string-name><surname>Dowling</surname> <given-names>J</given-names></string-name>, <string-name><surname>Holloway</surname> <given-names>L</given-names></string-name>, <string-name><surname>Haworth</surname> <given-names>A</given-names></string-name></person-group>. <article-title>A review of medical image data augmentation techniques for deep learning applications</article-title>. <source>J Med Imag Radiat Oncol</source>. <year>2021</year>;<volume>65</volume>(<issue>5</issue>):<fpage>545</fpage>&#x2013;<lpage>63</lpage>. doi:<pub-id pub-id-type="doi">10.1111/1754-9485.13261</pub-id>; <pub-id pub-id-type="pmid">34145766</pub-id></mixed-citation></ref>
<ref id="ref-69"><label>[69]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Garcea</surname> <given-names>F</given-names></string-name>, <string-name><surname>Serra</surname> <given-names>A</given-names></string-name>, <string-name><surname>Lamberti</surname> <given-names>F</given-names></string-name>, <string-name><surname>Morra</surname> <given-names>L</given-names></string-name></person-group>. <article-title>Data augmentation for medical imaging: a systematic literature review</article-title>. <source>Comput Biol Med</source>. <year>2023</year>;<volume>152</volume>(<issue>1</issue>):<fpage>106391</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.compbiomed.2022.106391</pub-id>; <pub-id pub-id-type="pmid">36549032</pub-id></mixed-citation></ref>
<ref id="ref-70"><label>[70]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kebaili</surname> <given-names>A</given-names></string-name>, <string-name><surname>Lapuyade-Lahorgue</surname> <given-names>J</given-names></string-name>, <string-name><surname>Ruan</surname> <given-names>S</given-names></string-name></person-group>. <article-title>Deep learning approaches for data augmentation in medical imaging: a review</article-title>. <source>J Imag</source>. <year>2023</year>;<volume>9</volume>(<issue>4</issue>):<fpage>81</fpage>. doi:<pub-id pub-id-type="doi">10.3390/jimaging9040081</pub-id>; <pub-id pub-id-type="pmid">37103232</pub-id></mixed-citation></ref>
<ref id="ref-71"><label>[71]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Sun</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Yuan</surname> <given-names>P</given-names></string-name>, <string-name><surname>Sun</surname> <given-names>Y</given-names></string-name></person-group>. <article-title>MM-GAN: 3D MRI data augmentation for medical image segmentation via generative adversarial networks</article-title>. In: <conf-name>2020 IEEE International Conference on Knowledge Graph (ICKG); 2020 Aug 9&#x2013;11</conf-name>; <publisher-loc>Nanjing, China</publisher-loc>. p. <fpage>227</fpage>&#x2013;<lpage>34</lpage>.</mixed-citation></ref>
<ref id="ref-72"><label>[72]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Goceri</surname> <given-names>E</given-names></string-name></person-group>. <article-title>Medical image data augmentation: techniques, comparisons and interpretations</article-title>. <source>Artif Intell Rev</source>. <year>2023</year>;<volume>56</volume>(<issue>11</issue>):<fpage>12561</fpage>&#x2013;<lpage>605</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s10462-023-10453-z</pub-id>; <pub-id pub-id-type="pmid">37362888</pub-id></mixed-citation></ref>
<ref id="ref-73"><label>[73]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Tschuchnig</surname> <given-names>ME</given-names></string-name>, <string-name><surname>Gadermayr</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Anomaly detection in medical imaging&#x2014;a mini review</article-title>. In: <conf-name>International Data Science Conference</conf-name>. <publisher-loc>Cham, Switzerland</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>2021</year>. p. <fpage>33</fpage>&#x2013;<lpage>8</lpage>.</mixed-citation></ref>
<ref id="ref-74"><label>[74]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Xia</surname> <given-names>X</given-names></string-name>, <string-name><surname>Pan</surname> <given-names>X</given-names></string-name>, <string-name><surname>Li</surname> <given-names>N</given-names></string-name>, <string-name><surname>He</surname> <given-names>X</given-names></string-name>, <string-name><surname>Ma</surname> <given-names>L</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>X</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>GAN-based anomaly detection: a review</article-title>. <source>Neurocomputing</source>. <year>2022</year>;<volume>493</volume>:<fpage>497</fpage>&#x2013;<lpage>535</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.neucom.2021.12.093</pub-id>.</mixed-citation></ref>
<ref id="ref-75"><label>[75]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Esmaeili</surname> <given-names>M</given-names></string-name>, <string-name><surname>Toosi</surname> <given-names>A</given-names></string-name>, <string-name><surname>Roshanpoor</surname> <given-names>A</given-names></string-name>, <string-name><surname>Changizi</surname> <given-names>V</given-names></string-name>, <string-name><surname>Ghazisaeedi</surname> <given-names>M</given-names></string-name>, <string-name><surname>Rahmim</surname> <given-names>A</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Generative adversarial networks for anomaly detection in biomedical imaging: a study on seven medical image datasets</article-title>. <source>IEEE Access</source>. <year>2023</year>;<volume>11</volume>:<fpage>17906</fpage>&#x2013;<lpage>21</lpage>. doi:<pub-id pub-id-type="doi">10.1109/access.2023.3244741</pub-id>.</mixed-citation></ref>
<ref id="ref-76"><label>[76]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Vyas</surname> <given-names>B</given-names></string-name>, <string-name><surname>Rajendran</surname> <given-names>RM</given-names></string-name></person-group>. <article-title>Generative adversarial networks for anomaly detection in medical images</article-title>. <source>Int J Multidisci Innova Res Method</source>. <year>2023</year>;<volume>2</volume>(<issue>4</issue>):<fpage>52</fpage>&#x2013;<lpage>8</lpage>.</mixed-citation></ref>
<ref id="ref-77"><label>[77]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Han</surname> <given-names>C</given-names></string-name>, <string-name><surname>Rundo</surname> <given-names>L</given-names></string-name>, <string-name><surname>Murao</surname> <given-names>K</given-names></string-name>, <string-name><surname>Noguchi</surname> <given-names>T</given-names></string-name>, <string-name><surname>Shimahara</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Milacski</surname> <given-names>Z&#x00C1;</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>MADGAN: unsupervised medical anomaly detection GAN using multiple adjacent brain MRI slice reconstruction</article-title>. <source>BMC Bioinformatics</source>. <year>2021</year>;<volume>22</volume>(<issue>S2</issue>):<fpage>31</fpage>. doi:<pub-id pub-id-type="doi">10.1186/s12859-020-03936-1</pub-id>; <pub-id pub-id-type="pmid">33902457</pub-id></mixed-citation></ref>
<ref id="ref-78"><label>[78]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zhang</surname> <given-names>H</given-names></string-name>, <string-name><surname>Guo</surname> <given-names>W</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>S</given-names></string-name>, <string-name><surname>Lu</surname> <given-names>H</given-names></string-name>, <string-name><surname>Zhao</surname> <given-names>X</given-names></string-name></person-group>. <article-title>Unsupervised deep anomaly detection for medical images using an improved adversarial autoencoder</article-title>. <source>J Digital Imag</source>. <year>2022</year>;<volume>35</volume>(<issue>2</issue>):<fpage>153</fpage>&#x2013;<lpage>61</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s10278-021-00558-8</pub-id>; <pub-id pub-id-type="pmid">35013826</pub-id></mixed-citation></ref>
<ref id="ref-79"><label>[79]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Guan</surname> <given-names>H</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Domain adaptation for medical image analysis: a survey</article-title>. <source>IEEE Transact Biomed Eng</source>. <year>2022</year>;<volume>69</volume>(<issue>3</issue>):<fpage>1173</fpage>&#x2013;<lpage>85</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tbme.2021.3117407</pub-id>; <pub-id pub-id-type="pmid">34606445</pub-id></mixed-citation></ref>
<ref id="ref-80"><label>[80]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Xie</surname> <given-names>X</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>J</given-names></string-name>, <string-name><surname>Li</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Shen</surname> <given-names>L</given-names></string-name>, <string-name><surname>Ma</surname> <given-names>K</given-names></string-name>, <string-name><surname>Zheng</surname> <given-names>Y</given-names></string-name></person-group>. <article-title>MI 2 GAN: generative adversarial network for medical image domain adaptation using mutual information constraint</article-title>. In: <conf-name>International Conference on Medical Image Computing and Computer-Assisted Intervention</conf-name>. <publisher-loc>Cham, Switzerland</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>2020</year>. p. <fpage>516</fpage>&#x2013;<lpage>25</lpage>.</mixed-citation></ref>
<ref id="ref-81"><label>[81]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Sun</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>G</given-names></string-name>, <string-name><surname>Ding</surname> <given-names>D</given-names></string-name>, <string-name><surname>Cheng</surname> <given-names>G</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>J</given-names></string-name>, <string-name><surname>Li</surname> <given-names>X</given-names></string-name></person-group>. <article-title>A GAN-based domain adaptation method for glaucoma diagnosis</article-title>. In: <conf-name>2020 International Joint Conference on Neural Networks (IJCNN); 2020 Jul 19&#x2013;24</conf-name>; <publisher-loc>Glasgow, UK</publisher-loc>. p. <fpage>1</fpage>&#x2013;<lpage>8</lpage>.</mixed-citation></ref>
<ref id="ref-82"><label>[82]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Sagheer</surname> <given-names>SVM</given-names></string-name>, <string-name><surname>George</surname> <given-names>SN</given-names></string-name></person-group>. <article-title>A review on medical image denoising algorithms</article-title>. <source>Biomed Signal Process Cont</source>. <year>2020</year>;<volume>61</volume>(<issue>11</issue>):<fpage>102036</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.bspc.2020.102036</pub-id>.</mixed-citation></ref>
<ref id="ref-83"><label>[83]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Isola</surname> <given-names>P</given-names></string-name>, <string-name><surname>Zhu</surname> <given-names>JY</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>T</given-names></string-name>, <string-name><surname>Efros</surname> <given-names>AA</given-names></string-name></person-group>. <article-title>Image-to-image translation with conditional adversarial networks</article-title>. In: <conf-name>Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition; 2017 Jul 21&#x2013;26</conf-name>; <publisher-loc>Honolulu, HI, USA</publisher-loc>. p. <fpage>1125</fpage>&#x2013;<lpage>34</lpage>.</mixed-citation></ref>
<ref id="ref-84"><label>[84]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kitchenham</surname> <given-names>B</given-names></string-name>, <string-name><surname>Brereton</surname> <given-names>OP</given-names></string-name>, <string-name><surname>Budgen</surname> <given-names>D</given-names></string-name>, <string-name><surname>Turner</surname> <given-names>M</given-names></string-name>, <string-name><surname>Bailey</surname> <given-names>J</given-names></string-name>, <string-name><surname>Linkman</surname> <given-names>S</given-names></string-name></person-group>. <article-title>Systematic literature reviews in software engineering-a systematic literature review</article-title>. <source>Informati Software Technol</source>. <year>2009</year>;<volume>51</volume>(<issue>1</issue>):<fpage>7</fpage>&#x2013;<lpage>15</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.infsof.2008.09.009</pub-id>.</mixed-citation></ref>
<ref id="ref-85"><label>[85]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ali</surname> <given-names>M</given-names></string-name>, <string-name><surname>Ali</surname> <given-names>M</given-names></string-name>, <string-name><surname>Hussain</surname> <given-names>M</given-names></string-name>, <string-name><surname>Koundal</surname> <given-names>D</given-names></string-name></person-group>. <article-title>Generative adversarial networks (GANs) for medical image processing: recent advancements</article-title>. <source>Arch Comput Methods Eng</source>. <year>2025</year>;<volume>32</volume>(<issue>2</issue>):<fpage>1185</fpage>&#x2013;<lpage>98</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s11831-024-10174-8</pub-id>.</mixed-citation></ref>
<ref id="ref-86"><label>[86]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Tu</surname> <given-names>H</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Zhao</surname> <given-names>Y</given-names></string-name></person-group>. <article-title>Unpaired image-to-image translation with diffusion adversarial network</article-title>. <source>Mathematics</source>. <year>2024</year>:<fpage>3178</fpage>. doi:<pub-id pub-id-type="doi">10.21203/rs.3.rs-4502713/v1</pub-id>.</mixed-citation></ref>
<ref id="ref-87"><label>[87]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>G</given-names></string-name>, <string-name><surname>Shi</surname> <given-names>H</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Wu</surname> <given-names>B</given-names></string-name></person-group>. <article-title>Unsupervised image-to-image translation via long-short cycle-consistent adversarial networks</article-title>. <source>Appl Intell</source>. <year>2023</year>;<volume>53</volume>(<issue>14</issue>):<fpage>17243</fpage>&#x2013;<lpage>59</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s10489-022-04389-0</pub-id>.</mixed-citation></ref>
<ref id="ref-88"><label>[88]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Sun</surname> <given-names>L</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>J</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Gong</surname> <given-names>M</given-names></string-name>, <string-name><surname>Yu</surname> <given-names>K</given-names></string-name>, <string-name><surname>Batmanghelich</surname> <given-names>K</given-names></string-name></person-group>. <article-title>Hierarchical amortized GAN for 3D high resolution medical image synthesis</article-title>. <source>IEEE J Biomed Health Inform</source>. <year>2022</year>;<volume>26</volume>(<issue>8</issue>):<fpage>3966</fpage>&#x2013;<lpage>75</lpage>. doi:<pub-id pub-id-type="doi">10.1109/jbhi.2022.3172976</pub-id>; <pub-id pub-id-type="pmid">35522642</pub-id></mixed-citation></ref>
<ref id="ref-89"><label>[89]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Xun</surname> <given-names>S</given-names></string-name>, <string-name><surname>Li</surname> <given-names>D</given-names></string-name>, <string-name><surname>Zhu</surname> <given-names>H</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>M</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>J</given-names></string-name>, <string-name><surname>Li</surname> <given-names>J</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Generative adversarial networks in medical image segmentation: a review</article-title>. <source>Comput Biol Med</source>. <year>2022</year>;<volume>140</volume>:<fpage>105063</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.compbiomed.2021.105063</pub-id>; <pub-id pub-id-type="pmid">34864584</pub-id></mixed-citation></ref>
<ref id="ref-90"><label>[90]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Armanious</surname> <given-names>K</given-names></string-name>, <string-name><surname>Jiang</surname> <given-names>C</given-names></string-name>, <string-name><surname>Fischer</surname> <given-names>M</given-names></string-name>, <string-name><surname>K&#x00FC;stner</surname> <given-names>T</given-names></string-name>, <string-name><surname>Hepp</surname> <given-names>T</given-names></string-name>, <string-name><surname>Nikolaou</surname> <given-names>K</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>MedGAN: medical image translation using GANs</article-title>. <source>Comput Med Imag Grap</source>. <year>2020</year>;<volume>79</volume>:<fpage>101684</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.compmedimag.2019.101684</pub-id>; <pub-id pub-id-type="pmid">31812132</pub-id></mixed-citation></ref>
<ref id="ref-91"><label>[91]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Zhang</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Sun</surname> <given-names>J</given-names></string-name>, <string-name><surname>Mok</surname> <given-names>GS</given-names></string-name></person-group>. <article-title>Low dose SPECT image denoising using a generative adversarial network</article-title>. <comment>arXiv:1907.11944. 2019</comment>.</mixed-citation></ref>
<ref id="ref-92"><label>[92]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Mahapatra</surname> <given-names>D</given-names></string-name>, <string-name><surname>Bozorgtabar</surname> <given-names>B</given-names></string-name>, <string-name><surname>Garnavi</surname> <given-names>R</given-names></string-name></person-group>. <article-title>Image super-resolution using progressive generative adversarial networks for medical image analysis</article-title>. <source>Comput Med Imag Grap</source>. <year>2019</year>;<volume>71</volume>(<issue>6</issue>):<fpage>30</fpage>&#x2013;<lpage>9</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.compmedimag.2018.10.005</pub-id>; <pub-id pub-id-type="pmid">30472408</pub-id></mixed-citation></ref>
<ref id="ref-93"><label>[93]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>You</surname> <given-names>C</given-names></string-name>, <string-name><surname>Li</surname> <given-names>G</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>X</given-names></string-name>, <string-name><surname>Shan</surname> <given-names>H</given-names></string-name>, <string-name><surname>Li</surname> <given-names>M</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>CT super-resolution GAN constrained by the identical, residual, and cycle learning ensemble (GAN-CIRCLE)</article-title>. <source>IEEE Transact Med Imag</source>. <year>2020</year>;<volume>39</volume>(<issue>1</issue>):<fpage>188</fpage>&#x2013;<lpage>203</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2019.2922960</pub-id>; <pub-id pub-id-type="pmid">31217097</pub-id></mixed-citation></ref>
<ref id="ref-94"><label>[94]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Yang</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Yan</surname> <given-names>P</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Yu</surname> <given-names>H</given-names></string-name>, <string-name><surname>Shi</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Mou</surname> <given-names>X</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Low-dose CT image denoising using a generative adversarial network with Wasserstein distance and perceptual loss</article-title>. <source>IEEE Transact Med Imag</source>. <year>2018</year>;<volume>37</volume>(<issue>6</issue>):<fpage>1348</fpage>&#x2013;<lpage>57</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2018.2827462</pub-id>; <pub-id pub-id-type="pmid">29870364</pub-id></mixed-citation></ref>
<ref id="ref-95"><label>[95]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Li</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Shen</surname> <given-names>L</given-names></string-name></person-group>. <article-title>cC-GAN: a robust transfer-learning framework for HEp-2 specimen image segmentation</article-title>. <source>IEEE Access</source>. <year>2018</year>;<volume>6</volume>:<fpage>14048</fpage>&#x2013;<lpage>58</lpage>. doi:<pub-id pub-id-type="doi">10.1109/access.2018.2808938</pub-id>.</mixed-citation></ref>
<ref id="ref-96"><label>[96]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Hu</surname> <given-names>B</given-names></string-name>, <string-name><surname>Tang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Eric</surname> <given-names>I</given-names></string-name>, <string-name><surname>Chang</surname> <given-names>C</given-names></string-name>, <string-name><surname>Fan</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Lai</surname> <given-names>M</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Unsupervised learning for cell-level visual representation in histopathology images with generative adversarial networks</article-title>. <source>IEEE J Biomed Health Inform</source>. <year>2018</year>;<volume>23</volume>(<issue>3</issue>):<fpage>1316</fpage>&#x2013;<lpage>28</lpage>. doi:<pub-id pub-id-type="doi">10.1109/jbhi.2018.2852639</pub-id>; <pub-id pub-id-type="pmid">29994411</pub-id></mixed-citation></ref>
<ref id="ref-97"><label>[97]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Nie</surname> <given-names>D</given-names></string-name>, <string-name><surname>Trullo</surname> <given-names>R</given-names></string-name>, <string-name><surname>Lian</surname> <given-names>J</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>L</given-names></string-name>, <string-name><surname>Petitjean</surname> <given-names>C</given-names></string-name>, <string-name><surname>Ruan</surname> <given-names>S</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Medical image synthesis with deep convolutional adversarial networks</article-title>. <source>IEEE Transact Biomed Eng</source>. <year>2018</year>;<volume>65</volume>(<issue>12</issue>):<fpage>2720</fpage>&#x2013;<lpage>30</lpage>.</mixed-citation></ref>
<ref id="ref-98"><label>[98]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Bissoto</surname> <given-names>A</given-names></string-name>, <string-name><surname>Perez</surname> <given-names>F</given-names></string-name>, <string-name><surname>Valle</surname> <given-names>E</given-names></string-name>, <string-name><surname>Avila</surname> <given-names>S</given-names></string-name></person-group>. <chapter-title>Skin lesion synthesis with generative adversarial networks</chapter-title>. In: <source>OR 2.0 context-aware operating theaters, computer assisted robotic endoscopy, clinical image-based procedures, and skin image analysis</source>. <publisher-loc>Cham, Switzerland</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>2018</year>. p. <fpage>294</fpage>&#x2013;<lpage>302</lpage>. doi:<pub-id pub-id-type="doi">10.1007/978-3-030-01201-4_32</pub-id>.</mixed-citation></ref>
<ref id="ref-99"><label>[99]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Udrea</surname> <given-names>A</given-names></string-name>, <string-name><surname>Mitra</surname> <given-names>GD</given-names></string-name></person-group>. <article-title>Generative adversarial neural networks for pigmented and non-pigmented skin lesions detection in clinical images</article-title>. In: <conf-name>2017 21st International Conference on Control Systems and Computer Science (CSCS); 2017 May 29&#x2013;31</conf-name>; <publisher-loc>Bucharest, Romania</publisher-loc>. p. <fpage>364</fpage>&#x2013;<lpage>8</lpage>.</mixed-citation></ref>
<ref id="ref-100"><label>[100]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>N</given-names></string-name>, <string-name><surname>Li</surname> <given-names>J</given-names></string-name></person-group>. <article-title>GAN-based architecture for low-dose computed tomography imaging denoising</article-title>. <comment>arXiv:2411.09512. 2024</comment>.</mixed-citation></ref>
<ref id="ref-101"><label>[101]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Yu</surname> <given-names>X</given-names></string-name>, <string-name><surname>Luan</surname> <given-names>S</given-names></string-name>, <string-name><surname>Lei</surname> <given-names>S</given-names></string-name>, <string-name><surname>Huang</surname> <given-names>J</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Xue</surname> <given-names>X</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Deep learning for fast denoising filtering in ultrasound localization microscopy</article-title>. <source>Phy Med Biol</source>. <year>2023</year>;<volume>68</volume>(<issue>20</issue>):<fpage>205002</fpage>. doi:<pub-id pub-id-type="doi">10.1088/1361-6560/acf98f</pub-id>; <pub-id pub-id-type="pmid">37703894</pub-id></mixed-citation></ref>
<ref id="ref-102"><label>[102]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ran</surname> <given-names>M</given-names></string-name>, <string-name><surname>Hu</surname> <given-names>J</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>H</given-names></string-name>, <string-name><surname>Sun</surname> <given-names>H</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>J</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Denoising of 3D magnetic resonance images using a residual encoder-decoder Wasserstein generative adversarial network</article-title>. <source>Med Image Anal</source>. <year>2019</year>;<volume>55</volume>(<issue>2</issue>):<fpage>165</fpage>&#x2013;<lpage>80</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.media.2019.05.001</pub-id>; <pub-id pub-id-type="pmid">31085444</pub-id></mixed-citation></ref>
<ref id="ref-103"><label>[103]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kaur</surname> <given-names>A</given-names></string-name>, <string-name><surname>Dong</surname> <given-names>G</given-names></string-name></person-group>. <article-title>A complete review on image denoising techniques for medical images</article-title>. <source>Neural Process Lett</source>. <year>2023</year>;<volume>55</volume>(<issue>6</issue>):<fpage>7807</fpage>&#x2013;<lpage>50</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s11063-023-11286-1</pub-id>.</mixed-citation></ref>
<ref id="ref-104"><label>[104]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Yi</surname> <given-names>X</given-names></string-name>, <string-name><surname>Babyn</surname> <given-names>P</given-names></string-name></person-group>. <article-title>Sharpness-aware low-dose CT denoising using conditional generative adversarial network</article-title>. <source>J Digital Imag</source>. <year>2018</year>;<volume>31</volume>(<issue>5</issue>):<fpage>655</fpage>&#x2013;<lpage>69</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s10278-018-0056-0</pub-id>; <pub-id pub-id-type="pmid">29464432</pub-id></mixed-citation></ref>
<ref id="ref-105"><label>[105]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Shan</surname> <given-names>H</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Kruger</surname> <given-names>U</given-names></string-name>, <string-name><surname>Kalra</surname> <given-names>MK</given-names></string-name>, <string-name><surname>Sun</surname> <given-names>L</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>3-D convolutional encoder-decoder network for low-dose CT via transfer learning from a 2-D trained network</article-title>. <source>IEEE Transact Med Imag</source>. <year>2018</year>;<volume>37</volume>(<issue>6</issue>):<fpage>1522</fpage>&#x2013;<lpage>34</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2018.2878429</pub-id>.</mixed-citation></ref>
<ref id="ref-106"><label>[106]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wolterink</surname> <given-names>JM</given-names></string-name>, <string-name><surname>Leiner</surname> <given-names>T</given-names></string-name>, <string-name><surname>Viergever</surname> <given-names>MA</given-names></string-name>, <string-name><surname>I&#x0161;gum</surname> <given-names>I</given-names></string-name></person-group>. <article-title>Generative adversarial networks for noise reduction in low-dose CT</article-title>. <source>IEEE Transact Med Imag</source>. <year>2017</year>;<volume>36</volume>(<issue>12</issue>):<fpage>2536</fpage>&#x2013;<lpage>45</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2017.2708987</pub-id>; <pub-id pub-id-type="pmid">28574346</pub-id></mixed-citation></ref>
<ref id="ref-107"><label>[107]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Huang</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>J</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Shan</surname> <given-names>H</given-names></string-name></person-group>. <article-title>DU-GAN: generative adversarial networks with dual-domain U-Net-based discriminators for low-dose CT denoising</article-title>. <source>IEEE Transact Instrument Measure</source>. <year>2022</year>;<volume>71</volume>:<fpage>1</fpage>&#x2013;<lpage>12</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tim.2021.3128703</pub-id>.</mixed-citation></ref>
<ref id="ref-108"><label>[108]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zhu</surname> <given-names>ML</given-names></string-name>, <string-name><surname>Zhao</surname> <given-names>LL</given-names></string-name>, <string-name><surname>Xiao</surname> <given-names>L</given-names></string-name></person-group>. <article-title>Image denoising based on GAN with optimization algorithm</article-title>. <source>Electronics</source>. <year>2022</year>;<volume>11</volume>(<issue>15</issue>):<fpage>2445</fpage>. doi:<pub-id pub-id-type="doi">10.3390/electronics11152445</pub-id>.</mixed-citation></ref>
<ref id="ref-109"><label>[109]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>R</given-names></string-name>, <string-name><surname>Lei</surname> <given-names>T</given-names></string-name>, <string-name><surname>Cui</surname> <given-names>R</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>B</given-names></string-name>, <string-name><surname>Meng</surname> <given-names>H</given-names></string-name>, <string-name><surname>Nandi</surname> <given-names>AK</given-names></string-name></person-group>. <article-title>Medical image segmentation using deep learning: a survey</article-title>. <source>IET Image Process</source>. <year>2022</year>;<volume>16</volume>(<issue>5</issue>):<fpage>1243</fpage>&#x2013;<lpage>67</lpage>. doi:<pub-id pub-id-type="doi">10.1049/ipr2.12419</pub-id>.</mixed-citation></ref>
<ref id="ref-110"><label>[110]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Cai</surname> <given-names>L</given-names></string-name>, <string-name><surname>Fang</surname> <given-names>H</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>N</given-names></string-name>, <string-name><surname>Ren</surname> <given-names>B</given-names></string-name></person-group>. <article-title>Counterfactual causal-effect intervention for interpretable medical visual question answering</article-title>. <source>IEEE Trans Med Imaging</source>. <year>2024</year>;<volume>43</volume>(<issue>12</issue>):<fpage>4430</fpage>&#x2013;<lpage>41</lpage>. doi:<pub-id pub-id-type="doi">10.1109/TMI.2024.3425533</pub-id>; <pub-id pub-id-type="pmid">38980786</pub-id></mixed-citation></ref>
<ref id="ref-111"><label>[111]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Song</surname> <given-names>W</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>X</given-names></string-name>, <string-name><surname>Guo</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Li</surname> <given-names>S</given-names></string-name>, <string-name><surname>Xia</surname> <given-names>B</given-names></string-name>, <string-name><surname>Hao</surname> <given-names>A</given-names></string-name></person-group>. <article-title>Centerformer: a novel cluster center enhanced transformer for unconstrained dental plaque segmentation</article-title>. <source>IEEE Trans Multimedia</source>. <year>2024</year>;<volume>26</volume>:<fpage>10965</fpage>&#x2013;<lpage>78</lpage>. doi:<pub-id pub-id-type="doi">10.1109/TMM.2024.3428349</pub-id>.</mixed-citation></ref>
<ref id="ref-112"><label>[112]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Dong</surname> <given-names>X</given-names></string-name>, <string-name><surname>Lei</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>T</given-names></string-name>, <string-name><surname>Thomas</surname> <given-names>M</given-names></string-name>, <string-name><surname>Tang</surname> <given-names>L</given-names></string-name>, <string-name><surname>Curran</surname> <given-names>WJ</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Automatic multiorgan segmentation in thorax CT images using U-net-GAN</article-title>. <source>Med Phy</source>. <year>2019</year>;<volume>46</volume>(<issue>5</issue>):<fpage>2157</fpage>&#x2013;<lpage>68</lpage>. doi:<pub-id pub-id-type="doi">10.1002/mp.13458</pub-id>; <pub-id pub-id-type="pmid">30810231</pub-id></mixed-citation></ref>
<ref id="ref-113"><label>[113]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Guo</surname> <given-names>B</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Yu</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Zhu</surname> <given-names>Y</given-names></string-name></person-group>. <article-title>Multi-modality medical image segmentation via adversarial learning with CV energy functional</article-title>. <source>Expert Syst Appl</source>. <year>2025</year>;<volume>283</volume>(<issue>8</issue>):<fpage>127467</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.eswa.2025.127467</pub-id>.</mixed-citation></ref>
<ref id="ref-114"><label>[114]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>G</given-names></string-name>, <string-name><surname>Ma</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Li</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Mao</surname> <given-names>K</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>L</given-names></string-name>, <string-name><surname>Zhao</surname> <given-names>Y</given-names></string-name></person-group>. <article-title>A skin lesion segmentation network with edge and body fusion</article-title>. <source>Appl Soft Comput</source>. <year>2025</year>;<volume>170</volume>(<issue>1</issue>):<fpage>112683</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.asoc.2024.112683</pub-id>.</mixed-citation></ref>
<ref id="ref-115"><label>[115]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Gunasagar</surname> <given-names>P</given-names></string-name>, <string-name><surname>Durgaprasad</surname> <given-names>V</given-names></string-name>, <string-name><surname>Abhijith</surname> <given-names>K</given-names></string-name>, <string-name><surname>Sai</surname> <given-names>RM</given-names></string-name>, <string-name><surname>Ameer</surname> <given-names>P</given-names></string-name></person-group>. <article-title>Enhancing facial expression synthesis through GAN with multi-scale dilated feature extraction and edge-enhanced facial features</article-title>. In: <conf-name>TENCON 2023-2023 IEEE Region 10 Conference (TENCON); 2023 Oct 31&#x2013;Nov 3</conf-name>; <publisher-loc>Chiang Mai, Thailand</publisher-loc>. p. <fpage>47</fpage>&#x2013;<lpage>52</lpage>.</mixed-citation></ref>
<ref id="ref-116"><label>[116]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Nimitha</surname> <given-names>U</given-names></string-name>, <string-name><surname>Ameer</surname> <given-names>P</given-names></string-name></person-group>. <article-title>MRI super-resolution using similarity distance and multi-scale receptive field based feature fusion GAN and pre-trained slice interpolation network</article-title>. <source>Magn Reson Imag</source>. <year>2024</year>;<volume>110</volume>(<issue>5</issue>):<fpage>195</fpage>&#x2013;<lpage>209</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.mri.2024.04.021</pub-id>; <pub-id pub-id-type="pmid">38653336</pub-id></mixed-citation></ref>
<ref id="ref-117"><label>[117]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>B</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>W</given-names></string-name>, <string-name><surname>Qian</surname> <given-names>J</given-names></string-name>, <string-name><surname>Feng</surname> <given-names>S</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Zuo</surname> <given-names>C</given-names></string-name></person-group>. <article-title>Single-shot super-resolved fringe projection profilometry (SSSR-FPP): 100,000 frames-per-second 3D imaging with deep learning</article-title>. <source>Light Sci Appl</source>. <year>2025</year>;<volume>14</volume>(<issue>1</issue>):<fpage>70</fpage>. doi:<pub-id pub-id-type="doi">10.1038/s41377-024-01721-w</pub-id>; <pub-id pub-id-type="pmid">39915449</pub-id></mixed-citation></ref>
<ref id="ref-118"><label>[118]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Ledig</surname> <given-names>C</given-names></string-name>, <string-name><surname>Theis</surname> <given-names>L</given-names></string-name>, <string-name><surname>Husz&#x00E1;r</surname> <given-names>F</given-names></string-name>, <string-name><surname>Caballero</surname> <given-names>J</given-names></string-name>, <string-name><surname>Cunningham</surname> <given-names>A</given-names></string-name>, <string-name><surname>Acosta</surname> <given-names>A</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Photo-realistic single image super-resolution using a generative adversarial network</article-title>. In: <conf-name>Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition; 2017 Jul 21&#x2013;26</conf-name>; <publisher-loc>Honolulu, HI, USA</publisher-loc>. p. <fpage>4681</fpage>&#x2013;<lpage>90</lpage>.</mixed-citation></ref>
<ref id="ref-119"><label>[119]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zhao</surname> <given-names>T</given-names></string-name>, <string-name><surname>Lin</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>W</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>Z</given-names></string-name></person-group>. <article-title>Learning-based quality assessment for image super-resolution</article-title>. <source>IEEE Transact Multim</source>. <year>2022</year>;<volume>24</volume>:<fpage>3570</fpage>&#x2013;<lpage>81</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmm.2021.3102401</pub-id>.</mixed-citation></ref>
<ref id="ref-120"><label>[120]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Chen</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Shi</surname> <given-names>F</given-names></string-name>, <string-name><surname>Christodoulou</surname> <given-names>AG</given-names></string-name>, <string-name><surname>Xie</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Li</surname> <given-names>D</given-names></string-name></person-group>. <article-title>Efficient and accurate MRI super-resolution using a generative adversarial network and 3D multi-level densely connected network</article-title>. In: <conf-name>International Conference on Medical Image Computing and Computer-Assisted Intervention</conf-name>. <publisher-loc>Cham, Switzerland</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>2018</year>. p. <fpage>91</fpage>&#x2013;<lpage>9</lpage>.</mixed-citation></ref>
<ref id="ref-121"><label>[121]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Gurcan</surname> <given-names>MN</given-names></string-name>, <string-name><surname>Boucheron</surname> <given-names>LE</given-names></string-name>, <string-name><surname>Can</surname> <given-names>A</given-names></string-name>, <string-name><surname>Madabhushi</surname> <given-names>A</given-names></string-name>, <string-name><surname>Rajpoot</surname> <given-names>NM</given-names></string-name>, <string-name><surname>Yener</surname> <given-names>B</given-names></string-name></person-group>. <article-title>Histopathological image analysis: a review</article-title>. <source>IEEE Rev Biomed Eng</source>. <year>2009</year>;<volume>2</volume>:<fpage>147</fpage>&#x2013;<lpage>71</lpage>. doi:<pub-id pub-id-type="doi">10.1109/rbme.2009.2034865</pub-id>; <pub-id pub-id-type="pmid">20671804</pub-id></mixed-citation></ref>
<ref id="ref-122"><label>[122]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Vu</surname> <given-names>CT</given-names></string-name>, <string-name><surname>Phan</surname> <given-names>TD</given-names></string-name>, <string-name><surname>Chandler</surname> <given-names>DM</given-names></string-name></person-group>. <article-title>A spectral and spatial measure of local perceived sharpness in natural images</article-title>. <source>IEEE Transact Image Process</source>. <year>2012</year>;<volume>21</volume>(<issue>3</issue>):<fpage>934</fpage>&#x2013;<lpage>45</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tip.2011.2169974</pub-id>; <pub-id pub-id-type="pmid">21965207</pub-id></mixed-citation></ref>
<ref id="ref-123"><label>[123]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Brenner</surname> <given-names>DJ</given-names></string-name>, <string-name><surname>Hall</surname> <given-names>EJ</given-names></string-name></person-group>. <article-title>Computed tomography&#x2014;an increasing source of radiation exposure</article-title>. <source>New Engl J Med</source>. <year>2007</year>;<volume>357</volume>(<issue>22</issue>):<fpage>2277</fpage>&#x2013;<lpage>84</lpage>. doi:<pub-id pub-id-type="doi">10.1056/nejmra072149</pub-id>; <pub-id pub-id-type="pmid">18046031</pub-id></mixed-citation></ref>
<ref id="ref-124"><label>[124]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Gu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zeng</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>H</given-names></string-name>, <string-name><surname>Wei</surname> <given-names>J</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>B</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>MedSRGAN: medical images super-resolution using generative adversarial networks</article-title>. <source>Multim Tools Applicat</source>. <year>2020</year>;<volume>79</volume>(<issue>29&#x2013;30</issue>):<fpage>21815</fpage>&#x2013;<lpage>40</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s11042-020-08980-w</pub-id>.</mixed-citation></ref>
<ref id="ref-125"><label>[125]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>R</given-names></string-name>, <string-name><surname>Fang</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Gu</surname> <given-names>J</given-names></string-name>, <string-name><surname>Guo</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>S</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>Y</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>High-resolution image reconstruction for portable ultrasound imaging devices</article-title>. <source>EURASIP J Adv Signal Process</source>. <year>2019</year>;<volume>2019</volume>(<issue>1</issue>):<fpage>1</fpage>&#x2013;<lpage>12</lpage>. doi:<pub-id pub-id-type="doi">10.1186/s13634-019-0649-x</pub-id>.</mixed-citation></ref>
<ref id="ref-126"><label>[126]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Luan</surname> <given-names>S</given-names></string-name>, <string-name><surname>Yu</surname> <given-names>X</given-names></string-name>, <string-name><surname>Lei</surname> <given-names>S</given-names></string-name>, <string-name><surname>Ma</surname> <given-names>C</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>X</given-names></string-name>, <string-name><surname>Xue</surname> <given-names>X</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Deep learning for fast super-resolution ultrasound microvessel imaging</article-title>. <source>Phy Med Biol</source>. <year>2023</year>;<volume>68</volume>(<issue>24</issue>):<fpage>245023</fpage>. doi:<pub-id pub-id-type="doi">10.1088/1361-6560/ad0a5a</pub-id>; <pub-id pub-id-type="pmid">37934040</pub-id></mixed-citation></ref>
<ref id="ref-127"><label>[127]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Nimitha</surname> <given-names>U</given-names></string-name>, <string-name><surname>Ameer</surname> <given-names>P</given-names></string-name></person-group>. <article-title>Multi image super resolution of MRI images using generative adversarial network</article-title>. <source>J Ambient Intell Human Comput</source>. <year>2024</year>;<volume>15</volume>(<issue>4</issue>):<fpage>2241</fpage>&#x2013;<lpage>53</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s12652-024-04751-9</pub-id>.</mixed-citation></ref>
<ref id="ref-128"><label>[128]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Jia</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>G</given-names></string-name>, <string-name><surname>Chi</surname> <given-names>H</given-names></string-name></person-group>. <article-title>Retinal fundus image super-resolution based on generative adversarial network guided with vascular structure prior</article-title>. <source>Scient Rep</source>. <year>2024</year>;<volume>14</volume>(<issue>1</issue>):<fpage>22786</fpage>. doi:<pub-id pub-id-type="doi">10.1038/s41598-024-74186-x</pub-id>; <pub-id pub-id-type="pmid">39354105</pub-id></mixed-citation></ref>
<ref id="ref-129"><label>[129]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>X</given-names></string-name>, <string-name><surname>Xie</surname> <given-names>L</given-names></string-name>, <string-name><surname>Dong</surname> <given-names>C</given-names></string-name>, <string-name><surname>Shan</surname> <given-names>Y</given-names></string-name></person-group>. <article-title>Real-esrgan: training real-world blind super-resolution with pure synthetic data</article-title>. In: <conf-name>Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision; 2021 Oct 11&#x2013;17</conf-name>; <publisher-loc>Montreal, BC, Canada</publisher-loc>. p. <fpage>1905</fpage>&#x2013;<lpage>14</lpage>.</mixed-citation></ref>
<ref id="ref-130"><label>[130]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zhao</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zheng</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhao</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Luo</surname> <given-names>L</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>S</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Automatic 2-D/3-D vessel enhancement in multiple modality images using a weighted symmetry filter</article-title>. <source>IEEE Transact Med Imag</source>. <year>2017</year>;<volume>37</volume>(<issue>2</issue>):<fpage>438</fpage>&#x2013;<lpage>50</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2017.2756073</pub-id>; <pub-id pub-id-type="pmid">28952938</pub-id></mixed-citation></ref>
<ref id="ref-131"><label>[131]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Jobson</surname> <given-names>DJ</given-names></string-name>, <string-name><surname>Rahman</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Woodell</surname> <given-names>GA</given-names></string-name></person-group>. <article-title>A multiscale retinex for bridging the gap between color images and the human observation of scenes</article-title>. <source>IEEE Transact Image Process</source>. <year>1997</year>;<volume>6</volume>(<issue>7</issue>):<fpage>965</fpage>&#x2013;<lpage>76</lpage>. doi:<pub-id pub-id-type="doi">10.1109/83.597272</pub-id>; <pub-id pub-id-type="pmid">18282987</pub-id></mixed-citation></ref>
<ref id="ref-132"><label>[132]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Chen</surname> <given-names>YS</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>YC</given-names></string-name>, <string-name><surname>Kao</surname> <given-names>MH</given-names></string-name>, <string-name><surname>Chuang</surname> <given-names>YY</given-names></string-name></person-group>. <article-title>Deep photo enhancer: unpaired learning for image enhancement from photographs with gans</article-title>. In: <conf-name>Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition; 2018 Jun 18&#x2013;23</conf-name>; <publisher-loc>Salt Lake City, UT, USA</publisher-loc>. p. <fpage>6306</fpage>&#x2013;<lpage>14</lpage>.</mixed-citation></ref>
<ref id="ref-133"><label>[133]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Alotaibi</surname> <given-names>A</given-names></string-name></person-group>. <article-title>Deep generative adversarial networks for image-to-image translation: a review</article-title>. <source>Symmetry</source>. <year>2020</year>;<volume>12</volume>(<issue>10</issue>):<fpage>1705</fpage>. doi:<pub-id pub-id-type="doi">10.3390/sym12101705</pub-id>.</mixed-citation></ref>
<ref id="ref-134"><label>[134]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Lore</surname> <given-names>KG</given-names></string-name>, <string-name><surname>Akintayo</surname> <given-names>A</given-names></string-name>, <string-name><surname>Sarkar</surname> <given-names>S</given-names></string-name></person-group>. <article-title>LLNet: a deep autoencoder approach to natural low-light image enhancement</article-title>. <source>Pattern Recognit</source>. <year>2017</year>;<volume>61</volume>(<issue>6</issue>):<fpage>650</fpage>&#x2013;<lpage>62</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.patcog.2016.06.008</pub-id>.</mixed-citation></ref>
<ref id="ref-135"><label>[135]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Gatys</surname> <given-names>LA</given-names></string-name></person-group>. <article-title>A neural algorithm of artistic style</article-title>. <comment>arXiv:1508.06576. 2015</comment>.</mixed-citation></ref>
<ref id="ref-136"><label>[136]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Zhu</surname> <given-names>JY</given-names></string-name>, <string-name><surname>Park</surname> <given-names>T</given-names></string-name>, <string-name><surname>Isola</surname> <given-names>P</given-names></string-name>, <string-name><surname>Efros</surname> <given-names>AA</given-names></string-name></person-group>. <article-title>Unpaired image-to-image translation using cycle-consistent adversarial networks</article-title>. In: <conf-name>Proceedings of the 2017 IEEE International Conference on Computer Vision; 2017 Oct 22&#x2013;29</conf-name>; <publisher-loc>Venice, Italy</publisher-loc>. p. <fpage>2223</fpage>&#x2013;<lpage>32</lpage>.</mixed-citation></ref>
<ref id="ref-137"><label>[137]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Yan</surname> <given-names>S</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>C</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>W</given-names></string-name>, <string-name><surname>Lyu</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Swin transformer-based GAN for multi-modal medical image translation</article-title>. <source>Front Oncol</source>. <year>2022</year>;<volume>12</volume>:<fpage>942511</fpage>. doi:<pub-id pub-id-type="doi">10.3389/fonc.2022.942511</pub-id>; <pub-id pub-id-type="pmid">36003791</pub-id></mixed-citation></ref>
<ref id="ref-138"><label>[138]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Liang</surname> <given-names>J</given-names></string-name>, <string-name><surname>Cao</surname> <given-names>J</given-names></string-name>, <string-name><surname>Sun</surname> <given-names>G</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>K</given-names></string-name>, <string-name><surname>Van Gool</surname> <given-names>L</given-names></string-name>, <string-name><surname>Timofte</surname> <given-names>R</given-names></string-name></person-group>. <article-title>Swinir: image restoration using swin transformer</article-title>. In: <conf-name>Proceedings of the 2021 IEEE/CVF iNternational Conference on Computer Vision; 2021 Oct 11&#x2013;17</conf-name>; <publisher-loc>Montreal, BC, Canada</publisher-loc>. p. <fpage>1833</fpage>&#x2013;<lpage>44</lpage>.</mixed-citation></ref>
<ref id="ref-139"><label>[139]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Chen</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Lin</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>X</given-names></string-name>, <string-name><surname>Ding</surname> <given-names>J</given-names></string-name>, <string-name><surname>Li</surname> <given-names>C</given-names></string-name>, <string-name><surname>Zeng</surname> <given-names>Y</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Multi-domain medical image translation generation for lung image classification based on generative adversarial networks</article-title>. <source>Comput Meth Prog Biomed</source>. <year>2023</year>;<volume>229</volume>(<issue>20</issue>):<fpage>107200</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.cmpb.2022.107200</pub-id>; <pub-id pub-id-type="pmid">36525713</pub-id></mixed-citation></ref>
<ref id="ref-140"><label>[140]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>&#x00D6;zbey</surname> <given-names>M</given-names></string-name>, <string-name><surname>Dalmaz</surname> <given-names>O</given-names></string-name>, <string-name><surname>Dar</surname> <given-names>SU</given-names></string-name>, <string-name><surname>Bedel</surname> <given-names>HA</given-names></string-name>, <string-name><surname>&#x00D6;zturk</surname> <given-names>&#x015E;</given-names></string-name>, <string-name><surname>G&#x00FC;ng&#x00F6;r</surname> <given-names>A</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Unsupervised medical image translation with adversarial diffusion models</article-title>. <source>IEEE Transact Med Imag</source>. <year>2023</year>;<volume>42</volume>(<issue>12</issue>):<fpage>3524</fpage>&#x2013;<lpage>39</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2023.3290149</pub-id>; <pub-id pub-id-type="pmid">37379177</pub-id></mixed-citation></ref>
<ref id="ref-141"><label>[141]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kim</surname> <given-names>KH</given-names></string-name>, <string-name><surname>Do</surname> <given-names>WJ</given-names></string-name>, <string-name><surname>Park</surname> <given-names>SH</given-names></string-name></person-group>. <article-title>Improving resolution of MR images with an adversarial network incorporating images with different contrast</article-title>. <source>Med Phy</source>. <year>2018</year>;<volume>45</volume>(<issue>7</issue>):<fpage>3120</fpage>&#x2013;<lpage>31</lpage>. doi:<pub-id pub-id-type="doi">10.1002/mp.12945</pub-id>; <pub-id pub-id-type="pmid">29729006</pub-id></mixed-citation></ref>
<ref id="ref-142"><label>[142]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Shitrit</surname> <given-names>O</given-names></string-name>, <string-name><surname>Riklin Raviv</surname> <given-names>T</given-names></string-name></person-group>. <chapter-title>Accelerated magnetic resonance imaging by adversarial neural network</chapter-title>. In: <source>Deep learning in medical image analysis and multimodal learning for clinical decision support</source>. <publisher-loc>Cham, Switzerland</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>2017</year>. p. <fpage>30</fpage>&#x2013;<lpage>8</lpage>. doi:<pub-id pub-id-type="doi">10.1007/978-3-319-67558-9_4</pub-id>.</mixed-citation></ref>
<ref id="ref-143"><label>[143]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Seitzer</surname> <given-names>M</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>G</given-names></string-name>, <string-name><surname>Schlemper</surname> <given-names>J</given-names></string-name>, <string-name><surname>Oktay</surname> <given-names>O</given-names></string-name>, <string-name><surname>W&#x00FC;rfl</surname> <given-names>T</given-names></string-name>, <string-name><surname>Christlein</surname> <given-names>V</given-names></string-name>, <etal>et al</etal></person-group>. <chapter-title>Adversarial and perceptual refinement for compressed sensing MRI reconstruction</chapter-title>. In: <source>Medical image computing and computer assisted intervention-MICCAI 2018</source>. <publisher-loc>Cham, Switzerland</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>2018</year>. p. <fpage>232</fpage>&#x2013;<lpage>40</lpage>. doi:<pub-id pub-id-type="doi">10.1007/978-3-030-00928-1_27</pub-id>.</mixed-citation></ref>
<ref id="ref-144"><label>[144]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Armanious</surname> <given-names>K</given-names></string-name>, <string-name><surname>Mecky</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Gatidis</surname> <given-names>S</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>B</given-names></string-name></person-group>. <article-title>Adversarial inpainting of medical image modalities</article-title>. In: <conf-name>ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); 2019 May 12&#x2013;17</conf-name>; <publisher-loc>Brighton, UK</publisher-loc>. p. <fpage>3267</fpage>&#x2013;<lpage>71</lpage>.</mixed-citation></ref>
<ref id="ref-145"><label>[145]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Yu</surname> <given-names>B</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>L</given-names></string-name>, <string-name><surname>Zu</surname> <given-names>C</given-names></string-name>, <string-name><surname>Lalush</surname> <given-names>DS</given-names></string-name>, <string-name><surname>Lin</surname> <given-names>W</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>3D conditional generative adversarial networks for high-quality PET image estimation at low dose</article-title>. <source>Neuroimage</source>. <year>2018</year>;<volume>174</volume>(<issue>7</issue>):<fpage>550</fpage>&#x2013;<lpage>62</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.neuroimage.2018.03.045</pub-id>; <pub-id pub-id-type="pmid">29571715</pub-id></mixed-citation></ref>
<ref id="ref-146"><label>[146]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Liao</surname> <given-names>H</given-names></string-name>, <string-name><surname>Huo</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Sehnert</surname> <given-names>WJ</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>SK</given-names></string-name>, <string-name><surname>Luo</surname> <given-names>J</given-names></string-name></person-group>. <chapter-title>Adversarial sparse-view CBCT artifact reduction</chapter-title>. In: <source>Medical image computing and computer assisted intervention-MICCAI 2018I</source>. <publisher-loc>Cham, Switzerland</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>2018</year>. p. <fpage>154</fpage>&#x2013;<lpage>62</lpage>. doi:<pub-id pub-id-type="doi">10.1007/978-3-030-00928-1_18</pub-id>.</mixed-citation></ref>
<ref id="ref-147"><label>[147]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Quan</surname> <given-names>TM</given-names></string-name>, <string-name><surname>Nguyen-Duc</surname> <given-names>T</given-names></string-name>, <string-name><surname>Jeong</surname> <given-names>WK</given-names></string-name></person-group>. <article-title>Compressed sensing MRI reconstruction using a generative adversarial network with a cyclic loss</article-title>. <source>IEEE Transact Med Imag</source>. <year>2018</year>;<volume>37</volume>(<issue>6</issue>):<fpage>1488</fpage>&#x2013;<lpage>97</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2018.2820120</pub-id>; <pub-id pub-id-type="pmid">29870376</pub-id></mixed-citation></ref>
<ref id="ref-148"><label>[148]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Mardani</surname> <given-names>M</given-names></string-name>, <string-name><surname>Gong</surname> <given-names>E</given-names></string-name>, <string-name><surname>Cheng</surname> <given-names>JY</given-names></string-name>, <string-name><surname>Vasanawala</surname> <given-names>SS</given-names></string-name>, <string-name><surname>Zaharchuk</surname> <given-names>G</given-names></string-name>, <string-name><surname>Xing</surname> <given-names>L</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Deep generative adversarial neural networks for compressive sensing MRI</article-title>. <source>IEEE Transact Med Imag</source>. <year>2018</year>;<volume>38</volume>(<issue>1</issue>):<fpage>167</fpage>&#x2013;<lpage>79</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2018.2858752</pub-id>; <pub-id pub-id-type="pmid">30040634</pub-id></mixed-citation></ref>
<ref id="ref-149"><label>[149]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kang</surname> <given-names>E</given-names></string-name>, <string-name><surname>Koo</surname> <given-names>HJ</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>DH</given-names></string-name>, <string-name><surname>Seo</surname> <given-names>JB</given-names></string-name>, <string-name><surname>Ye</surname> <given-names>JC</given-names></string-name></person-group>. <article-title>Cycle-consistent adversarial denoising network for multiphase coronary CT angiography</article-title>. <source>Med Phy</source>. <year>2019</year>;<volume>46</volume>(<issue>2</issue>):<fpage>550</fpage>&#x2013;<lpage>62</lpage>. doi:<pub-id pub-id-type="doi">10.1002/mp.13284</pub-id>; <pub-id pub-id-type="pmid">30449055</pub-id></mixed-citation></ref>
<ref id="ref-150"><label>[150]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Bhadra</surname> <given-names>S</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>W</given-names></string-name>, <string-name><surname>Anastasio</surname> <given-names>MA</given-names></string-name></person-group>. <chapter-title>Medical image reconstruction with image-adaptive priors learned by use of generative adversarial networks</chapter-title>. In: <source>Medical imaging 2020: physics of medical imaging</source>. Vol. <volume>11312</volume>. <publisher-loc>Bellingham, WA, USA</publisher-loc>: <publisher-name>SPIE</publisher-name>; <year>2020</year>. p. <fpage>206</fpage>&#x2013;<lpage>13</lpage>.</mixed-citation></ref>
<ref id="ref-151"><label>[151]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Rashid</surname> <given-names>ZM</given-names></string-name>, <string-name><surname>Alsawaff</surname> <given-names>ZH</given-names></string-name>, <string-name><surname>Yousif</surname> <given-names>AS</given-names></string-name>, <string-name><surname>Al-Nima</surname> <given-names>RRO</given-names></string-name></person-group>. <article-title>Optimization of PET image reconstruction for enhanced image quality in various tasks using a conventional PET scanner</article-title>. <source>J Elect Comput Eng</source>. <year>2025</year>;<volume>2025</volume>(<issue>1</issue>):<fpage>8108611</fpage>. doi:<pub-id pub-id-type="doi">10.1155/jece/8108611</pub-id>.</mixed-citation></ref>
<ref id="ref-152"><label>[152]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ahn</surname> <given-names>G</given-names></string-name>, <string-name><surname>Choi</surname> <given-names>BS</given-names></string-name>, <string-name><surname>Ko</surname> <given-names>S</given-names></string-name>, <string-name><surname>Jo</surname> <given-names>C</given-names></string-name>, <string-name><surname>Han</surname> <given-names>HS</given-names></string-name>, <string-name><surname>Lee</surname> <given-names>MC</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>High-resolution knee plain radiography image synthesis using style generative adversarial network adaptive discriminator augmentation</article-title>. <source>J Orthop Res&#x00AE;</source>. <year>2023</year>;<volume>41</volume>(<issue>1</issue>):<fpage>84</fpage>&#x2013;<lpage>93</lpage>. doi:<pub-id pub-id-type="doi">10.1002/jor.25325</pub-id>; <pub-id pub-id-type="pmid">35293648</pub-id></mixed-citation></ref>
<ref id="ref-153"><label>[153]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zhu</surname> <given-names>L</given-names></string-name>, <string-name><surname>He</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Huang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Zeng</surname> <given-names>J</given-names></string-name>, <string-name><surname>Lu</surname> <given-names>L</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>DualMMP-GAN: dual-scale multi-modality perceptual generative adversarial network for medical image segmentation</article-title>. <source>Comput Biol Med</source>. <year>2022</year>;<volume>144</volume>(<issue>10</issue>):<fpage>105387</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.compbiomed.2022.105387</pub-id>; <pub-id pub-id-type="pmid">35305502</pub-id></mixed-citation></ref>
<ref id="ref-154"><label>[154]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Che Azemin</surname> <given-names>MZ</given-names></string-name>, <string-name><surname>Mohd Tamrin</surname> <given-names>MI</given-names></string-name>, <string-name><surname>Hilmi</surname> <given-names>MR</given-names></string-name>, <string-name><surname>Mohd Kamal</surname> <given-names>K</given-names></string-name></person-group>. <article-title>Assessing the efficacy of StyleGAN 3 in generating realistic medical images with limited data availability</article-title>. In: <conf-name>Proceedings of the 2024 13th International Conference on Software and Computer Applications; 2024 Feb 1&#x2013;3</conf-name>; <publisher-loc>Bali Island, Indonesia</publisher-loc>. p. <fpage>192</fpage>&#x2013;<lpage>7</lpage>.</mixed-citation></ref>
<ref id="ref-155"><label>[155]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Zhao</surname> <given-names>A</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>M</given-names></string-name>, <string-name><surname>Shahin</surname> <given-names>AH</given-names></string-name>, <string-name><surname>Wuyts</surname> <given-names>W</given-names></string-name>, <string-name><surname>Jones</surname> <given-names>MG</given-names></string-name>, <string-name><surname>Jacob</surname> <given-names>J</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>4D VQ-GAN: synthesising medical scans at any time point for personalised disease progression modelling of idiopathic pulmonary fibrosis</article-title>. <comment>arXiv:2502.05713. 2025</comment>.</mixed-citation></ref>
<ref id="ref-156"><label>[156]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Antoniou</surname> <given-names>A</given-names></string-name>, <string-name><surname>Storkey</surname> <given-names>A</given-names></string-name>, <string-name><surname>Edwards</surname> <given-names>H</given-names></string-name></person-group>. <article-title>Data augmentation generative adversarial networks</article-title>. <comment>arXiv:1711.04340. 2017</comment>.</mixed-citation></ref>
<ref id="ref-157"><label>[157]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Frid-Adar</surname> <given-names>M</given-names></string-name>, <string-name><surname>Klang</surname> <given-names>E</given-names></string-name>, <string-name><surname>Amitai</surname> <given-names>M</given-names></string-name>, <string-name><surname>Goldberger</surname> <given-names>J</given-names></string-name>, <string-name><surname>Greenspan</surname> <given-names>H</given-names></string-name></person-group>. <article-title>Synthetic data augmentation using GAN for improved liver lesion classification</article-title>. In: <conf-name>IEEE 15th International Symposium On Biomedical Imaging (ISBI 2018); 2018 Apr 4&#x2013;7</conf-name>; <publisher-loc>Washington, DC, USA</publisher-loc>; <year>2018</year>. p. <fpage>289</fpage>&#x2013;<lpage>93</lpage>.</mixed-citation></ref>
<ref id="ref-158"><label>[158]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Gan</surname> <given-names>HS</given-names></string-name>, <string-name><surname>Ramlee</surname> <given-names>MH</given-names></string-name>, <string-name><surname>Al-Rimy</surname> <given-names>BAS</given-names></string-name>, <string-name><surname>Lee</surname> <given-names>YS</given-names></string-name>, <string-name><surname>Akkaraekthalin</surname> <given-names>P</given-names></string-name></person-group>. <article-title>Hierarchical knee image synthesis framework for generative adversarial network: data from the osteoarthritis initiative</article-title>. <source>IEEE Access</source>. <year>2022</year>;<volume>10</volume>:<fpage>55051</fpage>&#x2013;<lpage>61</lpage>. doi:<pub-id pub-id-type="doi">10.1109/access.2022.3175506</pub-id>.</mixed-citation></ref>
<ref id="ref-159"><label>[159]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Bowles</surname> <given-names>C</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>L</given-names></string-name>, <string-name><surname>Guerrero</surname> <given-names>R</given-names></string-name>, <string-name><surname>Bentley</surname> <given-names>P</given-names></string-name>, <string-name><surname>Gunn</surname> <given-names>R</given-names></string-name>, <string-name><surname>Hammers</surname> <given-names>A</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Gan augmentation: augmenting training data using generative adversarial networks</article-title>. <comment>arXiv:1810.10863. 2018</comment>.</mixed-citation></ref>
<ref id="ref-160"><label>[160]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kong</surname> <given-names>L</given-names></string-name>, <string-name><surname>Lian</surname> <given-names>C</given-names></string-name>, <string-name><surname>Huang</surname> <given-names>D</given-names></string-name>, <string-name><surname>Hu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>Q</given-names></string-name></person-group>. <article-title>Breaking the dilemma of medical image-to-image translation</article-title>. <source>Adv Neural Inform Process Syst</source>. <year>2021</year>;<volume>34</volume>:<fpage>1964</fpage>&#x2013;<lpage>78</lpage>.</mixed-citation></ref>
<ref id="ref-161"><label>[161]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Du</surname> <given-names>W</given-names></string-name>, <string-name><surname>Tian</surname> <given-names>S</given-names></string-name></person-group>. <article-title>Transformer and GAN-based super-resolution reconstruction network for medical images</article-title>. <source>Tsinghua Sci Technol</source>. <year>2024</year>;<volume>29</volume>(<issue>1</issue>):<fpage>197</fpage>&#x2013;<lpage>206</lpage>. doi:<pub-id pub-id-type="doi">10.26599/tst.2022.9010071</pub-id>.</mixed-citation></ref>
<ref id="ref-162"><label>[162]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zuo</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Tian</surname> <given-names>H</given-names></string-name>, <string-name><surname>Li</surname> <given-names>R</given-names></string-name>, <string-name><surname>Guo</surname> <given-names>J</given-names></string-name>, <string-name><surname>Hu</surname> <given-names>J</given-names></string-name>, <string-name><surname>Tang</surname> <given-names>L</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Hemisphere-separated cross-connectome aggregating learning via VAE-GAN for brain structural connectivity synthesis</article-title>. <source>IEEE Access</source>. <year>2023</year>;<volume>11</volume>:<fpage>48493</fpage>&#x2013;<lpage>505</lpage>. doi:<pub-id pub-id-type="doi">10.1109/access.2023.3276989</pub-id>.</mixed-citation></ref>
<ref id="ref-163"><label>[163]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zhang</surname> <given-names>R</given-names></string-name>, <string-name><surname>Lu</surname> <given-names>W</given-names></string-name>, <string-name><surname>Gao</surname> <given-names>J</given-names></string-name>, <string-name><surname>Tian</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Wei</surname> <given-names>X</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>C</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>RFI-GAN: a reference-guided fuzzy integral network for ultrasound image augmentation</article-title>. <source>Inform Sci</source>. <year>2023</year>;<volume>623</volume>(<issue>5</issue>):<fpage>709</fpage>&#x2013;<lpage>28</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.ins.2022.12.026</pub-id>.</mixed-citation></ref>
<ref id="ref-164"><label>[164]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zhang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>J</given-names></string-name>, <string-name><surname>Feng</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Wee</surname> <given-names>L</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>GAN-based one dimensional medical data augmentation</article-title>. <source>Soft Comput</source>. <year>2023</year>;<volume>27</volume>(<issue>15</issue>):<fpage>10481</fpage>&#x2013;<lpage>91</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s00500-023-08345-z</pub-id>.</mixed-citation></ref>
<ref id="ref-165"><label>[165]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Showrov</surname> <given-names>I</given-names></string-name>, <string-name><surname>Aziz</surname> <given-names>MT</given-names></string-name>, <string-name><surname>Nabil</surname> <given-names>HR</given-names></string-name>, <string-name><surname>Jim</surname> <given-names>JR</given-names></string-name>, <string-name><surname>Kabir</surname> <given-names>MM</given-names></string-name>, <string-name><surname>Mridha</surname> <given-names>M</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Generative adversarial networks (GANs) in medical imaging: advancements, applications and challenges</article-title>. <source>IEEE Access</source>. <year>2024</year>;<volume>12</volume>:<fpage>35728</fpage>&#x2013;<lpage>53</lpage>. doi:<pub-id pub-id-type="doi">10.1109/access.2024.3370848</pub-id>.</mixed-citation></ref>
<ref id="ref-166"><label>[166]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Saad</surname> <given-names>MM</given-names></string-name>, <string-name><surname>O&#x2019;Reilly</surname> <given-names>R</given-names></string-name>, <string-name><surname>Rehmani</surname> <given-names>MH</given-names></string-name></person-group>. <article-title>A survey on training challenges in generative adversarial networks for biomedical image analysis</article-title>. <source>Artif Intell Rev</source>. <year>2024</year>;<volume>57</volume>(<issue>2</issue>):<fpage>19</fpage>. doi:<pub-id pub-id-type="doi">10.1007/s10462-023-10624-y</pub-id>.</mixed-citation></ref>
<ref id="ref-167"><label>[167]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Biswas</surname> <given-names>S</given-names></string-name>, <string-name><surname>Rohdin</surname> <given-names>J</given-names></string-name>, <string-name><surname>Drahansk&#x1EF3;</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Synthetic retinal images from unconditional GANs</article-title>. In: <conf-name>2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC); 2019 Jul 23&#x2013;27</conf-name>; <publisher-loc>Berlin, Germany</publisher-loc>. p. <fpage>2736</fpage>&#x2013;<lpage>9</lpage>.</mixed-citation></ref>
<ref id="ref-168"><label>[168]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Abdelhalim</surname> <given-names>ISA</given-names></string-name>, <string-name><surname>Mohamed</surname> <given-names>MF</given-names></string-name>, <string-name><surname>Mahdy</surname> <given-names>YB</given-names></string-name></person-group>. <article-title>Data augmentation for skin lesion using self-attention based progressive generative adversarial network</article-title>. <source>Expert Syst Applicat</source>. <year>2021</year>;<volume>165</volume>(<issue>3</issue>):<fpage>113922</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.eswa.2020.113922</pub-id>.</mixed-citation></ref>
<ref id="ref-169"><label>[169]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Goel</surname> <given-names>T</given-names></string-name>, <string-name><surname>Murugan</surname> <given-names>R</given-names></string-name>, <string-name><surname>Mirjalili</surname> <given-names>S</given-names></string-name>, <string-name><surname>Chakrabartty</surname> <given-names>DK</given-names></string-name></person-group>. <article-title>Automatic screening of covid-19 using an optimized generative adversarial network</article-title>. <source>Cognit Computat</source>. <year>2024</year>;<volume>16</volume>(<issue>4</issue>):<fpage>1666</fpage>&#x2013;<lpage>81</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s12559-020-09785-7</pub-id>; <pub-id pub-id-type="pmid">33520007</pub-id></mixed-citation></ref>
<ref id="ref-170"><label>[170]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>Z</given-names></string-name>, <string-name><surname>She</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Ward</surname> <given-names>TE</given-names></string-name></person-group>. <article-title>Generative adversarial networks in computer vision: a survey and taxonomy</article-title>. <source>ACM Comput Surv (CSUR)</source>. <year>2021</year>;<volume>54</volume>(<issue>2</issue>):<fpage>1</fpage>&#x2013;<lpage>38</lpage>. doi:<pub-id pub-id-type="doi">10.1145/3439723</pub-id>.</mixed-citation></ref>
<ref id="ref-171"><label>[171]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Heusel</surname> <given-names>M</given-names></string-name>, <string-name><surname>Ramsauer</surname> <given-names>H</given-names></string-name>, <string-name><surname>Unterthiner</surname> <given-names>T</given-names></string-name>, <string-name><surname>Nessler</surname> <given-names>B</given-names></string-name>, <string-name><surname>Hochreiter</surname> <given-names>S</given-names></string-name></person-group>. <chapter-title>Gans trained by a two time-scale update rule converge to a local nash equilibrium</chapter-title>. In: <source>Advances in neural information processing systems</source>. <publisher-loc>Cambridge, MA, USA</publisher-loc>: <publisher-name>The MIT Press</publisher-name>; <year>2017</year>. doi:<pub-id pub-id-type="doi">10.18034/ajase.v8i1.9</pub-id>.</mixed-citation></ref>
<ref id="ref-172"><label>[172]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Mirjalili</surname> <given-names>S</given-names></string-name>, <string-name><surname>Lewis</surname> <given-names>A</given-names></string-name></person-group>. <article-title>The whale optimization algorithm</article-title>. <source>Adv Eng Soft</source>. <year>2016</year>;<volume>95</volume>(<issue>12</issue>):<fpage>51</fpage>&#x2013;<lpage>67</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.advengsoft.2016.01.008</pub-id>.</mixed-citation></ref>
<ref id="ref-173"><label>[173]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Dukler</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Li</surname> <given-names>W</given-names></string-name>, <string-name><surname>Lin</surname> <given-names>A</given-names></string-name>, <string-name><surname>Mont&#x00FA;far</surname> <given-names>G</given-names></string-name></person-group>. <article-title>Wasserstein of Wasserstein loss for learning generative models</article-title>. In: <conf-name>The 36th International Conference on Machine Learning; 2019 Jun 10&#x2013;15</conf-name>; <publisher-loc>Long Beach, CA, USA</publisher-loc>. p. <fpage>1716</fpage>&#x2013;<lpage>25</lpage>.</mixed-citation></ref>
<ref id="ref-174"><label>[174]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Laino</surname> <given-names>ME</given-names></string-name>, <string-name><surname>Cancian</surname> <given-names>P</given-names></string-name>, <string-name><surname>Politi</surname> <given-names>LS</given-names></string-name>, <string-name><surname>Della Porta</surname> <given-names>MG</given-names></string-name>, <string-name><surname>Saba</surname> <given-names>L</given-names></string-name>, <string-name><surname>Savevski</surname> <given-names>V</given-names></string-name></person-group>. <article-title>Generative adversarial networks in brain imaging: a narrative review</article-title>. <source>J Imaging</source>. <year>2022</year>;<volume>8</volume>(<issue>4</issue>):<fpage>83</fpage>. doi:<pub-id pub-id-type="doi">10.3390/jimaging8040083</pub-id>; <pub-id pub-id-type="pmid">35448210</pub-id></mixed-citation></ref>
<ref id="ref-175"><label>[175]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wolterink</surname> <given-names>JM</given-names></string-name>, <string-name><surname>Mukhopadhyay</surname> <given-names>A</given-names></string-name>, <string-name><surname>Leiner</surname> <given-names>T</given-names></string-name>, <string-name><surname>Vogl</surname> <given-names>TJ</given-names></string-name>, <string-name><surname>Bucher</surname> <given-names>AM</given-names></string-name>, <string-name><surname>I&#x0161;gum</surname> <given-names>I</given-names></string-name></person-group>. <article-title>Generative adversarial networks: a primer for radiologists</article-title>. <source>Radiographics</source>. <year>2021</year>;<volume>41</volume>(<issue>3</issue>):<fpage>840</fpage>&#x2013;<lpage>57</lpage>. doi:<pub-id pub-id-type="doi">10.1148/rg.2021200151</pub-id>; <pub-id pub-id-type="pmid">33891522</pub-id></mixed-citation></ref>
<ref id="ref-176"><label>[176]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Lee</surname> <given-names>M</given-names></string-name>, <string-name><surname>Seok</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Regularization methods for generative adversarial networks: an overview of recent studies</article-title>. <comment>arXiv:2005.09165. 2020</comment>.</mixed-citation></ref>
<ref id="ref-177"><label>[177]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Miyato</surname> <given-names>T</given-names></string-name>, <string-name><surname>Kataoka</surname> <given-names>T</given-names></string-name>, <string-name><surname>Koyama</surname> <given-names>M</given-names></string-name>, <string-name><surname>Yoshida</surname> <given-names>Y</given-names></string-name></person-group>. <article-title>Spectral normalization for generative adversarial networks</article-title>. <comment>arXiv:1802.05957. 2018</comment>.</mixed-citation></ref>
<ref id="ref-178"><label>[178]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Klambauer</surname> <given-names>G</given-names></string-name>, <string-name><surname>Unterthiner</surname> <given-names>T</given-names></string-name>, <string-name><surname>Mayr</surname> <given-names>A</given-names></string-name>, <string-name><surname>Hochreiter</surname> <given-names>S</given-names></string-name></person-group>. <chapter-title>Self-normalizing neural networks</chapter-title>. In: <source>Advances in neural information processing systems</source>. <publisher-loc>Cambridge, MA, USA</publisher-loc>: <publisher-name>The MIT Press</publisher-name>; <year>2017</year>.</mixed-citation></ref>
<ref id="ref-179"><label>[179]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Xu</surname> <given-names>L</given-names></string-name>, <string-name><surname>Zeng</surname> <given-names>X</given-names></string-name>, <string-name><surname>Huang</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Li</surname> <given-names>W</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>H</given-names></string-name></person-group>. <article-title>Low-dose chest X-ray image super-resolution using generative adversarial nets with spectral normalization</article-title>. <source>Biomed Signal Process Cont</source>. <year>2020</year>;<volume>55</volume>(<issue>6</issue>):<fpage>101600</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.bspc.2019.101600</pub-id>.</mixed-citation></ref>
<ref id="ref-180"><label>[180]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Hoang</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Nguyen</surname> <given-names>TD</given-names></string-name>, <string-name><surname>Le</surname> <given-names>T</given-names></string-name>, <string-name><surname>Phung</surname> <given-names>D</given-names></string-name></person-group>. <article-title>MGAN: training generative adversarial nets with multiple generators</article-title>. In: <conf-name>The 6th International Conference on Learning Representations; 2018 Apr 30&#x2013;May 3</conf-name>; <publisher-loc>Vancouver, BC, Canada</publisher-loc>. p. <fpage>1</fpage>&#x2013;<lpage>24</lpage>.</mixed-citation></ref>
<ref id="ref-181"><label>[181]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Wu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Yue</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Tan</surname> <given-names>X</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>W</given-names></string-name>, <string-name><surname>Lu</surname> <given-names>T</given-names></string-name></person-group>. <article-title>End-to-end chromosome Karyotyping with data augmentation using GAN</article-title>. In: <conf-name>2018 25th IEEE International Conference on Image Processing (ICIP); 2018 Oct 7&#x2013;10</conf-name>; <publisher-loc>Athens, Greece</publisher-loc>. p. <fpage>2456</fpage>&#x2013;<lpage>60</lpage>.</mixed-citation></ref>
<ref id="ref-182"><label>[182]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Neff</surname> <given-names>T</given-names></string-name>, <string-name><surname>Payer</surname> <given-names>C</given-names></string-name>, <string-name><surname>Stern</surname> <given-names>D</given-names></string-name>, <string-name><surname>Urschler</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Generative adversarial network based synthesis for supervised medical image segmentation</article-title>. In: <conf-name>Proceedings of the OAGM and ARW Joint Workshop; 2017 May 10&#x2013;12</conf-name>; <publisher-loc>Wien, Austria</publisher-loc>. p. <fpage>140</fpage>&#x2013;<lpage>5</lpage>.</mixed-citation></ref>
<ref id="ref-183"><label>[183]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Saad</surname> <given-names>MM</given-names></string-name>, <string-name><surname>Rehmani</surname> <given-names>MH</given-names></string-name>, <string-name><surname>O&#x2019;Reilly</surname> <given-names>R</given-names></string-name></person-group>. <article-title>Addressing the intra-class mode collapse problem using adaptive input image normalization in gan-based x-ray images</article-title>. In: <conf-name>2022 44th Annual International Conference of the IEEE Engineering in Medicine &#x0026; Biology Society (EMBC); 2022 Jul 11&#x2013;15</conf-name>; <publisher-loc>Glasgow, UK</publisher-loc>. p. <fpage>2049</fpage>&#x2013;<lpage>52</lpage>.</mixed-citation></ref>
<ref id="ref-184"><label>[184]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Saad</surname> <given-names>MM</given-names></string-name>, <string-name><surname>Rehmani</surname> <given-names>MH</given-names></string-name>, <string-name><surname>O&#x2019;Reilly</surname> <given-names>R</given-names></string-name></person-group>. <article-title>A self-attention guided multi-scale gradient GAN for diversified x-ray image synthesis</article-title>. In: <conf-name>Irish Conference on Artificial Intelligence and Cognitive Science; 2022 Dec 8&#x2013;9</conf-name>; <publisher-loc>Munster, Ireland</publisher-loc>. p. <fpage>18</fpage>&#x2013;<lpage>31</lpage>.</mixed-citation></ref>
<ref id="ref-185"><label>[185]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Xue</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Ye</surname> <given-names>J</given-names></string-name>, <string-name><surname>Long</surname> <given-names>LR</given-names></string-name>, <string-name><surname>Antani</surname> <given-names>S</given-names></string-name>, <string-name><surname>Cornwell</surname> <given-names>C</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Synthetic augmentation and feature-based filtering for improved cervical histopathology image classification</article-title>. In: <conf-name>International Conference on Medical Image Computing and Computer-Assisted Intervention; 2019 Oct 13&#x2013;17</conf-name>; <publisher-loc>Shenzhen, China</publisher-loc>: <publisher-name>Springer</publisher-name>. p. <fpage>387</fpage>&#x2013;<lpage>96</lpage>.</mixed-citation></ref>
<ref id="ref-186"><label>[186]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Kudo</surname> <given-names>A</given-names></string-name>, <string-name><surname>Kitamura</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Li</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Iizuka</surname> <given-names>S</given-names></string-name>, <string-name><surname>Simo-Serra</surname> <given-names>E</given-names></string-name></person-group>. <article-title>Virtual thin slice: 3D conditional GAN-based super-resolution for CT slice interval</article-title>. In: <conf-name>International Workshop on Machine Learning for Medical Image Reconstruction; 2019 Oct 17</conf-name>; <publisher-loc>Shenzhen, China</publisher-loc>. p. <fpage>91</fpage>&#x2013;<lpage>100</lpage>. doi:<pub-id pub-id-type="doi">10.1007/978-3-030-33843-5_9</pub-id>.</mixed-citation></ref>
<ref id="ref-187"><label>[187]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Segato</surname> <given-names>A</given-names></string-name>, <string-name><surname>Corbetta</surname> <given-names>V</given-names></string-name>, <string-name><surname>Di Marzo</surname> <given-names>M</given-names></string-name>, <string-name><surname>Pozzi</surname> <given-names>L</given-names></string-name>, <string-name><surname>De Momi</surname> <given-names>E</given-names></string-name></person-group>. <article-title>Data augmentation of 3D brain environment using deep convolutional refined auto-encoding alpha GAN</article-title>. <source>IEEE Transact Med Robot Bionics</source>. <year>2020</year>;<volume>3</volume>(<issue>1</issue>):<fpage>269</fpage>&#x2013;<lpage>72</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmrb.2020.3045230</pub-id>.</mixed-citation></ref>
<ref id="ref-188"><label>[188]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Kwon</surname> <given-names>G</given-names></string-name>, <string-name><surname>Han</surname> <given-names>C</given-names></string-name>, <string-name><surname>Kim</surname> <given-names>DS</given-names></string-name></person-group>. <article-title>Generation of 3D brain MRI using auto-encoding generative adversarial networks</article-title>. In: <conf-name>International Conference on Medical Image Computing and Computer-Assisted Intervention; 2019 Oct 13&#x2013;17</conf-name>; <publisher-loc>Shenzhen, China</publisher-loc>. p. <fpage>118</fpage>&#x2013;<lpage>26</lpage>.</mixed-citation></ref>
<ref id="ref-189"><label>[189]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Qin</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Zhu</surname> <given-names>P</given-names></string-name>, <string-name><surname>Xue</surname> <given-names>Y</given-names></string-name></person-group>. <article-title>A GAN-based image synthesis method for skin lesion classification</article-title>. <source>Comput Meth Prog Biomed</source>. <year>2020</year>;<volume>195</volume>(<issue>2</issue>):<fpage>105568</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.cmpb.2020.105568</pub-id>; <pub-id pub-id-type="pmid">32526536</pub-id></mixed-citation></ref>
<ref id="ref-190"><label>[190]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Modanwal</surname> <given-names>G</given-names></string-name>, <string-name><surname>Vellal</surname> <given-names>A</given-names></string-name>, <string-name><surname>Mazurowski</surname> <given-names>MA</given-names></string-name></person-group>. <article-title>Normalization of breast MRIs using cycle-consistent generative adversarial networks</article-title>. <source>Comput Meth Prog Biomed</source>. <year>2021</year>;<volume>208</volume>(<issue>4</issue>):<fpage>106225</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.cmpb.2021.106225</pub-id>; <pub-id pub-id-type="pmid">34198016</pub-id></mixed-citation></ref>
<ref id="ref-191"><label>[191]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Deng</surname> <given-names>J</given-names></string-name>, <string-name><surname>Dong</surname> <given-names>W</given-names></string-name>, <string-name><surname>Socher</surname> <given-names>R</given-names></string-name>, <string-name><surname>Li</surname> <given-names>LJ</given-names></string-name>, <string-name><surname>Li</surname> <given-names>K</given-names></string-name>, <string-name><surname>Fei-Fei</surname> <given-names>L</given-names></string-name></person-group>. <article-title>Imagenet: a large-scale hierarchical image database</article-title>. In: <conf-name>2009 IEEE Conference on Computer Vision and Pattern Recognition; 2009 Jun 20&#x2013;25</conf-name>; <publisher-loc>Miami, FL, USA</publisher-loc>. p. <fpage>248</fpage>&#x2013;<lpage>55</lpage>.</mixed-citation></ref>
<ref id="ref-192"><label>[192]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Sara</surname> <given-names>U</given-names></string-name>, <string-name><surname>Akter</surname> <given-names>M</given-names></string-name>, <string-name><surname>Uddin</surname> <given-names>MS</given-names></string-name></person-group>. <article-title>Image quality assessment through FSIM, SSIM, MSE and PSNR&#x2014;a comparative study</article-title>. <source>J Comput Communicat</source>. <year>2019</year>;<volume>7</volume>(<issue>3</issue>):<fpage>8</fpage>&#x2013;<lpage>18</lpage>. doi:<pub-id pub-id-type="doi">10.4236/jcc.2019.73002</pub-id>.</mixed-citation></ref>
<ref id="ref-193"><label>[193]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Korkinof</surname> <given-names>D</given-names></string-name>, <string-name><surname>Harvey</surname> <given-names>H</given-names></string-name>, <string-name><surname>Heindl</surname> <given-names>A</given-names></string-name>, <string-name><surname>Karpati</surname> <given-names>E</given-names></string-name>, <string-name><surname>Williams</surname> <given-names>G</given-names></string-name>, <string-name><surname>Rijken</surname> <given-names>T</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Perceived realism of high-resolution generative adversarial network-derived synthetic mammograms</article-title>. <source>Radi Artif Intell</source>. <year>2021</year>;<volume>3</volume>(<issue>2</issue>):<fpage>e190181</fpage>. doi:<pub-id pub-id-type="doi">10.1148/ryai.2020190181</pub-id>; <pub-id pub-id-type="pmid">33937856</pub-id></mixed-citation></ref>
<ref id="ref-194"><label>[194]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Lim</surname> <given-names>G</given-names></string-name>, <string-name><surname>Ng</surname> <given-names>WY</given-names></string-name>, <string-name><surname>Tan</surname> <given-names>TE</given-names></string-name>, <string-name><surname>Lim</surname> <given-names>J</given-names></string-name>, <string-name><surname>Lim</surname> <given-names>SH</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Synthetic artificial intelligence using generative adversarial network for retinal imaging in detection of age-related macular degeneration</article-title>. <source>Front Med</source>. <year>2023</year>;<volume>10</volume>:<fpage>1184892</fpage>. doi:<pub-id pub-id-type="doi">10.3389/fmed.2023.1184892</pub-id>; <pub-id pub-id-type="pmid">37425325</pub-id></mixed-citation></ref>
<ref id="ref-195"><label>[195]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Mardani</surname> <given-names>M</given-names></string-name>, <string-name><surname>Gong</surname> <given-names>E</given-names></string-name>, <string-name><surname>Cheng</surname> <given-names>JY</given-names></string-name>, <string-name><surname>Vasanawala</surname> <given-names>S</given-names></string-name>, <string-name><surname>Zaharchuk</surname> <given-names>G</given-names></string-name>, <string-name><surname>Alley</surname> <given-names>M</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Deep generative adversarial networks for compressed sensing automates MRI</article-title>. <comment>arXiv:1706.00051. 2017</comment>.</mixed-citation></ref>
<ref id="ref-196"><label>[196]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Li</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>H</given-names></string-name>, <string-name><surname>Xie</surname> <given-names>D</given-names></string-name>, <string-name><surname>Dreizin</surname> <given-names>D</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>F</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>Z</given-names></string-name></person-group>. <article-title>POCS-augmented CycleGAN for MR image reconstruction</article-title>. <source>Appl Sci</source>. <year>2021</year>;<volume>12</volume>(<issue>1</issue>):<fpage>114</fpage>. doi:<pub-id pub-id-type="doi">10.3390/app12010114</pub-id>; <pub-id pub-id-type="pmid">37465648</pub-id></mixed-citation></ref>
<ref id="ref-197"><label>[197]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Ehrhardt</surname> <given-names>J</given-names></string-name>, <string-name><surname>Wilms</surname> <given-names>M</given-names></string-name></person-group>. <chapter-title>Autoencoders and variational autoencoders in medical image analysis</chapter-title>. In: <person-group person-group-type="editor"><string-name><surname>Burgos</surname> <given-names>N</given-names></string-name>, <string-name><surname>Svoboda</surname> <given-names>D</given-names></string-name></person-group>, editors. <source>Biomedical image synthesis and simulation: methods and applications</source>. <publisher-loc>Cambridge, MA, USA</publisher-loc>: <publisher-name>Academic Press</publisher-name>; <year>2022</year>. p. <fpage>129</fpage>&#x2013;<lpage>62</lpage>. doi:<pub-id pub-id-type="doi">10.1016/b978-0-12-824349-7.00015-3</pub-id>.</mixed-citation></ref>
<ref id="ref-198"><label>[198]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kazerouni</surname> <given-names>A</given-names></string-name>, <string-name><surname>Aghdam</surname> <given-names>EK</given-names></string-name>, <string-name><surname>Heidari</surname> <given-names>M</given-names></string-name>, <string-name><surname>Azad</surname> <given-names>R</given-names></string-name>, <string-name><surname>Fayyaz</surname> <given-names>M</given-names></string-name>, <string-name><surname>Hacihaliloglu</surname> <given-names>I</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Diffusion models in medical imaging: a comprehensive survey</article-title>. <source>Med Image Anal</source>. <year>2023</year>;<volume>88</volume>(<issue>1</issue>):<fpage>102846</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.media.2023.102846</pub-id>; <pub-id pub-id-type="pmid">37295311</pub-id></mixed-citation></ref>
</ref-list>
</back></article>












