<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xml:lang="en" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CMES</journal-id>
<journal-id journal-id-type="nlm-ta">CMES</journal-id>
<journal-id journal-id-type="publisher-id">CMES</journal-id>
<journal-title-group>
<journal-title>Computer Modeling in Engineering &#x0026; Sciences</journal-title>
</journal-title-group>
<issn pub-type="epub">1526-1506</issn>
<issn pub-type="ppub">1526-1492</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">73387</article-id>
<article-id pub-id-type="doi">10.32604/cmes.2025.073387</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Enhancement of Medical Imaging Technique for Diabetic Retinopathy: Realistic Synthetic Image Generation Using GenAI</article-title>
<alt-title alt-title-type="left-running-head">Enhancement of Medical Imaging Technique for Diabetic Retinopathy: Realistic Synthetic Image Generation Using GenAI</alt-title>
<alt-title alt-title-type="right-running-head">Enhancement of Medical Imaging Technique for Diabetic Retinopathy: Realistic Synthetic Image Generation Using GenAI</alt-title>
</title-group>
<contrib-group>
<contrib id="author-1" contrib-type="author">
<name name-style="western"><surname>Palaniappan</surname><given-names>Damodharan</given-names></name><xref ref-type="aff" rid="aff-1">1</xref></contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western"><surname>Tak</surname><given-names>Tan Kuan</given-names></name><xref ref-type="aff" rid="aff-2">2</xref></contrib>
<contrib id="author-3" contrib-type="author">
<name name-style="western"><surname>Vijayan</surname><given-names>K.</given-names></name><xref ref-type="aff" rid="aff-3">3</xref></contrib>
<contrib id="author-4" contrib-type="author">
<name name-style="western"><surname>Maram</surname><given-names>Balajee</given-names></name><xref ref-type="aff" rid="aff-4">4</xref></contrib>
<contrib id="author-5" contrib-type="author">
<name name-style="western"><surname>Kshirsagar</surname><given-names>Pravin R</given-names></name><xref ref-type="aff" rid="aff-5">5</xref></contrib>
<contrib id="author-6" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Ahmad</surname><given-names>Naim</given-names></name><xref ref-type="aff" rid="aff-6">6</xref><xref rid="cor1" ref-type="corresp">&#x002A;</xref><email>nagqadir@kku.edu.sa</email></contrib>
<aff id="aff-1"><label>1</label><institution>Department of Information Technology, Marwadi University</institution>, <addr-line>Rajkot, 360003</addr-line>, <country>India</country></aff>
<aff id="aff-2"><label>2</label><institution>Engineering Cluster, Singapore Institute of Technology, Singapore</institution>, <addr-line>828608</addr-line>, <country>Singapore</country></aff>
<aff id="aff-3"><label>3</label><institution>Electronics and Communication Engineering Department, Sapthagiri NPS University</institution>, <addr-line>Bangalore, 560057</addr-line>, <country>India</country></aff>
<aff id="aff-4"><label>4</label><institution>School of Computer Science and Artificial Intelligence, SR University</institution>, <addr-line>Warangal, 506371</addr-line>, <country>India</country></aff>
<aff id="aff-5"><label>5</label><institution>Electronics Telecommunication Engineering, J D College of Engineering Management</institution>, <addr-line>Nagpur, 441501</addr-line>, <country>India</country></aff>
<aff id="aff-6"><label>6</label><institution>College of Computer Science, King Khalid University</institution>, <addr-line>Abha, 61421</addr-line>, <country>Saudi Arabia</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label>Corresponding Author: Naim Ahmad. Email: <email>nagqadir@kku.edu.sa</email></corresp>
</author-notes>
<pub-date date-type="collection" publication-format="electronic">
<year>2025</year>
</pub-date>
<pub-date date-type="pub" publication-format="electronic">
<day>23</day><month>12</month><year>2025</year>
</pub-date>
<volume>145</volume>
<issue>3</issue>
<fpage>4107</fpage>
<lpage>4127</lpage>
<history>
<date date-type="received">
<day>17</day>
<month>09</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>11</day>
<month>11</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2025 The Authors.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Published by Tech Science Press.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CMES_73387.pdf"></self-uri>
<abstract>
<p>A phase-aware cross-modal framework is presented that synthesizes UWF_FA from non-invasive UWF_RI for diabetic retinopathy (DR) stratification. A curated cohort of 1198 patients (2915 UWF_RI and 17,854 UWF_FA images) with strict registration quality supports training across three angiographic phases (initial, mid, final). The generator is based on a modified pix2pixHD with an added Gradient Variance Loss to better preserve microvasculature, and is evaluated using MAE, PSNR, SSIM, and MS-SSIM on held-out pairs. Quantitatively, the mid phase achieves the lowest MAE (98.76 <inline-formula id="ieqn-1"><mml:math id="mml-ieqn-1"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 42.67), while SSIM remains high across phases. Expert review shows substantial agreement (Cohen&#x2019;s <inline-formula id="ieqn-2"><mml:math id="mml-ieqn-2"><mml:mi>&#x03BA;</mml:mi></mml:math></inline-formula> &#x003D; 0.78&#x2013;0.82) and Turing-style misclassification of 50%&#x2013;70% of synthetic images as real, indicating strong perceptual realism. For downstream DR stratification, fusing multi-phase synthetic UWF_FA with UWF_RI in a Swin Transformer classifier yields significant gains over a UWF_RI-only baseline, with the full-phase setting (Set D) reaching AUC &#x003D; 0.910 and accuracy &#x003D; 0.829. These results support synthetic UWF_FA as a scalable, non-invasive complement to dye-based angiography that enhances screening accuracy while avoiding injection-related risks.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Diabetic retinopathy</kwd>
<kwd>synthetic image generation</kwd>
<kwd>GenAI</kwd>
<kwd>medical imaging</kwd>
<kwd>ultra-widefield retinal imaging</kwd>
<kwd>enhanced medical imaging datasets</kwd>
<kwd>multi-scale structural similarity</kwd>
</kwd-group>
<funding-group>
<award-group id="awg1">
<funding-source>Deanship of Research and Graduate Studies at King Khalid University through Large Research Project</funding-source>
<award-id>RGP2/417/46</award-id>
</award-group>
</funding-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>As a a common and deadly complication of diabetes mellitus, diabetic retinopathy (DR) is one of the leading causes of vision loss and blindness globally. According to the International Diabetes Federation, over 100 million people are currently affected by DR, and this number is expected to rise with the global diabetes burden, making early detection and management a public health priority. People in their middle years and later years seem to have it more often than younger generations [<xref ref-type="bibr" rid="ref-1">1</xref>]. People with diabetes often notice a gradual blurring of vision as the disease progresses; if left untreated, this might lead to permanent irreversible impairment. Diabetic retinopathy (DR) manifests itself in a variety of ways, but one of the most important is the development of microvascular abnormalities, which can have a major impact on retinal function [<xref ref-type="bibr" rid="ref-2">2</xref>]. Depending on the degree of damage, there are two main kinds of diabetic retinopathy (DR): non-proliferative diabetic retinopathy (NPDR) and proliferative diabetic retinopathy (PDR). In order to effectively treat and prevent vision loss, it is crucial to diagnose these phases early and accurately.</p>
<p>Ophthalmologists with extensive training are required to perform medical imaging procedures such as fundus photography and Optical Coherence Tomography (OCT), which are the backbone of the traditional diagnostic approach. However, due to the complexity and variation in image quality of retinal photographs, manual screening is both laborious and error prone. Moreover, fluorescein angiography (FA) the gold standard for visualizing retinal vasculature requires intravenous dye injection, which can cause side effects such as nausea, skin discoloration, and in rare cases severe allergic reactions; these risks limit routine use, especially in large-scale screening programs. Automating the diagnosis of diabetic retinopathy (DR) from raw retinal images has been made possible in recent years by ML techniques like Bayesian classifiers, K-Means clustering, Probabilistic Neural Networks (PNN), and Support Vector Machines (SVM) [<xref ref-type="bibr" rid="ref-3">3</xref>]. Methods such as these make use of blood vessels, haemorrhages, and exudates. Support vector machines (SVM) outperformed Bayes (94.4%) and PNN (89.6%), according to a comparative study that used 350 fundus pictures (100 for training and 250 for testing). This resulted in the highest classification accuracy (97.6%) [<xref ref-type="bibr" rid="ref-4">4</xref>]. Another evaluation using 130 images from the DIARETDB0 dataset gave more evidence of SVM&#x2019;s superior performance (95.38%) [<xref ref-type="bibr" rid="ref-5">5</xref>]. Despite these advancements, limitations in dataset diversity, imbalance, or data scarcity might occasionally impair the efficacy of deep learning and machine learning models. With this issue in mind, generative artificial intelligence (GenAI) has been getting a lot of attention, with a focus on GANs and diffusion models in particular. Generating synthetic images that closely resemble real patient data is within the capabilities of these AI kinds [<xref ref-type="bibr" rid="ref-6">6</xref>]. You may utilise these artificial images to enhance model generalisation, increase DR diagnosis accuracy, and refresh training datasets; they also solve big problems with data availability, privacy, and imbalance [<xref ref-type="bibr" rid="ref-7">7</xref>]. Plus, you can use them to make DR diagnostics work better.</p>
<p>Additionally, common image preprocessing techniques like denoising, edge detection, and contrast enhancement can miss small retinal features associated with DR. Modern DR classification methods are based on deep learning models such as Convolutional Neural Networks (CNNs). A limitation of these models is their reliance on massive annotated datasets, which limits their scalability [<xref ref-type="bibr" rid="ref-8">8</xref>]. The ability to generate varied and realistic retinal images across all stages of DR has led to the emergence of GenAI models, including conditional GANs (cGANs), CycleGANs, and diffusion models, as game-changing tools [<xref ref-type="bibr" rid="ref-9">9</xref>]. The training of diagnostic systems powered by artificial intelligence has been improved as a result [<xref ref-type="bibr" rid="ref-10">10</xref>]. Generative models not only enhance model training, but they also allow for the modelling of disease progression. For a better understanding of the disease, this allows researchers to build prediction models for DR staging and severity assessment. The ability to do so offers up new avenues for personalised and preventive treatment options [<xref ref-type="bibr" rid="ref-11">11</xref>]. Furthermore, synthetic data helps to allay privacy concerns, which in turn allows for more data interchange and collaboration among ophthalmic researchers, all while patients&#x2019; confidentiality is preserved [<xref ref-type="bibr" rid="ref-12">12</xref>].</p>
<p>In this work, we propose a GAN-based cross modal translation framework that synthesizes ultra-widefield fluorescein angiography (UWF_FA) directly from ultra widefield retinal images (UWF_RI). To the best of our knowledge, this is the first study to generate <bold>multi-phase synthetic UWF_FA</bold> (initial, mid, and final angiographic phases) and to validate both perceptual realism and clinical utility using quantitative metrics (MAE, PSNR, SSIM) and ophthalmologist Turing tests while integrating the synthetic phases with a Swin Transformer classifier for improved DR stratification. An overview of the proposed pipeline is shown in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>. By eliminating the risks of dye injection and preserving angiographic cues (nonperfusion, leakage, peripheral vasculature), our approach offers a scalable, non-invasive pathway to safer DR screening and triage.</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>Block diagram of the proposed cross-modal framework. UWF&#x2013;RI &#x003D; ultra-widefield retinal imaging; UWF&#x2013;FA &#x003D; ultra-widefield fluorescein angiography. The framework performs preprocessing and registration of UWF&#x2013;RI, phase-aware GAN synthesis of UWF&#x2013;FA (initial, mid, final), and fusion with a Swin Transformer classifier for DR stratification</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_73387-fig-1.tif"/>
</fig>
<p>Overall, DR screening methods might be improved by combining generative models with traditional and modern AI-based diagnostic techniques. This could lead to more accessible, efficient, and accurate screenings. Our study addresses a critical gap by delivering high fidelity, phase-aware synthetic angiography from non-invasive inputs, demonstrating measurable gains in downstream DR grading and pointing toward broader clinical deployment.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Related Work</title>
<p>Research on automated diabetic retinopathy (DR) analysis has evolved from classical feature engineering on colour fundus photographs to powerful deep models that segment vessels and lesions, grade disease severity, and increasingly aim to recreate angiography like information from non-invasive inputs [<xref ref-type="bibr" rid="ref-13">13</xref>]. Ultra widefield (UWF) modalities and angiography (FA, OCTA) capture nonperfusion, leakage, and peripheral pathology that standard views can miss, which is clinically important for risk stratification and treatment decisions [<xref ref-type="bibr" rid="ref-14">14</xref>]. This has motivated cross-modal synthesis (fundus or UWF retinal images to FA-like outputs) to support safer, scalable screening and triage while preserving clinically salient vascular cues [<xref ref-type="bibr" rid="ref-15">15</xref>].</p>
<sec id="s2_1">
<label>2.1</label>
<title>Traditional DR Classification Approaches</title>
<p>Early automated systems extracted hand-crafted descriptors vessel maps, microaneurysms, haemorrhages, and exudates from colour fundus images and applied SVMs, kNN, or Bayesian models to estimate DR severity. These pipelines required heavy preprocessing, were sensitive to illumination and device shifts, and often struggled on heterogeneous cohorts [<xref ref-type="bibr" rid="ref-16">16</xref>]. To alleviate data scarcity, works explored contrast enhancement and GAN-based augmentation to stabilise prognostication [<xref ref-type="bibr" rid="ref-17">17</xref>], but such strategies still operated on fundus-only inputs and lacked direct angiographic signals unless dye-based FA or OCTA was acquired [<xref ref-type="bibr" rid="ref-18">18</xref>]. Robust learning under imperfect labels (loss correction and noise-aware training) was also studied [<xref ref-type="bibr" rid="ref-19">19</xref>], and objective image-quality metrics such as SSIM, PSNR, MSE, and FSIM became standard for screening inputs and evaluating restorations [<xref ref-type="bibr" rid="ref-20">20</xref>].</p>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Deep Learning for DR Screening and Grading</title>
<p>Modern screening is dominated by convolutional and transformer-based architectures, spanning segmentation (vessel/lesion localisation) and end-to-end grading [<xref ref-type="bibr" rid="ref-21">21</xref>]. Lesion aware and multi-loss designs improved microvasculature delineation and robustness [<xref ref-type="bibr" rid="ref-22">22</xref>], while hybrid pipelines with vision transformers strengthened representation learning for synthesis and predict workflows [<xref ref-type="bibr" rid="ref-23">23</xref>]. Transfer learning and representation pretraining trends (including vision&#x2013;language supervision) further enhanced feature reuse across cohorts [<xref ref-type="bibr" rid="ref-24">24</xref>]. Persistent challenges include large annotation demands, class imbalance, and domain shift [<xref ref-type="bibr" rid="ref-25">25</xref>]. Diffusion models recently emerged as strong priors for denoising and augmentation on retinal imagery, improving fidelity and training stability [<xref ref-type="bibr" rid="ref-26">26</xref>]. Nevertheless, most pipelines are fundus-only and cannot directly model nonperfusion or leakage patterns that are best visualised on FA or OCTA; studies using UWF-FA demonstrate grading value but still require invasive dye [<xref ref-type="bibr" rid="ref-27">27</xref>].</p>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>Cross-Modal Translation and Generative Models for Retinal Imaging</title>
<p>Cross-modal retinal synthesis aims to recover angiography-like information from non-invasive images. Foundational conditional adversarial frameworks established paired image-to-image translation, and ophthalmic variants adapted architectures, objectives, and priors to encode retinal structure [<xref ref-type="bibr" rid="ref-28">28</xref>]. UWAT-GAN introduced an ultra-wide-angle, multi-scale generator to map UWF-RI to UWF-FA, but relied on only a few dozen matched pairs, limiting generalisation and phase modelling [<xref ref-type="bibr" rid="ref-29">29</xref>]. UWAFA-GAN increased data scale and integrated attention and registration enhancement to sharpen vasculature fidelity; however, it produced a single venous-phase frame and did not quantify downstream DR stratification gains. Other directions focused on targeted synthesis or controllability: lesion-centric DR-LL-GAN, Wasserstein-based retinal synthesis, controllable StyleGAN variants, and high-fidelity semantic manipulation. Some explored grading via the generative prior itself, but comprehensive multi-phase UWF-FA generation coupled with an integrated classifier and validated on downstream clinical tasks remains uncommon [<xref ref-type="bibr" rid="ref-30">30</xref>].</p>
</sec>
<sec id="s2_4">
<label>2.4</label>
<title>Generative AI in Broader Medical Imaging</title>
<p>Beyond ophthalmology, adversarial and diffusion paradigms matured for translation and restoration. Unsupervised adversarial diffusion enabled cross modality mapping without dense pairing [<xref ref-type="bibr" rid="ref-31">31</xref>], while image-to-image diffusion provided stable, controllable synthesis [<xref ref-type="bibr" rid="ref-32">32</xref>]. Disentangled and regularised GANs promoted factorised latent structure and interpretable controls [<xref ref-type="bibr" rid="ref-33">33</xref>]; surveys chart diffusion advantages in likelihood training, diversity, and perceptual quality [<xref ref-type="bibr" rid="ref-34">34</xref>]. At the same time, the community cautioned against overreliance on single realism metrics, advocating multi metric assessment [<xref ref-type="bibr" rid="ref-35">35</xref>] and task based validation principles directly relevant to retinal synthesis and evaluation [<xref ref-type="bibr" rid="ref-36">36</xref>]. Recent hybrid generative&#x2013;classification frameworks have further demonstrated the synergy between image synthesis and diagnostic modeling in diabetic retinopathy [<xref ref-type="bibr" rid="ref-37">37</xref>]. Gencer et al. (2025) combined GAN-based augmentation with denoising autoencoders and an EfficientNet-B0 classifier, achieving 99% accuracy, recall, and specificity on a high-resolution OCT dataset [<xref ref-type="bibr" rid="ref-38">38</xref>].</p>
</sec>
<sec id="s2_5">
<label>2.5</label>
<title>Clinical Context and Motivation</title>
<p>UWF imaging expands coverage to the periphery, where clinically significant lesions and nonperfusion can alter staging and management. Widefield OCTA studies link nonperfusion areas with DR severity and progression, underscoring the value of angiographic signals beyond colour fundus. Deep-learning pipelines trained directly on UWF-FA demonstrate strong grading potential, yet FA requires dye injection, with workflow implications and rare adverse reactions. These factors motivate non-invasive synthesis that preserves diagnostic utility while improving accessibility [<xref ref-type="bibr" rid="ref-39">39</xref>].</p>
</sec>
<sec id="s2_6">
<label>2.6</label>
<title>Gap Identification</title>
<p>In summary, existing retinal synthesis studies often train on limited paired data, restrict outputs to a single angiographic phase (<xref ref-type="table" rid="table-1">Table 1</xref>), or focus on narrow fields and lesion slices; few quantify downstream diagnostic gains with an integrated classifier. Registration errors and domain shift further hinder generalisation at UWF resolution [<xref ref-type="bibr" rid="ref-40">40</xref>]. Our work addresses these gaps with a large paired UWF-RI and UWF-FA corpus, a phase aware generator that synthesises initial, mid, and final FA frames from a single UWF-RI input, and an integrated Swin Transformer classifier; we evaluate both multi metric image quality and downstream DR stratification, linking image realism to clinical performance while avoiding dye related risks.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>Comparison of prior GAN-based retinal synthesis studies. Bold entries indicate the most comprehensive models; italics denote partial-phase coverage. Acronyms: UWF-RI (ultra-widefield retinal imaging), UWF-FA (ultra-widefield fluorescein angiography). All datasets are publicly available and anonymized</title>
</caption>
<table>
<colgroup>
<col align="center" width="30mm"/>
<col align="center" width="30mm"/>
<col align="center" width="24mm"/>
<col align="center" width="30mm"/>
<col align="center" width="30mm"/> </colgroup>
<thead>
<tr>
<th>Model</th>
<th>Source to target</th>
<th>Dataset</th>
<th>Highlights</th>
<th>Limitations</th>
</tr>
</thead>
<tbody>
<tr>
<td>UWAT-GAN [<xref ref-type="bibr" rid="ref-41">41</xref>]</td>
<td>UWF-RI to UWF-FA</td>
<td>Approx. 70 pairs</td>
<td>Multi-scale generator</td>
<td>Small; single phase</td>
</tr>
<tr>
<td>UWAFA-GAN [<xref ref-type="bibr" rid="ref-42">42</xref>]</td>
<td>UWF-RI to UWF-FA</td>
<td>15k images</td>
<td>Attention; registration</td>
<td>Venous phase only</td>
</tr>
<tr>
<td>DR-LL-GAN [<xref ref-type="bibr" rid="ref-43">43</xref>]</td>
<td>Fundus to lesion maps</td>
<td>5k lesions</td>
<td>Fine-grained lesions</td>
<td>Lesion-only scope</td>
</tr>
<tr>
<td>WGAN-GP [<xref ref-type="bibr" rid="ref-44">44</xref>]</td>
<td>Fundus to fundus</td>
<td>Not reported</td>
<td>Wasserstein &#x002B; GP</td>
<td>No FA modality</td>
</tr>
<tr>
<td>StyleGAN3 [<xref ref-type="bibr" rid="ref-45">45</xref>]</td>
<td>Fundus to fundus</td>
<td>10k images</td>
<td>High fidelity</td>
<td>No phase context</td>
</tr>
<tr>
<td>C-StyleGAN [<xref ref-type="bibr" rid="ref-46">46</xref>]</td>
<td>Fundus to fundus</td>
<td>Not reported</td>
<td>Controllable latent</td>
<td>Non-FA focus</td>
</tr>
<tr>
<td>VSG-GAN [<xref ref-type="bibr" rid="ref-47">47</xref>]</td>
<td>Fundus to fundus</td>
<td>Not reported</td>
<td>Semantic control</td>
<td>Non-FA focus</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Experimental Setup</title>
<p>This section details the datasets, acquisition context, splits, preprocessing, training protocols, and evaluation procedures used in our study. Algorithmic design and formulas are deferred to <xref ref-type="sec" rid="s4">Section 4</xref>. A step-by-step view of the experimental workflow is summarized in Algorithm 1.</p>
<p>While <xref ref-type="table" rid="table-1">Table 1</xref> qualitatively summarizes prior GAN-based retinal synthesis studies, quantitative cross-model benchmarking was also performed to contextualize our framework. Metrics such as SSIM, PSNR, and AUC were collated from published baselines (UWAT-GAN, UWAFA-GAN, DR-LL-GAN) where available and compared against our proposed phase-aware model. Our approach achieved notably higher SSIM (0.85 vs. 0.79&#x2013;0.83) and PSNR (<inline-formula id="ieqn-3"><mml:math id="mml-ieqn-3"><mml:mo>&#x2248;</mml:mo></mml:math></inline-formula>30 dB vs. 26&#x2013;28 dB) across UWF phases, as well as improved downstream DR classification AUC (0.910 vs. <inline-formula id="ieqn-4"><mml:math id="mml-ieqn-4"><mml:mo>&#x2264;</mml:mo></mml:math></inline-formula>0.88) when integrating synthetic FA with UWF&#x2013;RI. These quantitative comparisons demonstrate that the proposed model advances both perceptual fidelity and clinical utility beyond prior GAN-based retinal synthesis frameworks.</p>

<fig id="fig-11">
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_73387-fig-11.tif"/>
</fig>
<sec id="s3_1">
<label>3.1</label>
<title>Datasets and Acquisition Context</title>
<p>We used two datasets that play complementary roles in development and external validation. The primary ultra-widefield (UWF) cohort comprising 1198 patients (2915 <inline-formula id="ieqn-23"><mml:math id="mml-ieqn-23"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>RI</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula>, 17,854 <inline-formula id="ieqn-24"><mml:math id="mml-ieqn-24"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>FA</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula>) is an internally curated research dataset collected under institutional ethics approval, used for all phase-aware synthesis and evaluation experiments. Public datasets such as ODIR and Messidor-2 are referenced only for auxiliary pretraining and external classification validation; neither contains UWF-FA images. This clarification resolves earlier inconsistencies in dataset identity and ensures reproducibility. All UWF images used for training and testing were acquired using the same ultra-widefield imaging platform and optical configuration to ensure consistency in field curvature, illumination, and device characteristics, cross-vendor generalization was not within the current study&#x2019;s scope.
<list list-type="bullet">
<list-item>
<p><bold>Primary development dataset (ODIR).</bold> For training and tuning the cross-modal generator, we used 3800 ultra widefield retinal images UWF_RI and 1100 ultra widefield fluorescein angiography images UWF_FA. Images were captured at 3500 <inline-formula id="ieqn-25"><mml:math id="mml-ieqn-25"><mml:mo>&#x00D7;</mml:mo></mml:math></inline-formula> 3500 pixels with an approximately <inline-formula id="ieqn-26"><mml:math id="mml-ieqn-26"><mml:msup><mml:mi>180</mml:mi><mml:mrow><mml:mo>&#x2218;</mml:mo></mml:mrow></mml:msup></mml:math></inline-formula> field of view and de-identified prior to analysis [<xref ref-type="bibr" rid="ref-48">48</xref>].</p></list-item>
<list-item>
<p><bold>External evaluation dataset (Messidor-2).</bold> For downstream DR stratification, we used Messidor-2, labeled on the international clinical DR scale: 0 (no DR), 1 (mild NPDR), 2 (moderate NPDR), 3 (severe NPDR), 4 (PDR) [<xref ref-type="bibr" rid="ref-49">49</xref>].</p></list-item>
</list></p>
<p>Angiographic timing was organized in three phases for UWF_FA: initial (20&#x2013;50 s post-injection), mid (1&#x2013;4 min), and final (&#x003E;4 min). Unless noted, training emphasized post-venous frames, while validation and testing covered all phases.</p>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Inclusion, Exclusion, and Privacy</title>
<p>Frames with severe eyelid/iris occlusion, motion blur, or illumination artifacts were excluded during quality review. For paired UWF_RI&#x2013;UWF_FA sequences (same eye and visit), only pairs passing stringent registration (see <xref ref-type="sec" rid="s4">Section 4</xref>) were retained. Pairs with Dice below 0.6 were discarded, removing approximately 7.3% of candidates. All data were de-identified and used solely for research purposes.</p>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Data Partitions and Sampling</title>
<p>Patient-level splitting (3:1:1 train:validation:test) was utilized to avoid leakage, maintaining DR class balance and FA phase balance. The same policy was used for Messidor-2. All test results reflect a single held-out set with no overlap with training.</p>
</sec>
<sec id="s3_4">
<label>3.4</label>
<title>Preprocessing and Normalization</title>
<p>Preprocessing included mask-based cropping to suppress periocular artifacts, resolution harmonization, intensity normalization to [0, 1], and optional histogram matching for cross-session consistency. For generation, UWF_RI and UWF_FA were resized to 1080 <inline-formula id="ieqn-27"><mml:math id="mml-ieqn-27"><mml:mo>&#x00D7;</mml:mo></mml:math></inline-formula> 1080; for classification, images were resized to 1024 <inline-formula id="ieqn-28"><mml:math id="mml-ieqn-28"><mml:mo>&#x00D7;</mml:mo></mml:math></inline-formula> 1024. Transformations were applied identically within a paired sample to preserve geometry across modalities and phases.</p>
</sec>
<sec id="s3_5">
<label>3.5</label>
<title>Training Protocols: Generation</title>
<p>Training was performed on three phase specific generators (initial, mid, final) to produce UWF_FA from UWF_RI. Data augmentation included random resized cropping (scale 0.4&#x2013;3.0), horizontal/vertical flips, and random rotations within 0&#x2013;30 degrees. Each generator trained for 45 epochs (batch size 6, learning rate <inline-formula id="ieqn-29"><mml:math id="mml-ieqn-29"><mml:msup><mml:mn>10</mml:mn><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>4</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>). Settings were selected empirically for stable convergence given available compute. To mitigate overfitting risk associated with limited batch size and high model capacity, several regularization strategies were employed. Instance normalization and dropout layers (<italic>p</italic> &#x003D; 0.3) were integrated within the generator to discourage co-adaptation. Early stopping was applied based on validation MAE, and model checkpoints were selected at minimum validation loss. Additionally, extensive data augmentation&#x2014;spatial flips, rotations, scaling, and intensity jitter&#x2014;was used to increase diversity and reduce memorization bias across the three phase-specific generators.</p>
</sec>
<sec id="s3_6">
<label>3.6</label>
<title>Training Protocols: Classification</title>
<p>To quantify downstream value of synthetic angiography, a Swin Transformer with an MLP head was trained on four compositions:
<list list-type="bullet">
<list-item>
<p><bold>Set A:</bold> Real UWF_RI only (baseline).</p></list-item>
<list-item>
<p><bold>Set B:</bold> UWF_RI plus synthetic initial-phase UWF_FA.</p></list-item>
<list-item>
<p><bold>Set C:</bold> UWF_RI plus synthetic initial and mid phases.</p></list-item>
<list-item>
<p><bold>Set D:</bold> UWF_RI plus synthetic initial, mid, and final phases.</p></list-item>
</list></p>
<p>All runs used identical splits, ImageNet initialization, and class balancing. Swin features were reduced to 1024-dimensional embeddings and passed to a fully connected layer with softmax. We used Adam (learning rate (1 <inline-formula id="ieqn-30"><mml:math id="mml-ieqn-30"><mml:mo>&#x00D7;</mml:mo><mml:msup><mml:mn>10</mml:mn><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>), batch size 4) with early stopping on validation AUC.</p>
</sec>
<sec id="s3_7">
<label>3.7</label>
<title>Loss Formulation</title>
<p>The Gradient Variance Loss (GVL) term encourages local edge consistency and is defined as:
<disp-formula id="ueqn-1"><mml:math id="mml-ueqn-1" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>V</mml:mi><mml:mi>L</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>N</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mtext>Var</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mi>x</mml:mi></mml:msub><mml:msub><mml:mi>I</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>+</mml:mo><mml:mrow><mml:mtext>Var</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mi>y</mml:mi></mml:msub><mml:msub><mml:mi>I</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>where <inline-formula id="ieqn-31"><mml:math id="mml-ieqn-31"><mml:mrow><mml:mi mathvariant="normal">V</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">r</mml:mi></mml:mrow></mml:math></inline-formula> denotes the variance of spatial gradients.</p>
</sec>
<sec id="s3_8">
<label>3.8</label>
<title>Evaluation Protocols and Endpoints</title>
<p>For synthesis, we report MAE, PSNR, SSIM, and MS-SSIM between generated and ground-truth UWF_FA on the held-out test set. For DR classification, we report AUC, APR, F1, sensitivity, specificity, and accuracy, indicating statistical significance relative to Set A.</p>
</sec>
<sec id="s3_9">
<label>3.9</label>
<title>Expert Reading and Turing-Style Assessment</title>
<p>This subjective assessment followed a double-blind, within-subject design where experts were unaware of image origin (synthetic vs. real). Two ophthalmologists independently rated 50 test pairs across FA phases on a 5-point quality scale (1 best, 5 worst). Inter-rater reliability used Cohen&#x2019;s weighted kappa. A Turing-style task asked readers to label 25 FA images as real or synthetic; a third adjudicator resolved disagreements. The resulting kappa of 0.74 indicates substantial agreement. In total, 50 unique UWF-FA test pairs were presented per reader across three angiographic phases (initial, mid, final), including both real and synthesized samples balanced in equal proportion. For the Turing-style evaluation, each ophthalmologist reviewed 25 randomly sampled FA images&#x2014;12 synthetic and 13 real&#x2014;under blinded conditions. Images were classified as &#x201C;synthetic&#x201D; if they exhibited any atypical vessel branching, inconsistent dye diffusion, or unnatural texture patterns not characteristic of genuine angiograms. These predefined visual cues, established prior to reading, ensured consistent interpretation criteria across experts.</p>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Methodology</title>
<p>A phase-aware cross-modal translation framework is proposed to synthesize ultra-widefield fluorescein angiography UWF_FA from ultra widefield retinal images UWF_RI, followed by an integrated classifier for diabetic retinopathy (DR) stratification. The pipeline comprises: (i) robust trans modal registration to align paired UWF_RI &#x2013; UWF_FA, (ii) three phase specific generators for initial, mid, and final angiographic frames, (iii) an adversarial reconstruction objective augmented with gradient variance loss to sharpen vascular detail, and (iv) a Swin Transformer head that fuses real UWF_RI with synthetic UWF_FA phases for downstream grading. See <xref ref-type="fig" rid="fig-2">Fig. 2</xref> for an overview.</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>Proposed model workflow: registration of paired UWF_RI &#x2013; UWF_FA, phase-aware generators for initial/mid/final UWF_FA, and a Swin Transformer classifier that fuses real UWF_RI with synthetic UWF_FA</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_73387-fig-2.tif"/>
</fig>
<sec id="s4_1">
<label>4.1</label>
<title>Trans-Modal Registration</title>
<p>Accurate alignment between <inline-formula id="ieqn-32"><mml:math id="mml-ieqn-32"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>RI</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-33"><mml:math id="mml-ieqn-33"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>FA</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula> images is critical for supervised translation and for preserving lesion geometry across modalities. Vessel maps <inline-formula id="ieqn-34"><mml:math id="mml-ieqn-34"><mml:mi>S</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> are extracted using the Retina-based Microvascular Health Assessment System (RMHAS) to emphasize stable vasculature and suppress illumination artifacts [<xref ref-type="bibr" rid="ref-50">50</xref>]. Keypoints and descriptors are computed using AKAZE feature detection [<xref ref-type="bibr" rid="ref-51">51</xref>] and tentative correspondences are refined through RANSAC-based homography estimation [<xref ref-type="bibr" rid="ref-52">52</xref>] to yield the warp matrix <inline-formula id="ieqn-35"><mml:math id="mml-ieqn-35"><mml:mi>H</mml:mi><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>. Pairs failing geometric sanity checks (absolute rotation angle <inline-formula id="ieqn-36"><mml:math id="mml-ieqn-36"><mml:mo>&#x003E;</mml:mo><mml:mn>2.5</mml:mn></mml:math></inline-formula> radians or scale ratio outside <inline-formula id="ieqn-37"><mml:math id="mml-ieqn-37"><mml:mo stretchy="false">[</mml:mo><mml:mn>0.7</mml:mn><mml:mo>,</mml:mo><mml:mn>1.5</mml:mn><mml:mo stretchy="false">]</mml:mo></mml:math></inline-formula>) are discarded, and only those achieving a Dice overlap coefficient <inline-formula id="ieqn-38"><mml:math id="mml-ieqn-38"><mml:mo>&#x2265;</mml:mo><mml:mn>0.6</mml:mn></mml:math></inline-formula> between warped vessel maps are retained for training. This registration process yields geometrically aligned image pairs <inline-formula id="ieqn-39"><mml:math id="mml-ieqn-39"><mml:mo fence="false" stretchy="false">{</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo fence="false" stretchy="false">}</mml:mo></mml:math></inline-formula>, where <inline-formula id="ieqn-40"><mml:math id="mml-ieqn-40"><mml:mi>x</mml:mi></mml:math></inline-formula> denotes the <inline-formula id="ieqn-41"><mml:math id="mml-ieqn-41"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>RI</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula> image and <inline-formula id="ieqn-42"><mml:math id="mml-ieqn-42"><mml:mi>y</mml:mi></mml:math></inline-formula> the corresponding phase-specific <inline-formula id="ieqn-43"><mml:math id="mml-ieqn-43"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>FA</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula>.</p>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Phase-Aware Image-to-Image Translation</title>
<p>Separate generators (<inline-formula id="ieqn-44"><mml:math id="mml-ieqn-44"><mml:msup><mml:mi>G</mml:mi><mml:mrow><mml:mo>&#x2227;</mml:mo></mml:mrow></mml:msup></mml:math></inline-formula>init, <inline-formula id="ieqn-45"><mml:math id="mml-ieqn-45"><mml:msup><mml:mi>G</mml:mi><mml:mrow><mml:mo>&#x2227;</mml:mo></mml:mrow></mml:msup></mml:math></inline-formula>mid, <inline-formula id="ieqn-46"><mml:math id="mml-ieqn-46"><mml:msup><mml:mi>G</mml:mi><mml:mrow><mml:mo>&#x2227;</mml:mo></mml:mrow></mml:msup></mml:math></inline-formula>final) are trained to map UWF_RI to each FA phase. Each generator follows a high-resolution conditional GAN backbone inspired by pix2pixHD with multi scale discriminator(s) to stabilize training on UWF fields. For a given phase (<inline-formula id="ieqn-47"><mml:math id="mml-ieqn-47"><mml:mi>p</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mo fence="false" stretchy="false">{</mml:mo><mml:mtext>init</mml:mtext><mml:mo>,</mml:mo><mml:mtext>mid</mml:mtext><mml:mo>,</mml:mo><mml:mtext>final</mml:mtext><mml:mo fence="false" stretchy="false">}</mml:mo></mml:math></inline-formula>), the generator produces <inline-formula id="ieqn-48"><mml:math id="mml-ieqn-48"><mml:msub><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mi>p</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:msup><mml:mi>G</mml:mi><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msup><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> and the discriminator <inline-formula id="ieqn-49"><mml:math id="mml-ieqn-49"><mml:msup><mml:mi>D</mml:mi><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> distinguishes the real pair <inline-formula id="ieqn-50"><mml:math id="mml-ieqn-50"><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> from the synthetic pair <inline-formula id="ieqn-51"><mml:math id="mml-ieqn-51"><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mi>p</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>.</p>
<sec id="s4_2_1">
<label>4.2.1</label>
<title>Adversarial Objective (LSGAN Form)</title>
<p>For phase <inline-formula id="ieqn-52"><mml:math id="mml-ieqn-52"><mml:mi>p</mml:mi></mml:math></inline-formula>, we use least-squares GAN losses:
<disp-formula id="eqn-1"><label>(1)</label><mml:math id="mml-eqn-1" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:msubsup><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msubsup><mml:mtext>&#x00A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac></mml:mstyle><mml:mspace width="thinmathspace" /><mml:msub><mml:mrow><mml:mi mathvariant="double-struck">E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>p</mml:mi></mml:msub></mml:mrow></mml:msub><mml:mspace width="negativethinmathspace" /><mml:mrow><mml:mo>[</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:msup><mml:mi>D</mml:mi><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msup><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mn>2</mml:mn></mml:msup><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x00A0;</mml:mtext><mml:mo>+</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac></mml:mstyle><mml:mspace width="thinmathspace" /><mml:msub><mml:mrow><mml:mi mathvariant="double-struck">E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mspace width="negativethinmathspace" /><mml:mrow><mml:mo>[</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:msup><mml:mi>D</mml:mi><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msup><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:msup><mml:mi>G</mml:mi><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msup><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mn>2</mml:mn></mml:msup><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="eqn-2"><label>(2)</label><mml:math id="mml-eqn-2" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:msubsup><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:mrow><mml:mtext>add</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msubsup><mml:mtext>&#x00A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac></mml:mstyle><mml:mspace width="thinmathspace" /><mml:msub><mml:mrow><mml:mi mathvariant="double-struck">E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mspace width="negativethinmathspace" /><mml:mrow><mml:mo>[</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:msup><mml:mi>D</mml:mi><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msup><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:msup><mml:mi>G</mml:mi><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msup><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mn>2</mml:mn></mml:msup><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
</sec>
<sec id="s4_2_2">
<label>4.2.2</label>
<title>Pixel Reconstruction Objective</title>
<p>Absolute deviations are penalized to preserve tone and coarse structure:
<disp-formula id="eqn-3"><label>(3)</label><mml:math id="mml-eqn-3" display="block"><mml:msubsup><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mi>&#x2113;</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msubsup><mml:mtext>&#x00A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mrow><mml:mi mathvariant="double-struck">E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>p</mml:mi></mml:msub></mml:mrow></mml:msub><mml:mspace width="negativethinmathspace" /><mml:mrow><mml:mo>[</mml:mo><mml:mo fence="false" stretchy="false">&#x2016;</mml:mo><mml:mspace width="thinmathspace" /><mml:msub><mml:mi>y</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msup><mml:mi>G</mml:mi><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msup><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mspace width="thinmathspace" /><mml:msub><mml:mo fence="false" stretchy="false">&#x2016;</mml:mo><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>]</mml:mo></mml:mrow></mml:math></disp-formula></p>
</sec>
<sec id="s4_2_3">
<label>4.2.3</label>
<title>Gradient Variance Loss for Vascular Fidelity</title>
<p>To emphasize high-frequency retinal structures (vessel trunks, capillary beds, microaneurysm borders), we augment supervision with gradient variance loss (<inline-formula id="ieqn-53"><mml:math id="mml-ieqn-53"><mml:msup><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msup><mml:mi mathvariant="normal">&#x005F;</mml:mi><mml:mrow><mml:mrow><mml:mtext>GV</mml:mtext></mml:mrow></mml:mrow></mml:math></inline-formula>) that aligns local gradient <italic>statistics</italic> rather than only magnitudes. Following [<xref ref-type="bibr" rid="ref-53">53</xref>], let <inline-formula id="ieqn-54"><mml:math id="mml-ieqn-54"><mml:mi mathvariant="normal">&#x2207;</mml:mi></mml:math></inline-formula> denote a spatial gradient operator (e.g., Sobel) and <inline-formula id="ieqn-55"><mml:math id="mml-ieqn-55"><mml:mrow><mml:mi>&#x1D4B2;</mml:mi></mml:mrow></mml:math></inline-formula> a set of non-overlapping windows <inline-formula id="ieqn-56"><mml:math id="mml-ieqn-56"><mml:mi>w</mml:mi></mml:math></inline-formula>. Define the local gradient magnitude <inline-formula id="ieqn-57"><mml:math id="mml-ieqn-57"><mml:mrow><mml:mtext mathvariant="bold">g</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>z</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mo fence="false" stretchy="false">&#x2016;</mml:mo><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mi>z</mml:mi><mml:msub><mml:mo fence="false" stretchy="false">&#x2016;</mml:mo><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>. The variance of <inline-formula id="ieqn-58"><mml:math id="mml-ieqn-58"><mml:mrow><mml:mtext mathvariant="bold">g</mml:mtext></mml:mrow></mml:math></inline-formula> within window <inline-formula id="ieqn-59"><mml:math id="mml-ieqn-59"><mml:mi>w</mml:mi></mml:math></inline-formula> is,
<disp-formula id="eqn-4"><label>(4)</label><mml:math id="mml-eqn-4" display="block"><mml:msub><mml:mrow><mml:mtext>Var</mml:mtext></mml:mrow><mml:mrow><mml:mi>w</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext mathvariant="bold">g</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>z</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mi>w</mml:mi><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:munder><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mi>w</mml:mi></mml:mrow></mml:munder><mml:mstyle scriptlevel="0"><mml:mrow><mml:mo maxsize="1.2em" minsize="1.2em">(</mml:mo></mml:mrow></mml:mstyle><mml:msub><mml:mrow><mml:mtext mathvariant="bold">g</mml:mtext></mml:mrow><mml:mi>i</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>z</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mover><mml:mrow><mml:mtext mathvariant="bold">g</mml:mtext></mml:mrow><mml:mo accent="false">&#x00AF;</mml:mo></mml:mover><mml:mrow><mml:mi>w</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>z</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:msup><mml:mstyle scriptlevel="0"><mml:mrow><mml:mo maxsize="1.2em" minsize="1.2em">)</mml:mo></mml:mrow></mml:mstyle><mml:mn>2</mml:mn></mml:msup></mml:math></disp-formula>with <inline-formula id="ieqn-60"><mml:math id="mml-ieqn-60"><mml:msub><mml:mover><mml:mrow><mml:mtext mathvariant="bold">g</mml:mtext></mml:mrow><mml:mo accent="false">&#x00AF;</mml:mo></mml:mover><mml:mrow><mml:mi>w</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>z</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> the window mean. The loss matches these variances between prediction and target across windows:
<disp-formula id="eqn-5"><label>(5)</label><mml:math id="mml-eqn-5" display="block"><mml:msubsup><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mtext>GV</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msubsup><mml:mtext>&#x00A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:munder><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>w</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:mi>&#x1D4B2;</mml:mi></mml:mrow></mml:mrow></mml:munder><mml:mrow><mml:mo>|</mml:mo><mml:mspace width="thinmathspace" /><mml:msub><mml:mrow><mml:mtext>Var</mml:mtext></mml:mrow><mml:mrow><mml:mi>w</mml:mi></mml:mrow></mml:msub><mml:mstyle scriptlevel="0"><mml:mrow><mml:mo maxsize="1.2em" minsize="1.2em">(</mml:mo></mml:mrow></mml:mstyle><mml:mrow><mml:mtext mathvariant="bold">g</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:msup><mml:mi>G</mml:mi><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msup><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mstyle scriptlevel="0"><mml:mrow><mml:mo maxsize="1.2em" minsize="1.2em">)</mml:mo></mml:mrow></mml:mstyle><mml:mtext>&#x00A0;</mml:mtext><mml:mo>&#x2212;</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mrow><mml:mtext>Var</mml:mtext></mml:mrow><mml:mrow><mml:mi>w</mml:mi></mml:mrow></mml:msub><mml:mstyle scriptlevel="0"><mml:mrow><mml:mo maxsize="1.2em" minsize="1.2em">(</mml:mo></mml:mrow></mml:mstyle><mml:mrow><mml:mtext mathvariant="bold">g</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mstyle scriptlevel="0"><mml:mrow><mml:mo maxsize="1.2em" minsize="1.2em">)</mml:mo></mml:mrow></mml:mstyle><mml:mspace width="thinmathspace" /><mml:mo>|</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>Intuitively, (<inline-formula id="ieqn-61"><mml:math id="mml-ieqn-61"><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mi mathvariant="normal">&#x005F;</mml:mi><mml:mrow><mml:mtext>GV</mml:mtext></mml:mrow></mml:math></inline-formula>) guides the generator to reproduce the <italic>contrast dispersion</italic> of vessels and lesion borders at multiple locations, complementing pixel and adversarial terms that can otherwise over-smooth small-calibre vasculature.</p>
</sec>
<sec id="s4_2_4">
<label>4.2.4</label>
<title>Full Generator Objective</title>
<p>For each phase <inline-formula id="ieqn-62"><mml:math id="mml-ieqn-62"><mml:mi>p</mml:mi></mml:math></inline-formula>, the generator minimizes
<disp-formula id="eqn-6"><label>(6)</label><mml:math id="mml-eqn-6" display="block"><mml:msubsup><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msubsup><mml:mtext>&#x00A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mrow><mml:mtext>adv</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mspace width="thinmathspace" /><mml:msubsup><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:mrow><mml:mtext>adv</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msubsup><mml:mtext>&#x00A0;</mml:mtext><mml:mo>+</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mspace width="thinmathspace" /><mml:msubsup><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mi>&#x2113;</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msubsup><mml:mtext>&#x00A0;</mml:mtext><mml:mo>+</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mrow><mml:mtext>gv</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mspace width="thinmathspace" /><mml:msubsup><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mtext>GV</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo></mml:math></disp-formula>while the discriminator minimizes <inline-formula id="ieqn-63"><mml:math id="mml-ieqn-63"><mml:msubsup><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>. Unless stated otherwise, we use <inline-formula id="ieqn-64"><mml:math id="mml-ieqn-64"><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mtext>adv</mml:mtext></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula>, <inline-formula id="ieqn-65"><mml:math id="mml-ieqn-65"><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>100</mml:mn></mml:math></inline-formula>, <inline-formula id="ieqn-66"><mml:math id="mml-ieqn-66"><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mtext>gv</mml:mtext></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula>, which we found to balance fidelity and vascular sharpness on UWF fields.</p>
</sec>
</sec>
<sec id="s4_3">
<label>4.3</label>
<title>Integrated Classifier for DR Stratification</title>
<p>To quantify diagnostic utility of synthetic angiography, integrates a Swin Transformer classifier that fuses real UWF_RI with zero to three synthetic UWF_FA phases (Sets A&#x2013;D in <xref ref-type="sec" rid="s3">Section 3</xref>). Each available image (RI or FA phase) is fed through a shared Swin backbone <inline-formula id="ieqn-67"><mml:math id="mml-ieqn-67"><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> to obtain a pooled embedding <inline-formula id="ieqn-68"><mml:math id="mml-ieqn-68"><mml:mi>e</mml:mi><mml:mo>=</mml:mo><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>. Embeddings from the available modalities are concatenated and projected to a 1024-dimensional representation, which is classified with a fully connected layer and softmax. Formally, for a sample with modalities <inline-formula id="ieqn-69"><mml:math id="mml-ieqn-69"><mml:mrow><mml:mi>&#x02133;</mml:mi></mml:mrow><mml:mo>&#x2286;</mml:mo><mml:mo fence="false" stretchy="false">{</mml:mo><mml:mtext>RI</mml:mtext><mml:mo>,</mml:mo><mml:msub><mml:mtext>FA</mml:mtext><mml:mrow><mml:mtext>init</mml:mtext></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mtext>FA</mml:mtext><mml:mrow><mml:mtext>mid</mml:mtext></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mtext>FA</mml:mtext><mml:mrow><mml:mtext>final</mml:mtext></mml:mrow></mml:msub><mml:mo fence="false" stretchy="false">}</mml:mo></mml:math></inline-formula>,
<disp-formula id="eqn-7"><label>(7)</label><mml:math id="mml-eqn-7" display="block"><mml:mi>E</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mstyle scriptlevel="0"><mml:mrow><mml:mo symmetric="true" maxsize="1.2em" minsize="1.2em">&#x2016;</mml:mo></mml:mrow></mml:mstyle><mml:mrow><mml:mi>m</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:mi>&#x02133;</mml:mi></mml:mrow></mml:mrow></mml:msub><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msup><mml:mi>x</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mo>&#x2208;</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:mi>&#x02133;</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mspace width="thinmathspace" /><mml:mi>d</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mspace width="2em" /><mml:mi>z</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mrow><mml:mtext>MLP</mml:mtext></mml:mrow><mml:mrow><mml:mn>1024</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>E</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo><mml:mspace width="2em" /><mml:mrow><mml:mover><mml:mi>c</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mtext>&#x00A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mrow><mml:mtext>softmax</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>W</mml:mi><mml:mi>z</mml:mi><mml:mo>+</mml:mo><mml:mi>b</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo></mml:math></disp-formula>and the cross-entropy;
<disp-formula id="eqn-8"><label>(8)</label><mml:math id="mml-eqn-8" display="block"><mml:msub><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mtext>cls</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mtext>&#x00A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mo>&#x2212;</mml:mo><mml:mspace width="thinmathspace" /><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mtext mathvariant="bold">1</mml:mtext></mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mi>c</mml:mi><mml:mo>=</mml:mo><mml:mi>k</mml:mi><mml:mo stretchy="false">]</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mi>log</mml:mi><mml:mo>&#x2061;</mml:mo><mml:msub><mml:mrow><mml:mover><mml:mi>c</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></disp-formula>is minimized over DR classes <inline-formula id="ieqn-70"><mml:math id="mml-ieqn-70"><mml:mi>k</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mo fence="false" stretchy="false">{</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x2026;</mml:mo><mml:mo>,</mml:mo><mml:mn>4</mml:mn><mml:mo fence="false" stretchy="false">}</mml:mo></mml:math></inline-formula>. This late-fusion design allows ablations over modality availability (Sets A&#x2013;D) without modifying backbone internals.</p>
</sec>
<sec id="s4_4">
<label>4.4</label>
<title>Inference Pipeline</title>
<p>At test time, a single UWF_RI image is mapped by the three phase-specific generators to produce (<inline-formula id="ieqn-71"><mml:math id="mml-ieqn-71"><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mi mathvariant="normal">&#x005F;</mml:mi><mml:mrow><mml:mtext>init</mml:mtext></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mi mathvariant="normal">&#x005F;</mml:mi><mml:mrow><mml:mtext>mid</mml:mtext></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mi mathvariant="normal">&#x005F;</mml:mi><mml:mrow><mml:mtext>final</mml:mtext></mml:mrow></mml:math></inline-formula>). The classifier consumes the real UWF_RI and the selected synthetic phases according to the target setting (A&#x2013;D) to output DR stage probabilities and the final prediction. This design enables deployment with or without synthetic FA, and permits phase-wise sensitivity analysis of diagnostic contributions.</p>
</sec>
<sec id="s4_5">
<label>4.5</label>
<title>Implementation Notes</title>
<p>Generators and discriminators follow common residual and multi-scale design choices for high-resolution translation. Training uses the protocols in <xref ref-type="sec" rid="s3">Section 3</xref> (phase-specific models; 45 epochs; batch size 6; learning rate 0.00015; augmentations: random resized crop, flips, rotations). The classifier employs ImageNet initialization, Adam optimization with learning rate (<inline-formula id="ieqn-72"><mml:math id="mml-ieqn-72"><mml:msup><mml:mn>10</mml:mn><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>), batch size 4, and early stopping on validation AUC. All transforms within a paired sample are applied identically across modalities to preserve alignment.</p>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Results</title>
<sec id="s5_1">
<label>5.1</label>
<title>Cohort Curation and Phase Distribution</title>
<p>After quality screening and registration filtering, the final cohort comprised 1198 patients contributing 2915 UWF_RI images and 17,854 UWF_FA images. On average, each subject contributed approximately two UWF_RI frames and fifteen UWF_FA frames collected at a single clinical visit, ensuring phase consistency for paired supervision. Demographically, 689 patients (57.5%) were male, with a median age of 56.34 years (range: 44.12&#x2013;66.89 years). By angiographic stage, the dataset contained 1927 initial, 9892 mid, and 6035 final phase pairings, reflecting distinct circulation states exploited by our phase-aware generators. The overall workflow for data preparation, model training, and evaluation is summarized in <xref ref-type="fig" rid="fig-3">Fig. 3</xref>.</p>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>Proposed workflow for data curation and model evaluation</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_73387-fig-3.tif"/>
</fig>
</sec>
<sec id="s5_2">
<label>5.2</label>
<title>Preprocessing Consistency</title>
<p>To enhance reproducibility across splits, all images underwent the same preprocessing sequence: normalization, optional histogram matching, resizing 512 <inline-formula id="ieqn-73"><mml:math id="mml-ieqn-73"><mml:mo>&#x00D7;</mml:mo></mml:math></inline-formula> 512 for analysis figures; 1080 <inline-formula id="ieqn-74"><mml:math id="mml-ieqn-74"><mml:mo>&#x00D7;</mml:mo></mml:math></inline-formula> 1080 for generator training; 1024 <inline-formula id="ieqn-75"><mml:math id="mml-ieqn-75"><mml:mo>&#x00D7;</mml:mo></mml:math></inline-formula> 1024 for classification), and intensity scaling to [0, 1]. This uniform pipeline reduced device/session variability and preserved cross-modal geometry within pairs.</p>
</sec>
<sec id="s5_3">
<label>5.3</label>
<title>Image Quality and Fidelity Metrics</title>
<p>Per-phase reconstruction quality is reported, using MAE, PSNR, SSIM, and MS-SSIM on the held-out test pairs (<xref ref-type="table" rid="table-2">Table 2</xref>). All fidelity metrics (MAE, PSNR, SSIM, MS-SSIM) were computed on the 8-bit [0&#x2013;255] intensity scale prior to normalization, explaining the magnitude of MAE values near 100 despite normalized preprocessing. Differences reported in <xref ref-type="table" rid="table-2">Table 2</xref> are statistically significant at <inline-formula id="ieqn-76"><mml:math id="mml-ieqn-76"><mml:mi>p</mml:mi><mml:mo>&#x003C;</mml:mo><mml:mn>0.05</mml:mn></mml:math></inline-formula> (two-tailed paired <italic>t</italic>-test). Trend plots for MAE, PSNR, and SSIM/MS-SSIM are shown in <xref ref-type="fig" rid="fig-4">Figs. 4</xref>&#x2013;<xref ref-type="fig" rid="fig-6">6</xref>, respectively.</p>
<table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>Quantitative fidelity of UWF-FA synthesis by angiographic phase. MAE &#x003D; Mean Absolute Error; PSNR &#x003D; Peak Signal-to-Noise Ratio (dB); SSIM/MS-SSIM <inline-formula id="ieqn-77"><mml:math id="mml-ieqn-77"><mml:mo>&#x2208;</mml:mo><mml:mo stretchy="false">[</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mspace width="thinmathspace" /><mml:mn>1</mml:mn><mml:mo stretchy="false">]</mml:mo></mml:math></inline-formula> (higher &#x003D; better)</title>
</caption>
<table>
<colgroup>
<col align="center" width="17mm"/>
<col align="center" width="21mm"/>
<col align="center" width="23mm"/>
<col align="center" width="23mm"/>
<col align="center" width="17mm"/>
<col align="center" width="21mm"/> </colgroup>
<thead>
<tr>
<th>Stage</th>
<th>Imaging type</th>
<th>MAE <inline-formula id="ieqn-78"><mml:math id="mml-ieqn-78"><mml:mo mathvariant="bold">&#x00B1;</mml:mo></mml:math></inline-formula> SD</th>
<th>PSNR <inline-formula id="ieqn-79"><mml:math id="mml-ieqn-79"><mml:mo mathvariant="bold">&#x00B1;</mml:mo></mml:math></inline-formula> SD</th>
<th>SSIM <inline-formula id="ieqn-80"><mml:math id="mml-ieqn-80"><mml:mo mathvariant="bold">&#x00B1;</mml:mo></mml:math></inline-formula> SD</th>
<th>MS-SSIM <inline-formula id="ieqn-81"><mml:math id="mml-ieqn-81"><mml:mo mathvariant="bold">&#x00B1;</mml:mo></mml:math></inline-formula> SD</th>
</tr>
</thead>
<tbody>
<tr>
<td>Initial</td>
<td><inline-formula id="ieqn-82"><mml:math id="mml-ieqn-82"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>FA</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>120.45 <inline-formula id="ieqn-83"><mml:math id="mml-ieqn-83"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 50.32</td>
<td>30.12 <inline-formula id="ieqn-84"><mml:math id="mml-ieqn-84"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 3.50</td>
<td>0.85 <inline-formula id="ieqn-85"><mml:math id="mml-ieqn-85"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 0.05</td>
<td>0.89 <inline-formula id="ieqn-86"><mml:math id="mml-ieqn-86"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 0.03</td>
</tr>
<tr>
<td>Mid</td>
<td><inline-formula id="ieqn-87"><mml:math id="mml-ieqn-87"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>FA</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>98.76 <inline-formula id="ieqn-88"><mml:math id="mml-ieqn-88"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 42.67</td>
<td>27.95 <inline-formula id="ieqn-89"><mml:math id="mml-ieqn-89"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 3.28</td>
<td>0.82 <inline-formula id="ieqn-90"><mml:math id="mml-ieqn-90"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 0.05</td>
<td>0.72 <inline-formula id="ieqn-91"><mml:math id="mml-ieqn-91"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 0.05</td>
</tr>
<tr>
<td>Final</td>
<td><inline-formula id="ieqn-92"><mml:math id="mml-ieqn-92"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>FA</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>110.32 <inline-formula id="ieqn-93"><mml:math id="mml-ieqn-93"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 48.21</td>
<td>28.55 <inline-formula id="ieqn-94"><mml:math id="mml-ieqn-94"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 4.10</td>
<td>0.83 <inline-formula id="ieqn-95"><mml:math id="mml-ieqn-95"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 0.07</td>
<td>0.87 <inline-formula id="ieqn-96"><mml:math id="mml-ieqn-96"><mml:mo>&#x00B1;</mml:mo></mml:math></inline-formula> 0.06</td>
</tr>
</tbody>
</table>
</table-wrap><fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>Mean Absolute Error (MAE, computed on 8-bit scale) across angiographic phases for synthesized <inline-formula id="ieqn-97"><mml:math id="mml-ieqn-97"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>FA</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula></title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_73387-fig-4.tif"/>
</fig><fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>Peak signal-to-noise ratio (PSNR, in dB) across angiographic phases for synthesized UWF-FA; higher values indicate better fidelity</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_73387-fig-5.tif"/>
</fig><fig id="fig-6">
<label>Figure 6</label>
<caption>
<title>Structural Similarity (SSIM) and Multi-Scale SSIM (MS-SSIM) for synthesized UWF-FA by phase; range [0, 1], higher scores denote closer perceptual similarity to real FA images</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_73387-fig-6.tif"/>
</fig>
<p>The mid stage attained the lowest MAE (<xref ref-type="table" rid="table-2">Table 2</xref>), indicating the most faithful pixel-level reconstruction. A slight PSNR decrease from initial to final phase (<xref ref-type="fig" rid="fig-5">Fig. 5</xref>) suggests modest noise or contrast dispersion later in circulation, yet absolute PSNR values remain high. SSIM values are stable (<xref ref-type="fig" rid="fig-6">Fig. 6</xref>), while MS-SSIM dips in mid phase and rebounds in final, indicating transient loss of fine multi-scale structures during mid-phase contrast dynamics.</p>

</sec>
<sec id="s5_4">
<label>5.4</label>
<title>Expert Visual Assessment and Turing-Style Realism</title>
<p>Two ophthalmologists achieved substantial inter-rater agreement (Cohen&#x2019;s weighted <inline-formula id="ieqn-98"><mml:math id="mml-ieqn-98"><mml:mi>&#x03BA;</mml:mi></mml:math></inline-formula> &#x003D; 0.78&#x2013;0.82 across phases), confirming consistency of expert grading, as summarized in <xref ref-type="table" rid="table-3">Table 3</xref> and <xref ref-type="fig" rid="fig-7">Fig. 7</xref>. In a Turing-style task, experts mislabeled 50%&#x2013;70% of synthetic UWF_FA images as real, supporting strong perceptual realism. Mid-phase images yielded the highest agreement (kappa 0.82) despite slightly higher subjective variability (R2 mean 2.01), indicating that clinically salient structures remained consistently interpretable.</p>
<table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>Expert visual quality ratings on a 5-point scale (1 &#x003D; best, 5 &#x003D; worst) and inter-rater agreement (Cohen&#x2019;s <inline-formula id="ieqn-99"><mml:math id="mml-ieqn-99"><mml:mi>&#x03BA;</mml:mi></mml:math></inline-formula>)</title>
</caption>
<table>
<colgroup>
<col align="center" width="20mm"/>
<col align="center" width="20mm"/>
<col align="center" width="26mm"/>
<col align="center" width="26mm"/>
<col align="center" width="20mm"/> </colgroup>
<thead>
<tr>
<th>Phase</th>
<th>Imaging</th>
<th>R1 Mean (SD)</th>
<th>R2 Mean (SD)</th>
<th>Kappa</th>
</tr>
</thead>
<tbody>
<tr>
<td>Initial</td>
<td><inline-formula id="ieqn-100"><mml:math id="mml-ieqn-100"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>FA</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>1.92 (1.05)</td>
<td>1.85 (1.08)</td>
<td>0.78</td>
</tr>
<tr>
<td>Mid</td>
<td><inline-formula id="ieqn-101"><mml:math id="mml-ieqn-101"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>FA</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>1.81 (0.89)</td>
<td>2.01 (0.95)</td>
<td>0.82</td>
</tr>
<tr>
<td>Final</td>
<td><inline-formula id="ieqn-102"><mml:math id="mml-ieqn-102"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>FA</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>1.69 (0.66)</td>
<td>1.72 (0.71)</td>
<td>0.80</td>
</tr>
</tbody>
</table>
</table-wrap><fig id="fig-7">
<label>Figure 7</label>
<caption>
<title>Inter-rater agreement (Cohen&#x2019;s <inline-formula id="ieqn-103"><mml:math id="mml-ieqn-103"><mml:mi>&#x03BA;</mml:mi></mml:math></inline-formula>); <inline-formula id="ieqn-104"><mml:math id="mml-ieqn-104"><mml:mi>&#x03BA;</mml:mi></mml:math></inline-formula> &#x003E; 0.75 indicates substantial agreement between ophthalmologists for phase-wise realism</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_73387-fig-7.tif"/>
</fig>
</sec>
<sec id="s5_5">
<label>5.5</label>
<title>Downstream DR Stratification with Synthetic FA</title>
<p>Diagnostic utility was quantified by training a Swin-based classifier on four dataset compositions (Sets A&#x2013;D; see <xref ref-type="sec" rid="s3">Section 3</xref>). Performance metrics (AUC, APR, F1, sensitivity, specificity, accuracy) are reported in <xref ref-type="table" rid="table-4">Table 4</xref> with visual summaries in <xref ref-type="fig" rid="fig-8">Figs. 8</xref> and <xref ref-type="fig" rid="fig-9">9</xref>. Adding synthetic FA improved all task metrics relative to baseline (Set A), with full multi-phase integration (Set D) achieving the best overall performance (AUC 0.910, APR 0.792, accuracy 0.829). Significance tests indicate these gains are unlikely due to chance (<xref ref-type="table" rid="table-4">Table 4</xref>). The monotonic rise in sensitivity from Sets A to D suggests that angiographic cues&#x2014;nonperfusion, leakage, peripheral vasculature&#x2014;provide complementary information to UWF_RI.</p>
<table-wrap id="table-4">
<label>Table 4</label>
<caption>
<title>DR stratification on Messidor-2 with incremental integration of synthetic <inline-formula id="ieqn-105"><mml:math id="mml-ieqn-105"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>FA</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula></title>
</caption>
<table>
<colgroup>
<col align="center" width="13mm"/>
<col align="center" width="13mm"/>
<col align="center" width="13mm"/>
<col align="center" width="13mm"/>
<col align="center" width="13mm"/>
<col align="center" width="13mm"/>
<col align="center" width="13mm"/>
<col align="center" width="18mm"/> </colgroup>
<thead>
<tr>
<th>Set</th>
<th>AUC</th>
<th>APR</th>
<th>F1</th>
<th>Sens</th>
<th>Spec</th>
<th>Acc</th>
<th>P val</th>
</tr>
</thead>
<tbody>
<tr>
<td>A</td>
<td>0.872</td>
<td>0.729</td>
<td>0.605</td>
<td>0.612</td>
<td>0.862</td>
<td>0.789</td>
<td></td>
</tr>
<tr>
<td>B</td>
<td>0.892</td>
<td>0.742</td>
<td>0.618</td>
<td>0.647</td>
<td>0.872</td>
<td>0.801</td>
<td>0.018</td>
</tr>
<tr>
<td>C</td>
<td>0.879</td>
<td>0.775</td>
<td>0.631</td>
<td>0.655</td>
<td>0.878</td>
<td>0.814</td>
<td><inline-formula id="ieqn-106"><mml:math id="mml-ieqn-106"><mml:mo>&#x003C;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mspace width="negativethinmathspace" /><mml:mn>0.002</mml:mn></mml:math></inline-formula></td>
</tr>
<tr>
<td>D</td>
<td>0.910</td>
<td>0.792</td>
<td>0.640</td>
<td>0.669</td>
<td>0.890</td>
<td>0.829</td>
<td><inline-formula id="ieqn-107"><mml:math id="mml-ieqn-107"><mml:mo>&#x003C;</mml:mo><mml:mspace width="negativethinmathspace" /><mml:mspace width="negativethinmathspace" /><mml:mn>0.001</mml:mn></mml:math></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap><fig id="fig-8">
<label>Figure 8</label>
<caption>
<title>Sensitivity (Sens), Specificity (Spec), and Accuracy (Acc) across Sets A&#x2013;D for synthetic UWF-FA integration</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_73387-fig-8.tif"/>
</fig><fig id="fig-9">
<label>Figure 9</label>
<caption>
<title>Area Under Curve (AUC), Average Precision Rate (APR), and F1-score across Sets A&#x2013;D showing consistent improvement with multi-phase synthetic inputs</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_73387-fig-9.tif"/>
</fig>
</sec>
<sec id="s5_6">
<label>5.6</label>
<title>Qualitative Comparisons across Phases and Pathologies</title>
<p><xref ref-type="fig" rid="fig-10">Fig. 10</xref> illustrates side-by-side comparisons of real and synthesized UWF_FA across initial, mid, and final phases for representative conditions (retinal macroaneurysm, normal retina, DR, and retinal vein occlusion). The synthesized images preserve macrovessel topology and lesion boundaries while reproducing phase-dependent contrast filling, supporting both perceptual realism and clinical interpretability.</p>
<fig id="fig-10">
<label>Figure 10</label>
<caption>
<title>Qualitative gallery: real vs. synthetic <inline-formula id="ieqn-108"><mml:math id="mml-ieqn-108"><mml:msub><mml:mi>UWF</mml:mi><mml:mrow><mml:mtext>FA</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula> across initial, mid, and final phases for multiple retinal conditions</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_73387-fig-10.tif"/>
</fig>
</sec>
<sec id="s5_7">
<label>5.7</label>
<title>Summary of Findings</title>
<p>Across quantitative metrics, expert review, and downstream classification, the proposed phase-aware synthesis yields high-fidelity UWF_FA that enhances DR stratification when combined with UWF_RI. Mid-phase reconstructions exhibit the lowest MAE; multi-scale structure is transiently reduced during mid-phase but recovers in final-phase views; and multi-phase synthetic integration delivers the strongest end-task performance.</p>
</sec>
</sec>
<sec id="s6">
<label>6</label>
<title>Discussion</title>
<p>Presenting a phase-aware cross-modal translation framework that synthesizes ultra-widefield fluorescein angiography (UWF_FA) from non-invasive ultra widefield retinal images (UWF_RI), and demonstrated its clinical utility by integrating synthetic angiography into downstream diabetic retinopathy (DR) stratification. Across quantitative fidelity metrics, expert visual reading, and task performance, the multi phase design produced high quality angiograms that improved classification when fused with UWF_RI. See <xref ref-type="sec" rid="s5">Section 5</xref>, <xref ref-type="table" rid="table-2">Tables 2</xref> and <xref ref-type="table" rid="table-4">4</xref>, and <xref ref-type="fig" rid="fig-10">Fig. 10</xref>.</p>

<sec id="s6_1">
<label>6.1</label>
<title>Comparison with Prior Work</title>
<p>Our study advances beyond earlier UWF translation approaches by scaling paired data, enforcing strict registration quality, and learning phase-specific generators. Prior systems such as UWAT-GAN focused on single venous-phase synthesis with limited paired samples, which constrains generalization and fails to model temporal angiographic dynamics. In contrast, we produce initial, mid, and final phases under a unified training and evaluation protocol, enabling richer depiction of nonperfusion and leakage patterns. This broader temporal coverage and the integration with a modern classifier strengthen the link between perceptual realism and clinical task performance.</p>
</sec>
<sec id="s6_2">
<label>6.2</label>
<title>Clinical Relevance of Non-Invasive Multi-Phase FA</title>
<p>UWF imaging captures peripheral pathology that can shift disease staging and management. Angiographic information remains the gold standard for visualizing capillary dropout and leakage but requires dye injection with workflow burden and rare adverse reactions. Our results show that synthetic UWF_FA can complement UWF_RI to improve DR stratification (Set D vs. Set A in <xref ref-type="table" rid="table-4">Table 4</xref>), suggesting a practical path to safer, scalable screening and triage where invasive FA is unavailable or contraindicated.</p>

</sec>
<sec id="s6_3">
<label>6.3</label>
<title>Interpreting Quantitative Trends</title>
<p>The mid phase achieved the lowest MAE (<xref ref-type="table" rid="table-2">Table 2</xref>), indicating the most faithful pixel-level reconstruction during peak circulation contrast. The slight PSNR decline from initial to final phase (<xref ref-type="fig" rid="fig-5">Fig. 5</xref>) likely reflects dispersion and noise accumulation later in the sequence, yet absolute values remain high. SSIM is stable across phases (<xref ref-type="fig" rid="fig-6">Fig. 6</xref>); the MS-SSIM dip in the mid phase suggests transient loss of very fine multi-scale structure during dynamic filling, which is restored by the final phase. These patterns are consistent with the known temporal physiology of FA and the trade-offs between global fidelity and micro-structure preservation at high resolution.</p>

</sec>
<sec id="s6_4">
<label>6.4</label>
<title>Expert Reading and Turing-Style Realism</title>
<p>Two ophthalmologists provided consistent quality ratings with substantial agreement (<inline-formula id="ieqn-109"><mml:math id="mml-ieqn-109"><mml:mi>&#x03BA;</mml:mi></mml:math></inline-formula> 0.78&#x2013;0.82; <xref ref-type="table" rid="table-3">Table 3</xref>, <xref ref-type="fig" rid="fig-7">Fig. 7</xref>), and mislabeled 50%&#x2013;70% of synthetic images as real in a Turing-style task. These findings corroborate the perceptual validity of the synthesized angiograms and align with the quantitative fidelity metrics. Nevertheless, we emphasize that visual plausibility alone is insufficient; multi-metric assessment and task-based endpoints remain essential to guard against overfitting to any single realism measure. While the synthesized UWF-FA images exhibit high quantitative fidelity and strong clinical realism, they are intended to complement rather than replace conventional FA in diagnostic workflows. Synthetic FA generation offers a non-invasive, rapid, and low-risk alternative for scenarios where dye-based angiography is contraindicated or unavailable. In clinical decision-making, such images can support early screening, triage, and treatment planning, but final therapeutic decisions should remain based on physician review and, where appropriate, confirmatory FA acquisitions.</p>

</sec>
<sec id="s6_5">
<label>6.5</label>
<title>Downstream Utility: Why Multi-Phase Helps</title>
<p>Integrating synthetic FA incrementally improved all classification metrics relative to a UWF_RI only baseline, with the full multi-phase setting (Set D) yielding the best AUC, APR, sensitivity, and accuracy (<xref ref-type="table" rid="table-4">Table 4</xref> and <xref ref-type="fig" rid="fig-8">Figs. 8</xref> and <xref ref-type="fig" rid="fig-9">9</xref>). The monotonic gain in sensitivity from Sets A to D indicates that angiographic cues provide complementary information to color imaging particularly for detecting nonperfusion and peripheral vascular pathology not fully captured by UWF_RI. This supports the design choice of phase-specific generators and late-fusion classification.</p>

</sec>
<sec id="s6_6">
<label>6.6</label>
<title>Limitations and Threats to Validity</title>
<p>First, despite stringent registration (<xref ref-type="sec" rid="s4">Section 4</xref>), peripheral distortion and residual artifacts can degrade synthesis near the edges of UWF fields, potentially affecting lesion depiction. Second, although we curated large paired data, domain shift across devices and sites may persist; prospective multi-center validation is needed. Third, Messidor-2 served as external data for classification with synthetic FA derived from UWF_RI, which may not fully reflect clinical deployment where heterogeneous imaging protocols are common. Fourth, our evaluation focused on DR; generalization to other retinal diseases (e.g., RVO, RAM) is suggested by qualitative examples (<xref ref-type="fig" rid="fig-10">Fig. 10</xref>) but requires dedicated studies. Finally, reader studies included a limited number of experts; expanding to larger, geographically diverse panels will improve confidence intervals and reduce annotation bias.</p>
</sec>
<sec id="s6_7">
<label>6.7</label>
<title>Future Work</title>
<p>Several directions can further raise clinical readiness. (i) Data and domain expansion: incorporate additional vendors, populations, and pathologies, with patient-level temporal follow-up to study progression. (ii) Robust learning: add domain adaptation and self-supervised pretraining to mitigate shift, and uncertainty estimation to flag unreliable synthesis. (iii) Architectural advances: explore diffusion or hybrid adversarial&#x2013;diffusion objectives for sharper microvasculature while retaining stability. (iv) Lesion-aware evaluation: augment global metrics with vessel-wise and lesion-level endpoints, calibration, and decision-curve analysis to connect image quality with patient benefit. (v) Prospective trials: assess impact on referral decisions and treatment planning, compare against dye-based FA when ethically feasible, and evaluate operational benefits in teleophthalmology workflows. In summary, phase-aware synthesis of UWF_FA from UWF_RI yields high-fidelity, perceptually convincing angiograms and confers measurable gains in DR stratification when integrated with a transformer-based classifier. By reducing reliance on dye injection while preserving angiographic insight, the approach offers a scalable path to safer, more accessible screening. Continued work on domain robustness, lesion-aware validation, and prospective clinical studies will be key to realizing translational impact at scale.</p>
</sec>
</sec>
<sec id="s7">
<label>7</label>
<title>Conclusion</title>
<p>A phase-aware cross-modal framework was presented that synthesizes ultra-widefield fluorescein angiography (UWF_FA) from non-invasive UWF_RI and demonstrated clinical utility for diabetic retinopathy (DR) stratification. The approach combines strict trans-modal registration, three phase-specific generators (initial, mid, final), and a composite objective augmented with gradient variance loss to better preserve vascular detail. When fused with a transformer-based classifier, multi-phase synthetic angiography consistently outperformed a UWF_RI only baseline, with the full-phase setting (Set D) achieving the strongest end task metrics. See <xref ref-type="table" rid="table-4">Table 4</xref> and <xref ref-type="fig" rid="fig-8">Figs. 8</xref> and <xref ref-type="fig" rid="fig-9">9</xref>. From a fidelity standpoint, mid-phase synthesis attained the lowest MAE while SSIM remained stable and MS-SSIM exhibited a transient mid-phase dip consistent with dynamic contrast filling; expert review showed substantial agreement and frequent real-synthetic confusions, supporting perceptual realism. See <xref ref-type="table" rid="table-2">Tables 2</xref> and <xref ref-type="table" rid="table-3">3</xref> and <xref ref-type="fig" rid="fig-4">Figs. 4</xref>&#x2013;<xref ref-type="fig" rid="fig-7">7</xref>. Limitations include residual peripheral distortions, potential domain shift across devices/centers, and a modest reader cohort; broader disease coverage beyond DR also warrants formal evaluation. Future work will target robustness via domain adaptation and self/weak supervision, hybrid adversarial-diffusion objectives for sharper microvasculature, uncertainty calibration to flag unreliable synthesis, lesion/vessel-level endpoints with decision-curve analysis, and prospective multi center studies to quantify impact on referral and treatment workflows. In sum, phase aware UWF_FA synthesis offers a scalable, dye-free pathway to enrich screening and triage, improving DR stratification while mitigating risks of invasive angiography, and holds promise for safe, accessible retinal care at scale.</p>

</sec>
</body>
<back>
<ack>
<p>The authors extend their appreciation to the Deanship of Research and Graduate Studies at King Khalid University for funding this work through Large Research Project under grant number RGP2/417/46.</p>
</ack>
<sec>
<title>Funding Statement</title>
<p>The work was funded by the Deanship of Research and Graduate Studies at King Khalid University through Large Research Project under grant number RGP2/417/46.</p>
</sec>
<sec>
<title>Author Contributions</title>
<p>The authors confirm contribution to the paper as follows: major contribution to the writing of the article, model building, data extraction and main writer: Damodharan Palaniappan, Tan Kuan Tak and K. Vijayan; overall design and execution: Damodharan Palaniappan, Tan Kuan Tak, K. Vijayan, Balajee Maram, Pravin R Kshirsagar and Naim Ahmad; technical support in data processing and analysis: Damodharan Palaniappan, Tan Kuan Tak and K. Vijayan; overall design and execution: Damodharan Palaniappan, Tan Kuan Tak, K. Vijayan, Balajee Maram, Pravin R Kshirsagar and Naim Ahmad. All authors reviewed the results and approved the final version of the manuscript.</p>
</sec>
<sec sec-type="data-availability">
<title>Availability of Data and Materials</title>
<p>The data that support the findings of this study are available from the Corresponding Author, Naim Ahmad, upon reasonable request.</p>
</sec>
<sec>
<title>Ethics Approval</title>
<p>This study was conducted in accordance with the Declaration of Helsinki and approved by the institutional review board. All 1198 patient images were fully de-identified before analysis. Public datasets (ODIR, Messidor-2) are open-access and anonymized; therefore, no additional consent was required.</p>
</sec>
<sec sec-type="COI-statement">
<title>Conflicts of Interest</title>
<p>The authors declare no conflicts of interest to report regarding the present study.</p>
</sec>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Chatterjee</surname> <given-names>S</given-names></string-name>, <string-name><surname>Fruhling</surname> <given-names>A</given-names></string-name>, <string-name><surname>Kotiadis</surname> <given-names>K</given-names></string-name>, <string-name><surname>Gartner</surname> <given-names>D</given-names></string-name></person-group>. <article-title>Towards new frontiers of healthcare systems research using artificial intelligence and generative AI</article-title>. <source>Health Syst</source>. <year>2024</year>;<volume>13</volume>(<issue>4</issue>):<fpage>263</fpage>&#x2013;<lpage>73</lpage>. doi:<pub-id pub-id-type="doi">10.1080/20476965.2024.2402128</pub-id>; <pub-id pub-id-type="pmid">39584173</pub-id></mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Bajenaru</surname> <given-names>L</given-names></string-name>, <string-name><surname>Tomescu</surname> <given-names>M</given-names></string-name>, <string-name><surname>Grigorovici-Toganel</surname> <given-names>I</given-names></string-name></person-group>. <article-title>Leveraging generative Artificial Intelligence for advanced healthcare solutions</article-title>. <source>Rom J Inf Technol Autom Control</source>. <year>2024</year>;<volume>34</volume>(<issue>3</issue>):<fpage>149</fpage>&#x2013;<lpage>64</lpage>.</mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Bennani</surname> <given-names>T</given-names></string-name></person-group>. <article-title>Advancing healthcare with generativeAI: a multifaceted approach to reliable medical information and innovation [Ph.D. thesis]. Cambridge, MA, USA: Massachusetts Institute of Technology</article-title>; <year>2024</year>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Sai</surname> <given-names>S</given-names></string-name>, <string-name><surname>Gaur</surname> <given-names>A</given-names></string-name>, <string-name><surname>Sai</surname> <given-names>R</given-names></string-name>, <string-name><surname>Chamola</surname> <given-names>V</given-names></string-name>, <string-name><surname>Guizani</surname> <given-names>M</given-names></string-name>, <string-name><surname>Rodrigues</surname> <given-names>JJPC</given-names></string-name></person-group>. <article-title>Generative AI for transformative healthcare: a comprehensive study of emerging models, applications, case studies, and limitations</article-title>. <source>IEEE Access</source>. <year>2024</year>;<volume>12</volume>:<fpage>31078</fpage>&#x2013;<lpage>106</lpage>. doi:<pub-id pub-id-type="doi">10.1109/access.2024.3367715</pub-id>.</mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Yim</surname> <given-names>D</given-names></string-name>, <string-name><surname>Khuntia</surname> <given-names>J</given-names></string-name>, <string-name><surname>Parameswaran</surname> <given-names>V</given-names></string-name>, <string-name><surname>Meyers</surname> <given-names>A</given-names></string-name></person-group>. <article-title>Preliminary evidence of the use of generative AI in health care clinical services: systematic narrative review</article-title>. <source>JMIR Med Inform</source>. <year>2024</year>;<volume>12</volume>(<issue>1</issue>):<fpage>e52073</fpage>. doi:<pub-id pub-id-type="doi">10.2196/52073</pub-id>; <pub-id pub-id-type="pmid">38506918</pub-id></mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ling Ong</surname> <given-names>JC</given-names></string-name>, <string-name><surname>Michael</surname> <given-names>C</given-names></string-name>, <string-name><surname>Ng</surname> <given-names>N</given-names></string-name>, <string-name><surname>Elangovan</surname> <given-names>K</given-names></string-name>, <string-name><surname>Ting Tan</surname> <given-names>NY</given-names></string-name>, <string-name><surname>Jin</surname> <given-names>L</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Generative AI and large language models in reducing medication related harm and adverse drug events&#x2014;a scoping review</article-title>. <source>MedRxiv</source>. <year>2024</year>. doi:<pub-id pub-id-type="doi">10.1101/2024.09.13.24313606</pub-id>.</mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Teo</surname> <given-names>ZL</given-names></string-name>, <string-name><surname>Quek</surname> <given-names>CWN</given-names></string-name>, <string-name><surname>Wong</surname> <given-names>JLY</given-names></string-name>, <string-name><surname>Ting</surname> <given-names>DSW</given-names></string-name></person-group>. <article-title>Cybersecurity in the generative artificial intelligence era</article-title>. <source>Asia Pac J Ophthalmol</source>. <year>2024</year>;<volume>13</volume>(<issue>4</issue>):<fpage>100091</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.apjo.2024.100091</pub-id>; <pub-id pub-id-type="pmid">39209217</pub-id></mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Chen</surname> <given-names>R</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>W</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>B</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>X</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>P</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>S</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>EyeDiff: text-to-image diffusion model improves rare eye disease diagnosis</article-title>. <comment>arXiv:2411.10004. 2024</comment>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Ahmed</surname> <given-names>T</given-names></string-name>, <string-name><surname>Choudhury</surname> <given-names>S</given-names></string-name></person-group>. <article-title>An integrated approach to AI-generated content in e-health</article-title>. <comment>arXiv:2501.16348. 2025</comment>.</mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Gupta</surname> <given-names>M</given-names></string-name>, <string-name><surname>Gupta</surname> <given-names>S</given-names></string-name>, <string-name><surname>Palanisamy</surname> <given-names>G</given-names></string-name>, <string-name><surname>Nisha</surname> <given-names>JS</given-names></string-name>, <string-name><surname>Goutham</surname> <given-names>V</given-names></string-name>, <string-name><surname>Kumar</surname> <given-names>SA</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>A comprehensive survey on detection of ocular and non-ocular diseases using color fundus images</article-title>. <source>IEEE Access</source>. <year>2024</year>;<volume>12</volume>:<fpage>194296</fpage>&#x2013;<lpage>321</lpage>. doi:<pub-id pub-id-type="doi">10.1109/access.2024.3517700</pub-id>.</mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Mochan</surname> <given-names>A</given-names></string-name>, <string-name><surname>Farinha</surname> <given-names>J</given-names></string-name>, <string-name><surname>Bailey</surname> <given-names>G</given-names></string-name>, <string-name><surname>Rodriguez</surname> <given-names>L</given-names></string-name>, <string-name><surname>Zanca</surname> <given-names>F</given-names></string-name>, <string-name><surname>P&#x00F3;lvora</surname> <given-names>A</given-names></string-name>, <etal>et al</etal></person-group>. <source>Imaging the Future-Horizon scanning for emerging technologies and breakthrough innovations in the field of medical imaging and AI</source>. <publisher-loc>Luxembourg</publisher-loc>: <publisher-name>Publications Office of the European Union</publisher-name>; <year>2024</year>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Casu</surname> <given-names>M</given-names></string-name>, <string-name><surname>Guarnera</surname> <given-names>L</given-names></string-name>, <string-name><surname>Caponnetto</surname> <given-names>P</given-names></string-name>, <string-name><surname>Battiato</surname> <given-names>S</given-names></string-name></person-group>. <article-title>GenAI mirage: the impostor bias and the deepfake detection challenge in the era of artificial illusions</article-title>. <source>Forensic Sci Int Digit Investig</source>. <year>2024</year>;<volume>50</volume>(<issue>3</issue>):<fpage>301795</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.fsidi.2024.301795</pub-id>.</mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kawai</surname> <given-names>K</given-names></string-name>, <string-name><surname>Murakami</surname> <given-names>T</given-names></string-name>, <string-name><surname>Mori</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Ishihara</surname> <given-names>K</given-names></string-name>, <string-name><surname>Dodo</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Terada</surname> <given-names>N</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Clinically significant nonperfusion areas on widefield OCT angiography in diabetic retinopathy</article-title>. <source>Ophthalmol Sci</source>. <year>2023</year>;<volume>3</volume>(<issue>1</issue>):<fpage>100241</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.xops.2022.100241</pub-id>; <pub-id pub-id-type="pmid">36545265</pub-id></mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ashraf</surname> <given-names>M</given-names></string-name>, <string-name><surname>Shokrollahi</surname> <given-names>S</given-names></string-name>, <string-name><surname>Salongcay</surname> <given-names>RP</given-names></string-name>, <string-name><surname>Aiello</surname> <given-names>LP</given-names></string-name>, <string-name><surname>Silva</surname> <given-names>PS</given-names></string-name></person-group>. <article-title>Diabetic retinopathy and ultrawide field imaging</article-title>. <source>Semin Ophthalmol</source>. <year>2020</year>;<volume>35</volume>(<issue>1</issue>):<fpage>56</fpage>&#x2013;<lpage>65</lpage>. doi:<pub-id pub-id-type="doi">10.1080/08820538.2020.1729818</pub-id>; <pub-id pub-id-type="pmid">32167854</pub-id></mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>X</given-names></string-name>, <string-name><surname>Ji</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Ma</surname> <given-names>X</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Yi</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Zheng</surname> <given-names>H</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Automated grading of diabetic retinopathy with ultra-widefield fluorescein angiography and deep learning</article-title>. <source>J Diabetes Res</source>. <year>2021</year>;<volume>2021</volume>(<issue>1</issue>):<fpage>2611250</fpage>. doi:<pub-id pub-id-type="doi">10.1155/2021/2611250</pub-id>; <pub-id pub-id-type="pmid">34541004</pub-id></mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Akram</surname> <given-names>MU</given-names></string-name>, <string-name><surname>Khalid</surname> <given-names>S</given-names></string-name>, <string-name><surname>Tariq</surname> <given-names>A</given-names></string-name>, <string-name><surname>Khan</surname> <given-names>SA</given-names></string-name>, <string-name><surname>Azam</surname> <given-names>F</given-names></string-name></person-group>. <article-title>Detection and classification of retinal lesions for grading of diabetic retinopathy</article-title>. <source>Comput Biol Med</source>. <year>2014</year>;<volume>45</volume>(<issue>2</issue>):<fpage>161</fpage>&#x2013;<lpage>71</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.compbiomed.2013.11.014</pub-id>; <pub-id pub-id-type="pmid">24480176</pub-id></mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Alwakid</surname> <given-names>G</given-names></string-name>, <string-name><surname>Gouda</surname> <given-names>W</given-names></string-name>, <string-name><surname>Humayun</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Enhancement of diabetic retinopathy prognostication using deep learning, CLAHE, and ESRGAN</article-title>. <source>Diagnostics</source>. <year>2023</year>;<volume>13</volume>(<issue>14</issue>):<fpage>2375</fpage>. doi:<pub-id pub-id-type="doi">10.3390/diagnostics13142375</pub-id>; <pub-id pub-id-type="pmid">37510123</pub-id></mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Khan</surname> <given-names>MH</given-names></string-name>, <string-name><surname>Mungloo-Dilmohamud</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Jhumka</surname> <given-names>K</given-names></string-name>, <string-name><surname>Mungloo</surname> <given-names>NZ</given-names></string-name>, <string-name><surname>Pena-Reyes</surname> <given-names>C</given-names></string-name></person-group>. <article-title>Investigating on data augmentation and generative adversarial networks (GAN s) for diabetic retinopathy</article-title>. In: <conf-name>2022 International Conference on Electrical, Computer, Communications and Mechatronics Engineering (ICECCME); 2022 Nov 16&#x2013;18</conf-name>; <publisher-loc>Maldives. Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>. p. <fpage>1</fpage>&#x2013;<lpage>5</lpage>.</mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Patrini</surname> <given-names>G</given-names></string-name>, <string-name><surname>Rozza</surname> <given-names>A</given-names></string-name>, <string-name><surname>Krishna Menon</surname> <given-names>A</given-names></string-name>, <string-name><surname>Nock</surname> <given-names>R</given-names></string-name>, <string-name><surname>Qu</surname> <given-names>L</given-names></string-name></person-group>. <article-title>Making deep neural networks robust to label noise: a loss correction approach</article-title>. In: <conf-name>Proceedings of the 2017 IEEE conference on computer vision and pattern recognition; 2017 Jul 21&#x2013;26</conf-name>; <publisher-loc>Honolulu, HI, USA. Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>. p. <fpage>1944</fpage>&#x2013;<lpage>52</lpage>.</mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Dgani</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Greenspan</surname> <given-names>H</given-names></string-name>, <string-name><surname>Goldberger</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Training a neural network based on unreliable human annotation of medical images</article-title>. In: <conf-name>2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018)</conf-name>. <publisher-loc>Piscataway, NJ, USA: Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2018</year>. p. <fpage>39</fpage>&#x2013;<lpage>42</lpage>.</mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Mehralian</surname> <given-names>M</given-names></string-name>, <string-name><surname>Karasfi</surname> <given-names>B</given-names></string-name></person-group>. <article-title>RDCGAN: unsupervised representation learning with regularized deep convolutional generative adversarial networks</article-title>. In: <conf-name>2018 9th Conference on Artificial Intelligence and Robotics and 2nd Asia-Pacific International Symposium</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2018</year>. p. <fpage>31</fpage>&#x2013;<lpage>8</lpage>.</mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Park</surname> <given-names>KB</given-names></string-name>, <string-name><surname>Choi</surname> <given-names>SH</given-names></string-name>, <string-name><surname>Lee</surname> <given-names>JY</given-names></string-name></person-group>. <article-title>M-GAN: retinal blood vessel segmentation by balancing losses through stacked deep fully convolutional networks</article-title>. <source>IEEE Access</source>. <year>2020</year>;<volume>8</volume>:<fpage>146308</fpage>&#x2013;<lpage>22</lpage>. doi:<pub-id pub-id-type="doi">10.1109/access.2020.3015108</pub-id>.</mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Kamran</surname> <given-names>SA</given-names></string-name>, <string-name><surname>Hossain</surname> <given-names>KF</given-names></string-name>, <string-name><surname>Tavakkoli</surname> <given-names>A</given-names></string-name>, <string-name><surname>Zuckerbrod</surname> <given-names>SL</given-names></string-name>, <string-name><surname>Baker</surname> <given-names>SA</given-names></string-name></person-group>. <article-title>Vtgan: semi-supervised retinal image synthesis and disease prediction using vision transformers</article-title>. In: <conf-name>Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision; 2021 Oct 10&#x2013;17</conf-name>; <publisher-loc>Montreal, QC, Canada. Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>. p. <fpage>3235</fpage>&#x2013;<lpage>45</lpage>.</mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Radford</surname> <given-names>A</given-names></string-name>, <string-name><surname>Kim</surname> <given-names>JW</given-names></string-name>, <string-name><surname>Hallacy</surname> <given-names>C</given-names></string-name>, <string-name><surname>Ramesh</surname> <given-names>A</given-names></string-name>, <string-name><surname>Goh</surname> <given-names>G</given-names></string-name>, <string-name><surname>Agarwal</surname> <given-names>S</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Learning transferable visual models from natural language supervision</article-title>. In: <conf-name>International Conference on Machine Learning</conf-name>. <publisher-loc>Westminster, UK</publisher-loc>: <publisher-name>PMLR</publisher-name>; <year>2021</year>. p. <fpage>8748</fpage>&#x2013;<lpage>63</lpage>.</mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kong</surname> <given-names>L</given-names></string-name>, <string-name><surname>Lian</surname> <given-names>C</given-names></string-name>, <string-name><surname>Huang</surname> <given-names>D</given-names></string-name>, <string-name><surname>Li</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Hu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>Q</given-names></string-name></person-group>. <article-title>Breaking the dilemma of medical image-to-image translation</article-title>. <source>Adv Neural Inf Process Syst</source>. <year>2021</year>;<volume>34</volume>:<fpage>1964</fpage>&#x2013;<lpage>78</lpage>.</mixed-citation></ref>
<ref id="ref-26"><label>[26]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Nichol</surname> <given-names>AQ</given-names></string-name>, <string-name><surname>Dhariwal</surname> <given-names>P</given-names></string-name></person-group>. <article-title>Improved denoising diffusion probabilistic models</article-title>. In: <conf-name>International Conference on Machine Learning</conf-name>. <publisher-loc>Westminster, UK</publisher-loc>: <publisher-name>PMLR</publisher-name>; <year>2021</year>. p. <fpage>8162</fpage>&#x2013;<lpage>71</lpage>.</mixed-citation></ref>
<ref id="ref-27"><label>[27]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Nichol</surname> <given-names>A</given-names></string-name>, <string-name><surname>Dhariwal</surname> <given-names>P</given-names></string-name>, <string-name><surname>Ramesh</surname> <given-names>A</given-names></string-name>, <string-name><surname>Shyam</surname> <given-names>P</given-names></string-name>, <string-name><surname>Mishkin</surname> <given-names>P</given-names></string-name>, <string-name><surname>McGrew</surname> <given-names>B</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Glide: towards photorealistic image generation and editing with text-guided diffusion models</article-title>. <comment>arXiv:2112.10741. 2021</comment>.</mixed-citation></ref>
<ref id="ref-28"><label>[28]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Krichen</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Generative adversarial networks</article-title>. In: <conf-name>2023 14th International Conference on Computing Communication and Networking Technologies (ICCCNT)</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2023</year>. p. <fpage>1</fpage>&#x2013;<lpage>7</lpage>.</mixed-citation></ref>
<ref id="ref-29"><label>[29]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Isola</surname> <given-names>P</given-names></string-name>, <string-name><surname>Zhu</surname> <given-names>JY</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>T</given-names></string-name>, <string-name><surname>Efros</surname> <given-names>AA</given-names></string-name></person-group>. <article-title>Image-to-image translation with conditional adversarial networks</article-title>. In: <conf-name>Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2017</year>. p. <fpage>1125</fpage>&#x2013;<lpage>34</lpage>.</mixed-citation></ref>
<ref id="ref-30"><label>[30]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Poles</surname> <given-names>I</given-names></string-name>, <string-name><surname>D&#x2019;arnese</surname> <given-names>E</given-names></string-name>, <string-name><surname>Cellamare</surname> <given-names>LG</given-names></string-name>, <string-name><surname>Santambrogio</surname> <given-names>MD</given-names></string-name>, <string-name><surname>Yi</surname> <given-names>D</given-names></string-name></person-group>. <article-title>Repurposing the image generative potential: exploiting GANs to grade diabetic retinopathy</article-title>. In: <conf-name>Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2024</year>. p. <fpage>2305</fpage>&#x2013;<lpage>14</lpage>.</mixed-citation></ref>
<ref id="ref-31"><label>[31]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Chen</surname> <given-names>X</given-names></string-name>, <string-name><surname>Duan</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Houthooft</surname> <given-names>R</given-names></string-name>, <string-name><surname>Schulman</surname> <given-names>J</given-names></string-name>, <string-name><surname>Sutskever</surname> <given-names>I</given-names></string-name>, <string-name><surname>Abbeel</surname> <given-names>P</given-names></string-name></person-group>. <article-title>Infogan: interpretable representation learning by information maximizing generative adversarial nets</article-title>. In: <conf-name>NIPS&#x2019;16: Proceedings of the 30th International Conference on Neural Information Processing Systems; 2016 Dec 5&#x2013;10</conf-name>; <publisher-loc>Barcelona Spain. Red Hook, NY, USA</publisher-loc>: <publisher-name>Curran Associates Inc.</publisher-name> p. <fpage>2180</fpage>&#x2013;<lpage>8</lpage>.</mixed-citation></ref>
<ref id="ref-32"><label>[32]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Yang</surname> <given-names>L</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Song</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Hong</surname> <given-names>S</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>R</given-names></string-name>, <string-name><surname>Zhao</surname> <given-names>Y</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Diffusion models: a comprehensive survey of methods and applications</article-title>. <source>ACM Comput Surv</source>. <year>2023</year>;<volume>56</volume>(<issue>4</issue>):<fpage>1</fpage>&#x2013;<lpage>39</lpage>. doi:<pub-id pub-id-type="doi">10.1145/3626235</pub-id>.</mixed-citation></ref>
<ref id="ref-33"><label>[33]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Saharia</surname> <given-names>C</given-names></string-name>, <string-name><surname>Chan</surname> <given-names>W</given-names></string-name>, <string-name><surname>Chang</surname> <given-names>H</given-names></string-name>, <string-name><surname>Lee</surname> <given-names>C</given-names></string-name>, <string-name><surname>Ho</surname> <given-names>J</given-names></string-name>, <string-name><surname>Salimans</surname> <given-names>T</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Palette: image-to-image diffusion models</article-title>. In: <conf-name>ACM SIGGRAPH 2022 Conference Proceedings</conf-name>. <publisher-loc>New York, NY, USA</publisher-loc>: <publisher-name>ACM</publisher-name>; <year>2022</year>. p. <fpage>1</fpage>&#x2013;<lpage>10</lpage>.</mixed-citation></ref>
<ref id="ref-34"><label>[34]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ozbey</surname> <given-names>M</given-names></string-name>, <string-name><surname>Dalmaz</surname> <given-names>O</given-names></string-name>, <string-name><surname>Dar</surname> <given-names>SUH</given-names></string-name>, <string-name><surname>Bedel</surname> <given-names>HA</given-names></string-name>, <string-name><surname>Ozturk</surname> <given-names>S</given-names></string-name>, <string-name><surname>Gungor</surname> <given-names>A</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Unsupervised medical image translation with adversarial diffusion models</article-title>. <source>IEEE Trans Med Imaging</source>. <year>2023</year>;<volume>42</volume>(<issue>12</issue>):<fpage>3524</fpage>&#x2013;<lpage>39</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tmi.2023.3290149</pub-id>; <pub-id pub-id-type="pmid">37379177</pub-id></mixed-citation></ref>
<ref id="ref-35"><label>[35]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Binkowski</surname> <given-names>M</given-names></string-name>, <string-name><surname>Sutherland</surname> <given-names>DJ</given-names></string-name>, <string-name><surname>Arbel</surname> <given-names>M</given-names></string-name>, <string-name><surname>Gretton</surname> <given-names>A</given-names></string-name></person-group>. <article-title>Demystifying mmd gans</article-title>. <comment>arXiv:1801.01401. 2018</comment>.</mixed-citation></ref>
<ref id="ref-36"><label>[36]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Chong</surname> <given-names>MJ</given-names></string-name>, <string-name><surname>Forsyth</surname> <given-names>D</given-names></string-name></person-group>. <article-title>Effectively unbiased fid and inception score and where to find them</article-title>. In: <conf-name>Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2020 Jun 13&#x2013;19</conf-name>; <publisher-loc>Seattle, WA, USA. Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>. p. <fpage>6070</fpage>&#x2013;<lpage>9</lpage>.</mixed-citation></ref>
<ref id="ref-37"><label>[37]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Sara</surname> <given-names>U</given-names></string-name>, <string-name><surname>Akter</surname> <given-names>M</given-names></string-name>, <string-name><surname>Uddin</surname> <given-names>MS</given-names></string-name></person-group>. <article-title>Image quality assessment through FSIM, SSIM, MSE and PSNR&#x2014;a comparative study</article-title>. <source>J Comput Commun</source>. <year>2019</year>;<volume>7</volume>(<issue>3</issue>):<fpage>8</fpage>&#x2013;<lpage>18</lpage>. doi:<pub-id pub-id-type="doi">10.4236/jcc.2019.73002</pub-id>.</mixed-citation></ref>
<ref id="ref-38"><label>[38]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Gencer</surname> <given-names>K</given-names></string-name>, <string-name><surname>Gencer</surname> <given-names>G</given-names></string-name>, <string-name><surname>Ceran</surname> <given-names>TH</given-names></string-name>, <string-name><surname>Bilir</surname> <given-names>AE</given-names></string-name>, <string-name><surname>Do&#x01E7;an</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Photodiagnosis with deep learning: a GAN and autoencoder-based approach for diabetic retinopathy detection</article-title>. <source>Photodiagn Photodyn Ther</source>. <year>2025</year>;<volume>53</volume>(<issue>22</issue>):<fpage>104552</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.pdpdt.2025.104552</pub-id>; <pub-id pub-id-type="pmid">40064432</pub-id></mixed-citation></ref>
<ref id="ref-39"><label>[39]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Kalisapudi</surname> <given-names>SSA</given-names></string-name>, <string-name><surname>Raj</surname> <given-names>VD</given-names></string-name>, <string-name><surname>Vanam</surname> <given-names>S</given-names></string-name>, <string-name><surname>Anne</surname> <given-names>JC</given-names></string-name></person-group>. <article-title>Synthesizing realistic ARMD fundus images using generative adversarial networks (GANs)</article-title>. In: <conf-name>International Conference on Intelligent Computing and Communication</conf-name>. <publisher-loc>Cham, Switzerland</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>2022</year>. p. <fpage>587</fpage>&#x2013;<lpage>99</lpage>.</mixed-citation></ref>
<ref id="ref-40"><label>[40]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Alghamdi</surname> <given-names>M</given-names></string-name>, <string-name><surname>Abdel-Mottaleb</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Retinal image augmentation using composed GANs</article-title>. <source>Eng Technol Appl Sci Res</source>. <year>2024</year>;<volume>14</volume>(<issue>6</issue>):<fpage>18525</fpage>&#x2013;<lpage>31</lpage>. doi:<pub-id pub-id-type="doi">10.48084/etasr.8964</pub-id>.</mixed-citation></ref>
<ref id="ref-41"><label>[41]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Fang</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Wei</surname> <given-names>P</given-names></string-name>, <string-name><surname>Li</surname> <given-names>W</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>S</given-names></string-name>, <string-name><surname>Elazab</surname> <given-names>A</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Uwat-gan: fundus fluorescein angiography synthesis via ultra-wide-angle transformation multi-scale gan</article-title>. In: <conf-name>International Conference on Medical Image Computing and Computer-Assisted Intervention</conf-name>. <publisher-loc>Cham, Switzerland</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>2023</year>. p. <fpage>745</fpage>&#x2013;<lpage>55</lpage>.</mixed-citation></ref>
<ref id="ref-42"><label>[42]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ge</surname> <given-names>R</given-names></string-name>, <string-name><surname>Fang</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Wei</surname> <given-names>P</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Jiang</surname> <given-names>H</given-names></string-name>, <string-name><surname>Elazab</surname> <given-names>A</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>UWAFA-GAN: ultra-wide-angle fluorescein angiography transformation via multi-scale generation and registration enhancement</article-title>. <source>IEEE J Biomed Health Inform</source>. <year>2024</year>;<volume>28</volume>(<issue>8</issue>):<fpage>4820</fpage>&#x2013;<lpage>9</lpage>. doi:<pub-id pub-id-type="doi">10.1109/jbhi.2024.3394597</pub-id>; <pub-id pub-id-type="pmid">38683721</pub-id></mixed-citation></ref>
<ref id="ref-43"><label>[43]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Abbood</surname> <given-names>SH</given-names></string-name>, <string-name><surname>Abdull Hamed</surname> <given-names>HN</given-names></string-name>, <string-name><surname>Mohd Rahim</surname> <given-names>MS</given-names></string-name>, <string-name><surname>Alaidi</surname> <given-names>AHM</given-names></string-name>, <string-name><surname>Salim ALRikabi</surname> <given-names>HTH</given-names></string-name></person-group>. <article-title>DR-LL Gan: diabetic retinopathy lesions synthesis using generative adversarial network</article-title>. <source>Int J Online Biomed Eng</source>. <year>2022</year>;<volume>18</volume>(<issue>3</issue>):<fpage>151</fpage>&#x2013;<lpage>63</lpage>. doi:<pub-id pub-id-type="doi">10.3991/ijoe.v18i03.28005</pub-id>.</mixed-citation></ref>
<ref id="ref-44"><label>[44]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Das</surname> <given-names>S</given-names></string-name>, <string-name><surname>Walia</surname> <given-names>P</given-names></string-name></person-group>. <article-title>Enhancing early diabetic retinopathy detection through synthetic DR1 image generation: a StyleGAN3 approach</article-title>. <comment>arXiv:2501.00954. 2025</comment>.</mixed-citation></ref>
<ref id="ref-45"><label>[45]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Anaya-Sanchez</surname> <given-names>H</given-names></string-name>, <string-name><surname>Altamirano-Robles</surname> <given-names>L</given-names></string-name>, <string-name><surname>Diaz-Hernandez</surname> <given-names>R</given-names></string-name>, <string-name><surname>Zapotecas-Martinez</surname> <given-names>S</given-names></string-name></person-group>. <article-title>Wgan-gp for synthetic retinal image generation: enhancing sensor-based medical imaging for classification models</article-title>. <source>Sensors</source>. <year>2024</year>;<volume>25</volume>(<issue>1</issue>):<fpage>167</fpage>. doi:<pub-id pub-id-type="doi">10.3390/s25010167</pub-id>; <pub-id pub-id-type="pmid">39796958</pub-id></mixed-citation></ref>
<ref id="ref-46"><label>[46]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Pakdelmoez</surname> <given-names>S</given-names></string-name>, <string-name><surname>Omidikia</surname> <given-names>S</given-names></string-name>, <string-name><surname>Seyyedsalehi</surname> <given-names>SA</given-names></string-name>, <string-name><surname>Seyyedsalehi</surname> <given-names>SZ</given-names></string-name></person-group>. <article-title>Controllable retinal image synthesis using conditional StyleGAN and latent space manipulation for improved diagnosis and grading of diabetic retinopathy</article-title>. <comment>arXiv:2409.07422. 2024</comment>.</mixed-citation></ref>
<ref id="ref-47"><label>[47]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Liu</surname> <given-names>J</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>S</given-names></string-name>, <string-name><surname>He</surname> <given-names>P</given-names></string-name>, <string-name><surname>Wu</surname> <given-names>S</given-names></string-name>, <string-name><surname>Luo</surname> <given-names>X</given-names></string-name>, <string-name><surname>Deng</surname> <given-names>Y</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>VSG-GAN: a high-fidelity image synthesis method with semantic manipulation in retinal fundus image</article-title>. <source>Biophys J</source>. <year>2024</year>;<volume>123</volume>(<issue>17</issue>):<fpage>2815</fpage>&#x2013;<lpage>29</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.bpj.2024.02.019</pub-id>; <pub-id pub-id-type="pmid">38414236</pub-id></mixed-citation></ref>
<ref id="ref-48"><label>[48]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Pour</surname> <given-names>AM</given-names></string-name>, <string-name><surname>Seyedarabi</surname> <given-names>H</given-names></string-name>, <string-name><surname>Jahromi</surname> <given-names>SHA</given-names></string-name>, <string-name><surname>Javadzadeh</surname> <given-names>A</given-names></string-name></person-group>. <article-title>Automatic detection and monitoring of diabetic retinopathy using efficient convolutional neural networks and contrast limited adaptive histogram equalization</article-title>. <source>IEEE Access</source>. <year>2020</year>;<volume>8</volume>:<fpage>136668</fpage>&#x2013;<lpage>73</lpage>. doi:<pub-id pub-id-type="doi">10.1109/access.2020.3005044</pub-id>.</mixed-citation></ref>
<ref id="ref-49"><label>[49]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kaur</surname> <given-names>J</given-names></string-name>, <string-name><surname>Kaur</surname> <given-names>P</given-names></string-name></person-group>. <article-title>UNIConv: an enhanced U-Net based InceptionV3 convolutional model for DR semantic segmentation in retinal fundus images</article-title>. <source>Concurr Comput Pract Exp</source>. <year>2022</year>;<volume>34</volume>(<issue>21</issue>):<fpage>e7138</fpage>. doi:<pub-id pub-id-type="doi">10.1002/cpe.7138</pub-id>.</mixed-citation></ref>
<ref id="ref-50"><label>[50]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Shi</surname> <given-names>D</given-names></string-name>, <string-name><surname>He</surname> <given-names>S</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>J</given-names></string-name>, <string-name><surname>Zheng</surname> <given-names>Y</given-names></string-name>, <string-name><surname>He</surname> <given-names>M</given-names></string-name></person-group>. <article-title>One-shot retinal artery and vein segmentation via cross-modality pretraining</article-title>. <source>Ophthalmol Sci</source>. <year>2024</year>;<volume>4</volume>(<issue>2</issue>):<fpage>100363</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.xops.2023.100363</pub-id>; <pub-id pub-id-type="pmid">37868792</pub-id></mixed-citation></ref>
<ref id="ref-51"><label>[51]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Alcantarilla</surname> <given-names>PF</given-names></string-name>, <string-name><surname>Nuevo</surname> <given-names>J</given-names></string-name>, <string-name><surname>Bartoli</surname> <given-names>A</given-names></string-name></person-group>. <article-title>Fast explicit diffusion for accelerated features in nonlinear scale spaces</article-title>. In: <conf-name>Proceedings of the British Machine Vision Conference (BMVC); 2011 Aug 29&#x2013;Sep 2</conf-name>; <publisher-loc>Dundee, UK. Durham, UK</publisher-loc>: <publisher-name>BMVA Press</publisher-name>; <year>2011</year>. p. <fpage>1</fpage>&#x2013;<lpage>11</lpage>.</mixed-citation></ref>
<ref id="ref-52"><label>[52]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Fischler</surname> <given-names>MA</given-names></string-name>, <string-name><surname>Bolles</surname> <given-names>RC</given-names></string-name></person-group>. <article-title>Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography</article-title>. <source>Commun ACM</source>. <year>1981</year>;<volume>24</volume>(<issue>6</issue>):<fpage>381</fpage>&#x2013;<lpage>95</lpage>. doi:<pub-id pub-id-type="doi">10.1145/358669.358692</pub-id>.</mixed-citation></ref>
<ref id="ref-53"><label>[53]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Abrahamyan</surname> <given-names>L</given-names></string-name>, <string-name><surname>Truong</surname> <given-names>AM</given-names></string-name>, <string-name><surname>Philips</surname> <given-names>W</given-names></string-name>, <string-name><surname>Deligiannis</surname> <given-names>N</given-names></string-name></person-group>. <article-title>Gradient variance loss for structure-enhanced image super-resolution</article-title>. In: <conf-name>ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</conf-name>; <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2022</year>. p. <fpage>3219</fpage>&#x2013;<lpage>23</lpage>.</mixed-citation></ref>
</ref-list>
</back></article>