<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xml:lang="en" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CMC</journal-id>
<journal-id journal-id-type="nlm-ta">CMC</journal-id>
<journal-id journal-id-type="publisher-id">CMC</journal-id>
<journal-title-group>
<journal-title>Computers, Materials &#x0026; Continua</journal-title>
</journal-title-group>
<issn pub-type="epub">1546-2226</issn>
<issn pub-type="ppub">1546-2218</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">39188</article-id>
<article-id pub-id-type="doi">10.32604/cmc.2023.039188</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>A Hybrid Attention-Based Residual Unet for Semantic Segmentation of Brain Tumor</article-title>
<alt-title alt-title-type="left-running-head">A Hybrid Attention-Based Residual Unet for Semantic Segmentation of Brain Tumor</alt-title>
<alt-title alt-title-type="right-running-head">A Hybrid Attention-Based Residual Unet for Semantic Segmentation of Brain Tumor</alt-title>
</title-group>
<contrib-group>
<contrib id="author-1" contrib-type="author">
<name name-style="western"><surname>Khan</surname><given-names>Wajiha Rahim</given-names></name><xref ref-type="aff" rid="aff-1">1</xref></contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western"><surname>Madni</surname><given-names>Tahir Mustafa</given-names></name><xref ref-type="aff" rid="aff-1">1</xref></contrib>
<contrib id="author-3" contrib-type="author">
<name name-style="western"><surname>Janjua</surname><given-names>Uzair Iqbal</given-names></name><xref ref-type="aff" rid="aff-1">1</xref></contrib>
<contrib id="author-4" contrib-type="author">
<name name-style="western"><surname>Javed</surname><given-names>Umer</given-names></name><xref ref-type="aff" rid="aff-2">2</xref></contrib>
<contrib id="author-5" contrib-type="author">
<name name-style="western"><surname>Khan</surname><given-names>Muhammad Attique</given-names></name><xref ref-type="aff" rid="aff-3">3</xref></contrib>
<contrib id="author-6" contrib-type="author">
<name name-style="western"><surname>Alhaisoni</surname><given-names>Majed</given-names></name><xref ref-type="aff" rid="aff-4">4</xref></contrib>
<contrib id="author-7" contrib-type="author">
<name name-style="western"><surname>Tariq</surname><given-names>Usman</given-names></name><xref ref-type="aff" rid="aff-5">5</xref></contrib>
<contrib id="author-8" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Cha</surname><given-names>Jae-Hyuk</given-names></name><xref ref-type="aff" rid="aff-6">6</xref><email>chajh@hanyang.ac.kr</email></contrib>
<aff id="aff-1"><label>1</label><institution>Department of Computer Science, COMSATS University Islamabad</institution>, <addr-line>Islamabad, 45550</addr-line>, <country>Pakistan</country></aff>
<aff id="aff-2"><label>2</label><institution>Department of Electrical and Computer Engineering, COMSATS University Islamabad, Wah Campus</institution>, <country>Pakistan</country></aff>
<aff id="aff-3"><label>3</label><institution>Department of Computer Science, HITEC University</institution>, <addr-line>Taxila</addr-line>, <country>Pakistan</country></aff>
<aff id="aff-4"><label>4</label><institution>Department of Computer Science, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University</institution>, <addr-line>Riyadh, 11671</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff-5"><label>5</label><institution>Department of Management Information Systems, CoBA, Prince Sattam Bin Abdulaziz University</institution>, <addr-line>Al-Kharj, 16273</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff-6"><label>6</label><institution>Department of Computer Science, Hanyang University</institution>, <addr-line>Seoul, 04763</addr-line>, <country>Korea</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label>Corresponding Author: Jae-Hyuk Cha. Email: <email>chajh@hanyang.ac.kr</email></corresp>
</author-notes>
<pub-date date-type="collection" publication-format="electronic">
<year>2023</year></pub-date>
<pub-date date-type="pub" publication-format="electronic"><day>09</day>
<month>6</month>
<year>2023</year></pub-date>
<volume>76</volume>
<issue>1</issue>
<fpage>647</fpage>
<lpage>664</lpage>
<history>
<date date-type="received">
<day>13</day>
<month>1</month>
<year>2023</year>
</date>
<date date-type="accepted">
<day>10</day>
<month>3</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2023 Khan et al.</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Khan et al.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CMC_39188.pdf"></self-uri>
<abstract>
<p>Segmenting brain tumors in Magnetic Resonance Imaging (MRI) volumes is challenging due to their diffuse and irregular shapes. Recently, 2D and 3D deep neural networks have become famous for medical image segmentation because of the availability of labelled datasets. However, 3D networks can be computationally expensive and require significant training resources. This research proposes a 3D deep learning model for brain tumor segmentation that uses lightweight feature extraction modules to improve performance without compromising contextual information or accuracy. The proposed model, called Hybrid Attention-Based Residual Unet (HA-RUnet), is based on the Unet architecture and utilizes residual blocks to extract low- and high-level features from MRI volumes. Attention and Squeeze-Excitation (SE) modules are also integrated at different levels to learn attention-aware features adaptively within local and global receptive fields. The proposed model was trained on the BraTS-2020 dataset and achieved a dice score of 0.867, 0.813, and 0.787, as well as a sensitivity of 0.93, 0.88, and 0.83 for Whole Tumor, Tumor Core, and Enhancing Tumor, on test dataset respectively. Experimental results show that the proposed HA-RUnet model outperforms the ResUnet and AResUnet base models while having a smaller number of parameters than other state-of-the-art models. Overall, the proposed HA-RUnet model can improve brain tumor segmentation accuracy and facilitate appropriate diagnosis and treatment planning for medical practitioners.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>MRI volumes</kwd>
<kwd>residual Unet</kwd>
<kwd>BraTs-2020</kwd>
<kwd>squeeze-excitation (SE)</kwd>
</kwd-group>
<funding-group>
<award-group id="awg1">
<funding-source>&#x201C;Human Resources Program in Energy Technology&#x201D; of the Korea Institute of Energy Technology Evaluation and Planning (KETEP)</funding-source>
<award-id>20204010600090</award-id>
</award-group>
<award-group id="awg2">
<funding-source>Pakistan&#x0027;s National Center of Artificial Intelligence</funding-source>
</award-group>
</funding-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>A tumor is the abnormal growth of cells within the brain tissues. These unusual cells can be fatal if not diagnosed early [<xref ref-type="bibr" rid="ref-1">1</xref>]. In brain tumors, the most typical type of malignancies reported in adults are gliomas, and they are graded as High-Grade Gliomas (HGG) that increase rapidly and Low-Grade Gliomas (LGG) that grow slowly [<xref ref-type="bibr" rid="ref-2">2</xref>]. Meningiomas are more accessible to segment. However, glioblastomas are challenging to identify because of their high contrast and diffusion [<xref ref-type="bibr" rid="ref-3">3</xref>]. Its appearance varies in size, shape, and form, making it more challenging. Most widely used medical imaging modalities, i.e., Computed Tomography (CT) scans, X-rays, and Magnetic Resonance Imaging (MRI), are commonly used to diagnose diseases. High-resolution MRI is commonly used for brain tumor diagnosis and treatment planning [<xref ref-type="bibr" rid="ref-4">4</xref>]. In 2015, [<xref ref-type="bibr" rid="ref-5">5</xref>] suggested a unique, fully convolutional-based network (FCN) end-to-end model for natural image segmentation that helped transform the entire image segmentation field.</p>
<p>Similarly, [<xref ref-type="bibr" rid="ref-6">6</xref>] Introduced an FCN termed Unet for segmentation in medical images, inspired by the FCN architecture. Unet has two parts: One is an encoder, and the other is the decoder. It also consists of several skip connections between encoder-decoder parts to merge cropped feature maps, significantly improving medical image segmentation performance. Unet has now established itself as a benchmark for tackling the brain tumor segmentation problem. Meanwhile, many enhanced Unet models, such as Residual Unet [<xref ref-type="bibr" rid="ref-7">7</xref>] and Ensemble Unet [<xref ref-type="bibr" rid="ref-8">8</xref>], have also been proposed to enhance the performance of tumor segmentation tasks.</p>
<p>Brain tumor segmentation is a challenging task that has been the focus of extensive research in recent years due to its potential for improving patient diagnosis, treatment, and prognosis. The Brain Tumor Segmentation (BraTS) dataset has emerged as a popular benchmark for evaluating the performance of different segmentation models, providing a standardized platform for comparing results across studies.</p>
<p>However, the BraTS dataset poses several challenges that make accurate segmentation difficult. For instance, the images in the dataset exhibit significant variability in terms of tumor size, shape, location, and appearance, as well as in the presence of confounding factors such as edema and necrosis. Additionally, the dataset is highly imbalanced, with a relatively small number of tumor regions compared to normal brain tissue regions, which can affect the performance of segmentation models.</p>
<p>To address these challenges, researchers have developed various segmentation techniques [<xref ref-type="bibr" rid="ref-9">9</xref>&#x2013;<xref ref-type="bibr" rid="ref-11">11</xref>] that incorporate advanced machine learning algorithms, such as convolutional neural networks (CNNs), and feature engineering methods. Some of these techniques use sophisticated pre-processing methods, such as intensity normalization, bias correction, and skull stripping, to improve the quality of the input data and enhance the performance of segmentation models.</p>
<p>Several studies have explored the potential of these techniques in the context of the BraTS dataset, showing promising results regarding segmentation accuracy and efficiency. However, further research is still needed to develop robust and reliable segmentation models that can address the challenges of brain tumor segmentation in the BraTS dataset.</p>
<p>As Unet must progressively recover down-sampled images due to intrinsic stridden convolution and pooling layers, attention methods can bridge the information flow between shallow and deep layers to facilitate the up-sampling. As a result, this improved the accuracy of the pre-existing Unet models for the MRI segmentation tasks by increasing its capacity to capture multiple local and global responses. Inspired by the performance of Unet, this research integrates the residual attention and Squeeze-Excitation (SE) modules in Residual Unet to explore further the impacts of both local and global features mapping for the segmentation of brain tumors in MRIs. Notably, a 3D deep model for brain tumour segmentation is proposed by stacking attention and Squeeze-Excitation modules at different levels into the Residual Unet for performance enhancement. The following are the contributions of our work:
<list list-type="bullet">
<list-item>
<p>This research proposes a 3D deep and light-weighted, end-to-end model to address the challenge of accurately segmenting brain tumors in MR scans by successfully embedding the residual blocks, attention, and Squeeze-Excitation (SE) modules into the basic Unet model.</p></list-item>
<list-item>
<p>The proposed 3D Residual Unet model integrates low- and high-level features using residual blocks. A series of Attention modules are added on Skip-Connections between corresponding contracting and expanding parts to adaptively enhance local responses of different down-sampled feature maps used for feature recovery in the subsequent up-sampled phase to reduce the semantic gap between both parts. Finally, Squeeze-Excitation (SE) modules are added before up-sampling in the expanding part, recalibrating the channels to suppress the less relevant features and learn the inter-channel relationships.</p></list-item>
<list-item>
<p>Experiments on the publicly available dataset, i.e., BraTs-2020, show that the proposed 3D deep model with attention and Squeeze-Excitation modules outperforms its baseline model and shows competitive performance with a smaller number of parameters for segmenting the brain tumors.</p></list-item>
</list></p>
<p>The rest of the sections in this paper are organized as Section 2, which contains the state-of-the-art methods for semantic segmentation of tumors. Section 3 gives a detailed overview of our proposed methodology, including the dataset, pre-processing strategy, and deep Unet model. Section 4 evaluates the proposed model, presents experimental findings, and compares it to existing models. Section 5 contains the conclusion.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Related Work</title>
<p>Healthcare professionals consider medical image segmentation as a crucial task, which helps them to pinpoint a patient&#x2019;s area of interest and perform a diagnosis. Traditional approaches such as the watershed transform, graph cut, and other methods required handcrafted features and prior information on the input data to obtain impressive performance. However, this still does not fill the gap between computational cost and domain specialists in healthcare. Introducing deep neural networks (DNNs) has been a significant breakthrough in medical image processing. Unlike previous approaches, DNNs can automatically acquire discriminative features, making them capable of segmenting images effectively. Convolutional Neural Networks (CNNs) are the most common deep network for segmenting videos and images. In 2014, a novel fully convolutional-based neural network (FCN) was proposed to enhance the effectiveness of CNN for segmentation in medical images. In 2015, a deep CNN method was proposed for segmenting tissues in the brain using multiple modalities of magnetic resonance images. CNNs are not fully capable of separating tumors with uneven densities and indistinct boundaries. Unet was designed to have a structure comparable to a fully convolutional network to address the volumetric aspect of medical imaging. The attention mechanism has been widely employed in computer vision for various tasks, and its use in medical imaging tasks is minimal. In 2018, a novel attention gate model for diagnosis in medical images called attention Unet was proposed, which could learn to focus on target structures of diverse sizes and shapes. Another study created a hybrid attention-aware deep model to identify liver and extract malignancies from liver CT images in the same year. Customized U-Net architectures have been utilized for effective medical image segmentation.</p>
<p><xref ref-type="table" rid="table-1">Table 1</xref> presents a summary of the methods and their respective applications in medical image segmentation from the year 2007 to 2022. The methods included traditional approaches such as the watershed transform, graph cut, and other methods, deep neural networks (DNNs), and customized U-Net architectures. The attention mechanism has also been introduced for medical image segmentation to identify target structures of diverse sizes and shapes. These methods have demonstrated impressive performance in various medical imaging applications, including tissue segmentation in the brain, brain tumor segmentation, and malignancy extraction from liver CT images.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>Related work on medical image segmentation</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Years</th>
<th>Methods</th>
<th>Applications</th>
</tr>
</thead>
<tbody>
<tr>
<td>2007 till 2014</td>
<td>Watershed transform, graph cut, and other traditional approaches [<xref ref-type="bibr" rid="ref-12">12</xref>&#x2013;<xref ref-type="bibr" rid="ref-15">15</xref>]</td>
<td>Medical image segmentation</td>
</tr>
<tr>
<td>2014</td>
<td>Fully convolutional-based neural network (FCN) [<xref ref-type="bibr" rid="ref-16">16</xref>&#x2013;<xref ref-type="bibr" rid="ref-19">19</xref>]</td>
<td>Medical image segmentation</td>
</tr>
<tr>
<td>2015</td>
<td>Deep CNN method [<xref ref-type="bibr" rid="ref-20">20</xref>,<xref ref-type="bibr" rid="ref-21">21</xref>]</td>
<td>Tissue segmentation in the brain using multiple modalities of MRI</td>
</tr>
<tr>
<td>2016</td>
<td>CNN-based deep model [<xref ref-type="bibr" rid="ref-22">22</xref>,<xref ref-type="bibr" rid="ref-23">23</xref>]</td>
<td>Brain tumor segmentation using MRIs with tiny filters</td>
</tr>
<tr>
<td>2017 till 2018</td>
<td>Attention Unet and Hybrid attention-aware deep model [<xref ref-type="bibr" rid="ref-24">24</xref>&#x2013;<xref ref-type="bibr" rid="ref-31">31</xref>]</td>
<td>Medical image segmentation for diagnosis, Liver CT images, malignancy extraction</td>
</tr>
<tr>
<td>2019 till 2022</td>
<td>Customized U-Net architectures and different attention mechanisms [<xref ref-type="bibr" rid="ref-32">32</xref>&#x2013;<xref ref-type="bibr" rid="ref-44">44</xref>]</td>
<td>Medical image segmentation</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3">
<label>3</label>
<title>Methodology</title>
<p>This research proposes a new hybrid attention-based deep model for the semantic segmentation of brain tumors. The proposed deep learning model comprises three core blocks that segment the brain&#x2019;s tumorous cells. Initially, a 3D residual Unet is developed. Many residual blocks are layered together to acquire multi-scale information by merging low-level features with high-level ones in brain MRIs. Multiple attention modules are embedded at the skip connections between the encoder and decoder to learn multiple attention-aware features of different tumor parts. Finally, squeeze and excitation blocks enhance essential features by recalibrating the feature maps across various channels.</p>
<sec id="s3_1">
<label>3.1</label>
<title>The Proposed Deep Learning Model</title>
<p>Our proposed deep-learning model is depicted in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>. It is relatively like a regular Unet, simultaneously embedded with residual, attention, and squeeze-excitation mechanisms better suited for segmentation tasks in the medical image when the availability of correctly annotated data is limited. The contracting path of the model propagates the contextual information through residual, allowing for a more complex extraction of hierarchical feature maps. The expanding path takes these hierarchical features of varying complexity and reconstructs them from coarse to fine. The Unet, like deep models, benefits from introducing long-range connections across the contracting and expanding parts, allowing distinct hierarchical feature maps to be combined and making the model more robust.</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>The overview of the proposed pipeline. First is the input module that comprises four modalities, i.e., Flair, T1, T2, and T1ce, which are stacked together to form 3D volumes. The second is the proposed deep Residual-based Unet model with attention and Squeeze-Excitations blocks to process the input volumes and segment the Tumor. The third is the output module with three labels</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_39188-fig-1.tif"/>
</fig>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Residual Learning Mechanism</title>
<p>The proposed model consists of Residual blocks as basic building blocks, with basic architecture illustrated in <xref ref-type="fig" rid="fig-2">Fig. 2</xref>. The deep models&#x2019; depth is critical in all computer vision tasks. However, the vanishing of the gradient is a typical problem during the backpropagation in a deep neural network, resulting in poor training outcomes. Reference [<xref ref-type="bibr" rid="ref-23">23</xref>], presented a residual learning network to learn the identity maps. Residual blocks are layered except for the first and final layers to unlock the learning capacity of our proposed deep model. A shortcut identity mapping achieves the residual learning mechanism as fast connections through element-wise addition operation, significantly improving accuracy and training speed without extra parameters. The residual learning mechanism is presented in <xref ref-type="disp-formula" rid="eqn-1">Eq. (1)</xref>:</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>The residual-block used in the proposed deep model. The final output is the sum of Identity mapping and features from three sub blocks containing batch normalization, relu activation, and convolution blocks</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_39188-fig-2.tif"/>
</fig>
<p><disp-formula id="eqn-1"><label>(1)</label><mml:math id="mml-eqn-1" display="block"><mml:msub><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>c</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>I</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>I</mml:mi><mml:mo>+</mml:mo><mml:msub><mml:mrow><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>c</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>I</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>Here, <inline-formula id="ieqn-1"><mml:math id="mml-ieqn-1"><mml:msub><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>c</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>I</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is the output of the residual operation, which is the summation of I; the first input, and <inline-formula id="ieqn-2"><mml:math id="mml-ieqn-2"><mml:msub><mml:mrow><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>c</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> i.e., a mapping function for the residual path to be learned here <italic>i</italic> represents all spatial positions while all channel&#x0027;s indexes are depicted by: <italic>c &#x2208; {1, ..., C}</italic> of input, and C represent the total channels.</p>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Residual-Attention Learning Mechanism</title>
<p>If attention modules are naively stacked in any deep model, the performance of that model will surely suffer. A study [<xref ref-type="bibr" rid="ref-24">24</xref>] presented a novel attention-learning mechanism to solve this problem. This Residual-attention learning mechanism is based on Residua-Bocks and splits into two branches: Trunk-Branch and Soft-Mask-Branch; original features are processed in the Trunk-Branch with t number of residual blocks. The Soft-Mask-Branch constructs the identity mapping using the r number of residual blocks in encoder-decoder type structures. <xref ref-type="disp-formula" rid="eqn-2">(2)</xref> depicts the final output of the residual-attention module for the Residual-Attention learning mechanism</p>
<p><disp-formula id="eqn-2"><label>(2)</label><mml:math id="mml-eqn-2" display="block"><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>A</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>c</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>I</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>I</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>I</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>Here, S(I) is the Output of Soft-Mask-Branch and has values between 0 and 1; if S(I) has values near 0, then <inline-formula id="ieqn-3"><mml:math id="mml-ieqn-3"><mml:msub><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>c</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> (I) will have the feature maps F(I) Trunk-Branch processes them. The Soft-Mask-Branch performs the primary function of enhancing good features and lessening the noise from the trunk branch in the proposed attention-learning mechanism by identifying similar features and minimizing the noise from the Trunk-Branch. As mentioned earlier, the soft mask branch has two parts consisting of D Residual-Blocks, with max-pooling and up-sampling operations, and r skip connections between the corresponding parts. At the end of the soft branch, two layers of convolution and one Sigmoid layer are added to normalize the output from the attention module. The attention module&#x0027;s depth increases concerning the base model&#x0027;s depth. The attention module is illustrated in <xref ref-type="fig" rid="fig-3">Fig. 3</xref>.</p>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>(a) Is the residual-based attention module&#x2019;s architecture used in our proposed model. The architecture of soft-mask-branch is depicted in (b). Here p, t, and r represent the hyperparameters. Soft-mask branch has an encoder-decoder structure. D denotes the depth of these blocks. In our model, D is equal to {1, 2, 3, 4} according to the skip-connections of the proposed model</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_39188-fig-3.tif"/>
</fig>
<p>In general, the original feature maps are conserved by the trunk branch in the proposed attention mechanism, and it pays attention to the features of the brain tumor through the soft mask branch.</p>
</sec>
<sec id="s3_4">
<label>3.4</label>
<title>Squeeze-Excitation Learning Mechanism</title>
<p>Every deep model creates valuable feature maps by combining H &#x00D7; W (spatial) and C (Channel) information inside local-receptive fields at different stages to perform correctly. Soft-Mask-Branch&#x2019;s attention varies adaptably based on the trunk-branch&#x2019;s features in the attention mechanism mentioned above. However, the &#x201C;Squeeze-Excitation&#x201D; (SE) module in the expanding part of the proposed model addresses this limitation of the Residual-attention mechanism. This SE module [<xref ref-type="bibr" rid="ref-26">26</xref>] finds the interdependencies across distinct channels to recalibrate channel-wise feature maps. <xref ref-type="fig" rid="fig-4">Fig. 4</xref> depicts the architecture of the SE module. As a result of this channel-wise feature extraction mechanism, the model can concentrate on the channel dependencies that the model ignores as it recreates the feature maps. The SE module works the following way: First, each channel&#x2019;s features (H &#x00D7; W) are compressed into an actual number using global average pooling. Then, a nonlinear transformation is performed using a fully connected neural network to obtain each feature channel&#x2019;s weight. Finally, the normalized weights obtained above are weighted to the Residual-based channels. These weighted features of respective channels are then concatenated with the up-sampled features maps of the residual blocks, which are then fed into subsequent layers of the corresponding expanding part.</p>
<fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>The architecture of the SE (Squeeze-Excitation) module</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_39188-fig-4.tif"/>
</fig>
<p>The SE normalization technique has shown the potential to enhance the performance of CNNs in brain tumor segmentation. By weighting the feature maps adaptively, SE normalization allows the model to focus on the most significant features that distinguish between different brain tissue types and tumors. At the same time, it can down-weight features prone to noise or artefacts, resulting in improved accuracy and robustness of CNN-based segmentation models for brain tumor analysis.</p>
<p>Generally, a SE block is a structural unit that allows the proposed model to execute dynamic channel-wise feature recalibration to increase its segmentation power. Furthermore, SE blocks shed light on the residual attention mechanism&#x2019;s incapacity to appropriately describe channel-wise feature relationships.</p>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Experiments and Results</title>
<p>Extensive experimentation is performed in this study to evaluate the effectiveness of the proposed tumor segmentation model. All the experiments were performed using python language using the Keras and Tensor Flow frameworks on NVIDIA Titan 1080 TI GPU with a 12 GB GPU workstation. The rest of this section provides the details of data pre-processing, loss function, evaluation matrices, results and finally, the comparison with baseline models.</p>
<sec id="s4_1">
<label>4.1</label>
<title>Dataset Pre-processing</title>
<p>This study used the publicly available BraTs-2020 dataset [<xref ref-type="bibr" rid="ref-11">11</xref>] to evaluate the proposed model. BraTs-2020 contains the pre-operative MRI scans collected from multiple institutes to segment inherently diverse and complex brain tumors with four commonly used modalities. These are T1 weighted (T1), T1 weighted and contrast-enhanced (T1Gd), T2 weighted (T2), and fluid-attenuated inversion recovery (FLAIR) for each case, as shown in <xref ref-type="fig" rid="fig-5">Fig. 5</xref>. All the MRI scans are of 240 &#x00D7; 240 resolution, and 155 slices for each modality with manually annotated ground-truth labels for every scan. The BraTs-2020 training set comprises 369 patients. The hyperparameters were defined as a learning rate of 0.001, a dropout rate of 0.2 the number of epochs was 100 with a batch size of 1.</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>BraTs-2020 dataset with four different modalities with segmented masks and labels</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_39188-fig-5.tif"/>
</fig>
<p>Similarly, the validation set includes 125 patients. However, the ground-truth labels are made public for the training set. In contrast, withheld labels for validation sets validate the proposed model&#x2019;s generalization and scalability using the online validation approach.</p>
<p>Segmenting brain tumors in MRIs is a difficult challenge because of the brain&#x2019;s and biological tissue&#x2019;s complex structure and the medical imaging quality. Due to the restricted processing resources, this model downsized each image from 240 &#x00D7; 240 &#x00D7; 155 voxels to 128 &#x00D7; 128 &#x00D7; 128 voxels. Input MRI scans must be normalized because pixel intensities vary due to manufacturers, acquisition conditions, and sequences. Each MRI sequence&#x2019;s non-zero voxels are normalized independently through zero-score normalization and shrink input data. <xref ref-type="disp-formula" rid="eqn-3">(3)</xref> illustrates Z-Score normalization:</p>
<p><disp-formula id="eqn-3"><label>(3)</label><mml:math id="mml-eqn-3" display="block"><mml:mi>Z</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>I</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>&#x03BC;</mml:mi></mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mfrac></mml:math></disp-formula></p>
<p>Here, Z is the output image obtained after normalization on the input image I, &#x03BC; is the mean, and &#x03B4; is the standard deviation. There is no additional image processing. All MR images fed into our model have a final input size of 128 &#x00D7; 128 &#x00D7; 128 &#x00D7; 4, where 128 &#x00D7; 128 indicates the spatial dimensions of the input image (i.e., width and height), 128 represents the number of slices, and 4 represents the four MRI modalities. This research stacked all the modalities together to generate four channels to feed into the model. All the annotated training sets&#x2019; masks are handled the same way. The ultimate size of the segmented masks input into the model is 128 &#x00D7; 128 &#x00D7; 128 &#x00D7; 4. The first three numbers reflect the resolution and the number of scan slices. 4 represents the number of classes provided, namely: label 0 (Background), label 4 (ET (GD Enhancing-Tumor)), label 2 (ED (Peritumoral-Edema)), and label 1 (NCR/NET (Necrotic and Non-Enhancing Tumor-Core)). Even though deep models are mainly resistant to noise, data processing is still necessary for improving brain tumor segmentation accuracy. In this research, we performed data processing to make the available brain MRI scans better suited to train the deep model according to the available resources, as shown in <xref ref-type="fig" rid="fig-6">Fig. 6</xref>.</p>
<fig id="fig-6">
<label>Figure 6</label>
<caption>
<title>Dataset pre-processing steps</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_39188-fig-6.tif"/>
</fig>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Loss Function</title>
<p>The proposed model determines if a pixel represents a specific brain tumor tissue or a regular pixel. However, severe class imbalance problems are usually exhibited using the BraTs dataset for segmentation. <xref ref-type="table" rid="table-2">Table 2</xref> depicts the distribution of tumor subclasses and healthy tissues in the Brats-2020 training set.</p>
<table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>Given Labels and the Rate of sub-class distribution in the training set of BraTS-2020</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Label</th>
<th>Sub classes</th>
<th>Rate</th>
</tr>
</thead>
<tbody>
<tr>
<td>0</td>
<td>Background</td>
<td>98.46%</td>
</tr>
<tr>
<td>1</td>
<td>NCR/NET (Necrotic and Non-Enhancing-Tumor-Core)</td>
<td>1.02%</td>
</tr>
<tr>
<td>2</td>
<td>ED (Peritumoral-Edema)</td>
<td>0.29%</td>
</tr>
<tr>
<td>4</td>
<td>ET (Enhancing-Tumor)</td>
<td>0.23%</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>As brain tumors, mainly the TC (tumor core), are often tiny in the overall MRI Scans, CE (Cross-Entropy) loss is ineffective in solving this class imbalance problem in the segmentation task. The well-known, class-wise Dice overlap coefficient is adopted as a loss function in such tasks, outperforming the CE. For multi-class segmentation problems, the Dice-Coefficient for each respective class can be determined using <xref ref-type="disp-formula" rid="eqn-4">Eq. (4)</xref>:</p>
<p><disp-formula id="eqn-4"><label>(4)</label><mml:math id="mml-eqn-4" display="block"><mml:mi>L</mml:mi><mml:mi>o</mml:mi><mml:mi>s</mml:mi><mml:msub><mml:mi>s</mml:mi><mml:mrow><mml:mi>D</mml:mi><mml:mi>i</mml:mi><mml:mi>c</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>c</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:munderover><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover><mml:mi>s</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>i</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mi>g</mml:mi><mml:mi>t</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>i</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover><mml:mi>s</mml:mi><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>i</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:mi>g</mml:mi><mml:mi>t</mml:mi><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>i</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
<p>Here N represents the number of volumes (subjects) that, in our case, is 365, <italic>s (c, i)</italic> &#x2208; [0,1,2,3] represents the predicted output (softmax output) of the model, and <italic>gt (c, i)</italic> &#x2208; [0,1,2,3] is the given mask with four labels for tumor sub-classes, respectively. This loss function is used to check the similarity index of multi-class samples directly.</p>
</sec>
<sec id="s4_3">
<label>4.3</label>
<title>Evaluation Metrics</title>
<p>An overlap and a distance metric are two primary evaluation metrics for the deep model&#x0027;s evaluation. The Dice Similarity Coefficient (DSC) is a regularly used metric for determining overlap between two sets. It compares two groups, S1 and S2, by normalizing their intersection sizes over the average of their sizes. It is defined as follows in the context of ground truth comparison in <xref ref-type="disp-formula" rid="eqn-5">Eq. (5)</xref>:</p>
<p><disp-formula id="eqn-5"><label>(5)</label><mml:math id="mml-eqn-5" display="block"><mml:mi>D</mml:mi><mml:mi>S</mml:mi><mml:mi>C</mml:mi><mml:mo>=</mml:mo><mml:mn>2</mml:mn><mml:mfrac><mml:mrow><mml:mo>|</mml:mo><mml:mi>S</mml:mi><mml:mn>1</mml:mn><mml:mo>&#x2229;</mml:mo><mml:mi>S</mml:mi><mml:mn>2</mml:mn><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mrow><mml:mo>|</mml:mo><mml:mi>S</mml:mi><mml:mn>1</mml:mn><mml:mo>|</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mo>|</mml:mo><mml:mi>S</mml:mi><mml:mn>2</mml:mn><mml:mo>|</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
<p>Sensitivity, specificity, and Hausdorff distance are statistical decision theory metrics determined using <xref ref-type="disp-formula" rid="eqn-6">Eqs. (6)</xref> to <xref ref-type="disp-formula" rid="eqn-8">(8)</xref>, respectively.</p>
<p><disp-formula id="eqn-6"><label>(6)</label><mml:math id="mml-eqn-6" display="block"><mml:mrow><mml:mi mathvariant="italic">S</mml:mi><mml:mi mathvariant="italic">e</mml:mi><mml:mi mathvariant="italic">n</mml:mi><mml:mi mathvariant="italic">s</mml:mi><mml:mi mathvariant="italic">i</mml:mi><mml:mi mathvariant="italic">t</mml:mi><mml:mi mathvariant="italic">i</mml:mi><mml:mi mathvariant="italic">v</mml:mi><mml:mi mathvariant="italic">i</mml:mi><mml:mi mathvariant="italic">t</mml:mi><mml:mi mathvariant="italic">y</mml:mi></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>p</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
<p><disp-formula id="eqn-7"><label>(7)</label><mml:math id="mml-eqn-7" display="block"><mml:mrow><mml:mi mathvariant="italic">S</mml:mi><mml:mi mathvariant="italic">p</mml:mi><mml:mi mathvariant="italic">e</mml:mi><mml:mi mathvariant="italic">c</mml:mi><mml:mi mathvariant="italic">i</mml:mi><mml:mi mathvariant="italic">f</mml:mi><mml:mi mathvariant="italic">i</mml:mi><mml:mi mathvariant="italic">c</mml:mi><mml:mi mathvariant="italic">i</mml:mi><mml:mi mathvariant="italic">t</mml:mi><mml:mi mathvariant="italic">y</mml:mi></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>n</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
<p><disp-formula id="eqn-8"><label>(8)</label><mml:math id="mml-eqn-8" display="block"><mml:mi>H</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>A</mml:mi><mml:mo>,</mml:mo><mml:mi>B</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>b</mml:mi><mml:mi>&#x03F5;</mml:mi><mml:mi>B</mml:mi></mml:mrow></mml:msub><mml:mi>a</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>b</mml:mi></mml:math></disp-formula></p>
<p>Hausdorff-Distance is complementary with the Dice coefficient metric to quantify the maximum distance between the margins of original and predicted volumes. In <xref ref-type="disp-formula" rid="eqn-8">(8)</xref>, A (Actual Mask) and B (Predicted Mask) are the two sets of points being compared, and H (A,B) is the directed Hausdorff distances between them. Here a is a point in set A, b is a point in set B, and the function max and min (a,b) calculate the maximum and minimum Euclidean distance between a point in set A and a point in set B. It penalizes outliers heavily, as a model prediction might have voxel-perfect overlap. However, its value will be significant if the predicted voxel is far from the actual segmentation mask. This metric may appear noisier than the Dice-Coefficient, but it helps determine the segmentation&#x2019;s clinical utility. For example, suppose the predicted segmented mask includes surrounding healthy brain tissues. It will require manual adjustment from the radiation oncologist, even if the overlap of predicted and actual masks assessed through the Dice metric is good enough to avoid fatal effects on the patient.</p>
</sec>
<sec id="s4_4">
<label>4.4</label>
<title>Comparison with SOTA Methods</title>
<p>In this study, a new method for brain tumor segmentation was proposed that achieves state-of-the-art performance on the BraTS 2020 challenge dataset. Our method leverages the power of deep learning. It combines several innovative components, such as a dual-attention mechanism, a residual-attention mechanism, and a squeeze and excitation attention mechanism, to improve the segmentation&#x0027;s accuracy and robustness. We extensively evaluate our approach using standard evaluation metrics and statistical tests and show that it outperforms previous state-of-the-art methods on BraTS 2020. In particular, our method achieves an average dice score of 0.87 on the Training set and 0.86 on the test set, with less number of parameters as shown in <xref ref-type="table" rid="table-3">Table 3</xref>. Furthermore, in <xref ref-type="table" rid="table-4">Tables 4</xref>&#x2013;<xref ref-type="table" rid="table-7">7</xref>, we performed an ablation study to investigate the contribution of each component to the overall performance.</p>
<table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>Comparison with SOTA methods</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Methods</th>
<th>Average dice score<break/>(Training set)</th>
<th>Average dice score<break/>(Validation set)</th>
<th>Number of parameters</th>
</tr>
</thead>
<tbody>
<tr>
<td>SA-Net [<xref ref-type="bibr" rid="ref-45">45</xref>]</td>
<td>0.86</td>
<td>0.85</td>
<td>16.5M</td>
</tr>
<tr>
<td>Vanilla U-Net</td>
<td>0.85</td>
<td>0.83</td>
<td>17.8M</td>
</tr>
<tr>
<td>Proposed method</td>
<td>0.87</td>
<td>0.86</td>
<td>13.2M</td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-4">
<label>Table 4</label>
<caption>
<title>Evaluation metrics of the proposed model on BraTS_2020 Train and Validation sets. WT (Whole-Tumor), TC (Tumor-Core), and ET (Enhancing-Tumor)</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th rowspan="2">Results</th>
<th rowspan="2">Labels</th>
<th colspan="3" align="center">Dice-coefficient</th>
<th colspan="3" align="center">Sensitivity</th>
<th colspan="3" align="center">Specificity</th>
<th colspan="3" align="left">Hausdorff-distance</th>
</tr>
<tr>
<th>ET</th>
<th>WT</th>
<th>TC</th>
<th>ET</th>
<th>WT</th>
<th>TC</th>
<th>ET</th>
<th>WT</th>
<th>TC</th>
<th>ET</th>
<th>WT</th>
<th>TC</th>
</tr> 
</thead>
<tbody>
<tr>
<td rowspan="5">Training (365 MRI cases)</td>
<td>Mean</td>
<td>0.75</td>
<td>0.92</td>
<td>0.86</td>
<td>0.86</td>
<td>0.93</td>
<td>0.85</td>
<td>1.00</td>
<td>0.93</td>
<td>0.98</td>
<td>23.49</td>
<td>5.10</td>
<td>4.30</td>
</tr>
<tr>
<td>StdDev</td>
<td>0.23</td>
<td>0.06</td>
<td>0.20</td>
<td>0.23</td>
<td>0.08</td>
<td>0.23</td>
<td>0.03</td>
<td>0.05</td>
<td>0.02</td>
<td>86.11</td>
<td>11.11</td>
<td>7.84</td>
</tr>
<tr>
<td>Median</td>
<td>0.89</td>
<td>0.94</td>
<td>0.94</td>
<td>0.94</td>
<td>0.95</td>
<td>0.94</td>
<td>0.99</td>
<td>0.94</td>
<td>0.99</td>
<td>1.41</td>
<td>2.24</td>
<td>1.73</td>
</tr>
<tr>
<td>25quantile</td>
<td>0.82</td>
<td>0.90</td>
<td>0.88</td>
<td>0.88</td>
<td>0.91</td>
<td>0.88</td>
<td>0.96</td>
<td>0.99</td>
<td>0.98</td>
<td>1.00</td>
<td>1.73</td>
<td>1.41</td>
</tr>
<tr>
<td>75quantile</td>
<td>0.93</td>
<td>0.95</td>
<td>0.96</td>
<td>0.96</td>
<td>0.97</td>
<td>0.97</td>
<td>0.99</td>
<td>0.96</td>
<td>0.99</td>
<td>2.00</td>
<td>3.74</td>
<td>3.00</td>
</tr>
<tr>
<td rowspan="5">Validation (125 MRI cases)</td>
<td>Mean</td>
<td>0.65</td>
<td>0.86</td>
<td>0.72</td>
<td>0.74</td>
<td>0.84</td>
<td>0.69</td>
<td>0.79</td>
<td>0.94</td>
<td>0.97</td>
<td>7.22</td>
<td>9.14</td>
<td>10.79</td>
</tr>
<tr>
<td>StdDev</td>
<td>0.30</td>
<td>0.10</td>
<td>0.26</td>
<td>0.24</td>
<td>0.15</td>
<td>0.30</td>
<td>0.03</td>
<td>0.06</td>
<td>0.04</td>
<td>11.15</td>
<td>13.66</td>
<td>14.50</td>
</tr>
<tr>
<td>Median</td>
<td>0.79</td>
<td>0.89</td>
<td>0.83</td>
<td>0.82</td>
<td>0.89</td>
<td>0.81</td>
<td>0.99</td>
<td>0.96</td>
<td>0.99</td>
<td>3.00</td>
<td>4.69</td>
<td>5.87</td>
</tr>
<tr>
<td>25quantile</td>
<td>0.57</td>
<td>0.85</td>
<td>0.63</td>
<td>0.65</td>
<td>0.79</td>
<td>0.51</td>
<td>1.00</td>
<td>0.92</td>
<td>0.96</td>
<td>1.73</td>
<td>3.00</td>
<td>2.83</td>
</tr>
<tr>
<td>75quantile</td>
<td>0.87</td>
<td>0.92</td>
<td>0.91</td>
<td>0.89</td>
<td>0.93</td>
<td>0.91</td>
<td>1.00</td>
<td>1.00</td>
<td>1.00</td>
<td>6.06</td>
<td>7.68</td>
<td>12.57</td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-5">
<label>Table 5</label>
<caption>
<title>Summary of each model</title>
</caption>
<table frame="hsides" >
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Models</th>
<th>Time cost (Per epoch)</th>
<th>Trainable parameters</th>
</tr>
</thead>
<tbody>
<tr>
<td>Base model (Residual Unet)</td>
<td>2.3 min</td>
<td>4,212,947</td>
</tr>
<tr>
<td>Residual unet &#x002B; Attention</td>
<td>4 min</td>
<td>12,514,616</td>
</tr>
<tr>
<td>Proposed model (Residual Unet &#x002B; Attention &#x002B; SE)</td>
<td>5.2 min</td>
<td>13,253,348</td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-6">
<label>Table 6</label>
<caption>
<title>Comparison with base models on dice-coefficient and hausdorff distance-95 through the online evaluation process</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th rowspan="2">Results</th>
<th rowspan="2">Model</th>
<th align="center" colspan="3">Mean dice-coefficient</th>
<th align="center" colspan="3">Mean hausdorff distance-95</th>
</tr>
<tr>
<th>ET</th>
<th>WT</th>
<th>TC</th>
<th>ET</th>
<th>WT</th>
<th>TC</th>
</tr>
</thead>
<tbody>
<tr>
<td></td>
<td>Base model (Residual Unet)</td>
<td>0.378</td>
<td>0.558</td>
<td>0.415</td>
<td>52.43</td>
<td>38.14</td>
<td>27.12</td>
</tr>
<tr>
<td>Training set</td>
<td>Residual Unet &#x002B; Attention</td>
<td>0.693</td>
<td>0.877</td>
<td>0.80</td>
<td>36.81</td>
<td>15.60</td>
<td>11.07</td>
</tr>
<tr>
<td></td>
<td>Proposed model (Residual Unet &#x002B; Attention &#x002B; SE)</td>
<td>0.807</td>
<td>0.909</td>
<td>0.843</td>
<td>23.49</td>
<td>5.101</td>
<td>4.295</td>
</tr>
<tr>
<td></td>
<td>Base model (Residual Unet)</td>
<td>0.215</td>
<td>0.434</td>
<td>0.296</td>
<td>28.92</td>
<td>32.54</td>
<td>26.05</td>
</tr>
<tr>
<td>Validation set</td>
<td>Residual Unet &#x002B; Attention</td>
<td>0.64</td>
<td>0.816</td>
<td>0.679</td>
<td>13.97</td>
<td>25.55</td>
<td>13.58</td>
</tr>
<tr>
<td></td>
<td>Proposed model (Residual Unet &#x002B; Attention &#x002B; SE)</td>
<td>0.787</td>
<td>0.869</td>
<td>0.813</td>
<td>7.217</td>
<td>9.138</td>
<td>10.79</td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-7">
<label>Table 7</label>
<caption>
<title>Comparison of deep models by Sensitivity and Specificity through the online evaluation process</title>
</caption>
<table frame="hsides" >
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th rowspan="2">Results</th>
<th rowspan="2">Model</th>
<th align="center" colspan="3">Mean sensitivity</th>
<th align="center" colspan="3">Mean specificity</th>
</tr>
<tr>
<th>ET</th>
<th>WT</th>
<th>TC</th>
<th>ET</th>
<th>WT</th>
<th>TC</th>
</tr>
</thead>
<tbody>
<tr>
<td></td>
<td>Base model (Residual Unet)</td>
<td>0.664</td>
<td>0.784</td>
<td>0.674</td>
<td>0.938</td>
<td>0.905</td>
<td>0.942</td>
</tr>
<tr>
<td>Training set</td>
<td>Residual Unet &#x002B; Attention</td>
<td>0.765</td>
<td>0.891</td>
<td>0.794</td>
<td>0.965</td>
<td>0.917</td>
<td>0.969</td>
</tr>
<tr>
<td></td>
<td>Proposed model (Residual Unet &#x002B; Attention &#x002B; SE)</td>
<td>0.863</td>
<td>0.985</td>
<td>0.849</td>
<td>0.999</td>
<td>0.925</td>
<td>0.983</td>
</tr>
<tr>
<td></td>
<td>Base model (Residual Unet)</td>
<td>0.634</td>
<td>0.739</td>
<td>0.552</td>
<td>0.897</td>
<td>0.886</td>
<td>0.891</td>
</tr>
<tr>
<td>Validation set</td>
<td>Residual Unet &#x002B; Attention</td>
<td>0.748</td>
<td>0.842</td>
<td>0.659</td>
<td>0.997</td>
<td>0.987</td>
<td>0.997</td>
</tr>
<tr>
<td></td>
<td>Proposed model (Residual Unet &#x002B; Attention &#x002B; SE)</td>
<td>0.836</td>
<td>0.936</td>
<td>0.889</td>
<td>0.987</td>
<td>0.938</td>
<td>0.997</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s4_5">
<label>4.5</label>
<title>Results and Discussion</title>
<p>The proposed 3D deep model completely utilizes volume information while capturing three-dimensional data. The proposed 3D lightweight model integrates different modules to create a precise segmentation mask. Meanwhile, training with large MRI volumes of 128 &#x00D7; 128 &#x00D7; 128 &#x00D7; 4 gives significantly more contextual-based information than training with small patches. The brain volumes received by the model are passed from top to bottom during the contracting phase. Lower features were then passed top-down in the contracting path through different pooling layers and reversed during the expanding phase, with doubled resolution thanks to up-sampling. The attention blocks are used to realize the links between both parts. The feature maps from these attention blocks were concatenated with corresponding up-sampling and SE level of expanding part. The final brain tumor segmentation probability map was generated using an activation layer (i.e., Softmax). <xref ref-type="table" rid="table-4">Table 4</xref> presents our model&#x0027;s training and validation results (Mean, Standard Deviation, Median and 25 &#x0026; 75 Quantile) obtained through the online evaluation system for Brats-2020.</p>

<p>Our proposed deep learning model relies on three primary modules. 1st: a simple 3D basic Unet model that uses residual blocks rather than convolutional blocks like traditional Unet. 2nd: Residual Attention module added to the base model on skip Connections. 3rd: SE (Squeeze and excitation) module to derive information from the feature maps; SE module is utilized with upsampling layer after each residual unit in the expanding part. In this research, we trained the proposed deep model in this way; first, our base residual Unet model was trained, then added the attention modules, and finally added SE modules in the proposed model. For comparison, we acquired quantitative evaluations of each tumor class with baseline models regarding Dice-Coefficient and Hausdorff-Distance. This research trained all three models without weight initialization for the convolutional kernel parameters with Adam optimizer and Dice Loss. LR &#x003D; 0.001 was the initial learning rate. The time cost and the total number of trainable parameters for each model are depicted in <xref ref-type="table" rid="table-5">Table 5</xref>.</p>

<p>In this research, we evaluate the performance after adding each corresponding module in the baseline model through BraTs-Online Evaluation-System for training and validation sets. <xref ref-type="fig" rid="fig-7">Fig. 7</xref> gives a graphical representation of evaluation metrics for all models, while <xref ref-type="table" rid="table-6">Tables 6</xref> and <xref ref-type="table" rid="table-7">7</xref> depict the numeric comparison of Evaluation metrics of the proposed model with base models.</p>
<fig id="fig-7">
<label>Figure 7</label>
<caption>
<title>Boxplot demonstrates the evaluation metric for three deep models, including the base model, base model with attention, and proposed model. (a) Dice-coefficient for both training and validation sets, (b) Hausdroff95 for training and validation sets, (c) Sensitivity for training and validation sets (d) Specificity for training and validation sets. The different tumor types explained in Section 4.1 represent the abbreviation in the legend. (orange) ET is an acronym for Enhancing-Tumor (yellow) TC stands for Tumor-Core, and (green) WT is for Whole-Tumor</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_39188-fig-7a.tif"/><graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_39188-fig-7b.tif"/>
</fig>
<p>By comparing the (1) and (3) rows of <xref ref-type="table" rid="table-6">Table 6</xref>, this research finds that our final model&#x2019;s Dice Score of WT, TC, and ET of base model Residual Unet increased for training and validation sets. The Hausdorff-Distance of all tumors was also reduced by adding deep supervision through two additional modules in the base model. These two modules&#x2019; inclusion helped extract other discriminative tumor features, as seen by the improvement in most measures. By comparing the results of the last two rows in <xref ref-type="table" rid="table-5">Table 5</xref>, the Dice of all tumor classes in the base model with both attention and SE modules were better than those of the base model with only the attention module.</p>

<p>The proposed 3D deep model has fewer parameters than the traditional Unet. Adding Attention and SE modules to the base model increases the trainable parameters. However, the above-mentioned results demonstrated the effectiveness of each module individually that is added base residual Unet model along with model&#x2019;s depth. A few segmented examples of our proposed model and baseline models are shown in <xref ref-type="fig" rid="fig-8">Fig. 8</xref>.</p>
<fig id="fig-8">
<label>Figure 8</label>
<caption>
<title>The segmented outputs for unseen MRIs from the BraTs training set, from left to right: four modalities (FLAIR, T1, T1CE, T2), ground truth, and prediction of our models; for this instance, 1st Row: Patient ID 366, slice 90; 2nd Row: Patient ID 369, slice 100; Colors: Red is (Tumor-Core), Yellow is (GD-Enhancing-Tumor), and Green is (Peritumoral-Edema)</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_39188-fig-8.tif"/>
</fig>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Conclusion</title>
<p>This research presented a light-weighted 3D deep model for the segmentation task of tumors using brain MRIs by embedding attention, Squeeze-Excitation mechanisms, and residual blocks into Unet to perform pixel-to-pixel segmentation. By including a series of attention modules on the skip connections and Squeeze-Excitation modules on the corresponding expanding path, the proposed model adaptively enhances local responses of residual features maps by reducing the noise and rescaling these features maps by assigning channel weights in the recovery process of the expanding part. The effectiveness of the proposed model is demonstrated through experimental results. Our model outperformed the baseline models, i.e., Residual Unet and Residual-attention Unet, and showed comparable performance with other state-of-the-art brain tumor segmentation methods with the fewest parameters on the benchmark of the BraTs database. The proposed model expands the Unet family and can be extended to different medical imaging domains critical for real-world applications. However, there is a drawback in using resized MRI volumes as inputs to train the segmentation model, which leads to the potential loss of information on brain tumors due to the limitation of available computational resources. Therefore, in the future, some post-processing methods will be explored to improve the segmentation results further. In addition, to make the architecture more general for various datasets and more flexible for everyday medical imaging tasks, this research will further explore other powerful feature extraction modules to improve extra performance.</p>
</sec>
</body>
<back>
<sec><title>Funding Statement</title>
<p>This work was supported by &#x201C;Human Resources Program in Energy Technology&#x201D; of the Korea Institute of Energy Technology Evaluation and Planning (KETEP), granted financial resources from the Ministry of Trade, Industry &#x0026; Energy, Republic of Korea. (No. 20204010600090). This study was conducted at the Medical Imaging and Diagnostics Lab at COMSATS University Islamabad and was funded by Pakistan&#x0027;s National Center of Artificial Intelligence.</p>
</sec>
<sec sec-type="data-availability"><title>Availability of Data and Materials</title>
<p>A publicly available BRATS 2020 dataset was used for analyzing our model. This dataset can be found at <ext-link ext-link-type="uri" xlink:href="http://ipp.cbica.upenn.edu">http://ipp.cbica.upenn.edu</ext-link>.</p>
</sec>
<sec sec-type="COI-statement"><title>Conflicts of Interest</title>
<p>The authors declare that they have no conflicts of interest to report regarding the present study.</p>
</sec>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>N.</given-names> <surname>Gordillo</surname></string-name>, <string-name><given-names>E.</given-names> <surname>Montseny</surname></string-name> and <string-name><given-names>P.</given-names> <surname>Sobrevilla</surname></string-name></person-group>, &#x201C;<article-title>State of the art survey on MRI brain tumor segmentation</article-title>,&#x201D; <source>Magnetic Resonance Imaging</source>, vol. <volume>31</volume>, no. <issue>2</issue>, pp. <fpage>1426</fpage>&#x2013;<lpage>1438</lpage>, <year>2013</year>; <pub-id pub-id-type="pmid">23790354</pub-id></mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>F.</given-names> <surname>Demir</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Akbulut</surname></string-name>, <string-name><given-names>B.</given-names> <surname>Ta&#x015F;c&#x0131;</surname></string-name> and <string-name><given-names>K.</given-names> <surname>Demir</surname></string-name></person-group>, &#x201C;<article-title>Improving brain tumor classification performance with an effective approach based on new deep learning model named 3ACL from 3D MRI data</article-title>,&#x201D; <source>Biomedical Signal Processing and Control</source>, vol. <volume>81</volume>, no. <issue>11</issue>, pp. <fpage>104424</fpage>, <year>2023</year>.</mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>X.</given-names> <surname>Du</surname></string-name>, <string-name><given-names>Q.</given-names> <surname>He</surname></string-name>, <string-name><given-names>B.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>N.</given-names> <surname>Li</surname></string-name> and <string-name><given-names>W.</given-names> <surname>Li</surname></string-name></person-group>, &#x201C;<article-title>Diagnostic accuracy of diffusion-weighted imaging in differentiating glioma recurrence from posttreatment-related changes: A meta-analysis</article-title>,&#x201D; <source>Expert Review of Anticancer Therapy</source>, vol. <volume>22</volume>, no. <issue>4</issue>, pp. <fpage>123</fpage>&#x2013;<lpage>130</lpage>, <year>2022</year>; <pub-id pub-id-type="pmid">34727815</pub-id></mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Chattopadhyay</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Maitra</surname></string-name></person-group>, &#x201C;<article-title>MRI-based brain tumor image detection using CNN based deep learning method</article-title>,&#x201D; <source>Neuroscience Informatics</source>, vol. <volume>21</volume>, no. <issue>2</issue>, pp. <fpage>100060</fpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>R.</given-names> <surname>Mehta</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Filos</surname></string-name>, <string-name><given-names>U.</given-names> <surname>Baid</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Sako</surname></string-name> and <string-name><given-names>R.</given-names> <surname>McKinley</surname></string-name></person-group>, &#x201C;<article-title>QU-BraTS: MICCAI BraTS, 2020 challenge on quantifying uncertainty in brain tumor segmentation-analysis of ranking scores and benchmarking results</article-title>,&#x201D; <source>Journal of Machine Learning for Biomedical Imaging</source>, vol. <volume>1</volume>, no. <issue>2</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>21</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Long</surname></string-name>, <string-name><given-names>E.</given-names> <surname>Shelhamer</surname></string-name> and <string-name><given-names>T.</given-names> <surname>Darrell</surname></string-name></person-group>, &#x201C;<article-title>Fully convolutional networks for semantic segmentation</article-title>,&#x201D; in <conf-name>Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition</conf-name>, <publisher-loc>NY, USA</publisher-loc>, pp. <fpage>3431</fpage>&#x2013;<lpage>3440</lpage>, <year>2015</year>. </mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Dong</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Yang</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Liu</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Guo</surname></string-name></person-group>, &#x201C;<article-title>Automatic brain tumor detection and segmentation using U-Net based fully convolutional networks</article-title>,&#x201D; in <conf-name>Annual Conf. on Medical Image Understanding and Analysis</conf-name>, <publisher-loc>NY, USA</publisher-loc>, pp. <fpage>506</fpage>&#x2013;<lpage>517</lpage>, <year>2017</year>. </mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>K. V.</given-names> <surname>Kumari</surname></string-name> and <string-name><given-names>S. S.</given-names> <surname>Barpanda</surname></string-name></person-group>, &#x201C;<article-title>Residual UNet with dual attention&#x2014;An ensemble residual UNet with dual attention for multi-modal and multi-class brain MRI segmentation</article-title>,&#x201D; <source>International Journal of Imaging Systems and Technology</source>, vol. <volume>4</volume>, no. <issue>5</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>23</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Hu</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Gu</surname></string-name> and <string-name><given-names>X.</given-names> <surname>Gu</surname></string-name></person-group>, &#x201C;<article-title>Mutual ensemble learning for brain tumor segmentation</article-title>,&#x201D; <source>Neurocomputing</source>, vol. <volume>504</volume>, no. <issue>5</issue>, pp. <fpage>68</fpage>&#x2013;<lpage>81</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Ng</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Huang</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Ong</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Foong</surname></string-name> and <string-name><given-names>W.</given-names> <surname>Nowinski</surname></string-name></person-group>, &#x201C;<article-title>Medical image segmentation using watershed segmentation with texture-based region merging</article-title>,&#x201D; in <conf-name>2008 30th Annual Int. Conf. of the IEEE Engineering in Medicine and Biology Society</conf-name>, <publisher-loc>NY, USA</publisher-loc>, pp. <fpage>4039</fpage>&#x2013;<lpage>4042</lpage>, <year>2008</year>. </mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Hamza</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Alhaisoni</surname></string-name>, <string-name><given-names>A. Al</given-names> <surname>Hejaili</surname></string-name>, <string-name><given-names>K. A.</given-names> <surname>Shaban</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Alsubai</surname></string-name></person-group>, &#x201C;<article-title>D2BOF-COVIDNet: A framework of deep bayesian optimization and fusion-assisted optimal deep features for COVID-19 classification using chest x-ray and MRI scans</article-title>,&#x201D; <source>Diagnostics</source>, vol. <volume>13</volume>, no. <issue>2</issue>, pp. <fpage>101</fpage>, <year>2023</year>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Ajmal</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Akram</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Alqahtani</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Alhaisoni</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Armghan</surname></string-name></person-group>, &#x201C;<article-title>BF2SkNet: Best deep learning features fusion-assisted framework for multiclass skin lesion classification</article-title>,&#x201D; <source>Neural Computing and Applications</source>, vol. <volume>5</volume>, no. <issue>2</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>17</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Alhaisoni</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Alqahtani</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Alsubai</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Alharbi</surname></string-name></person-group>, &#x201C;<article-title>Multimodal brain tumor detection and classification using deep saliency map and improved dragonfly optimization algorithm</article-title>,&#x201D; <source>International Journal of Imaging Systems and Technology</source>, vol. <volume>21</volume>, no. <issue>4</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>21</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Alhaisoni</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Nazir</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Alqahtani</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Binbusayyis</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Alsubai</surname></string-name></person-group>, &#x201C;<article-title>A healthcare system for COVID19 classification using multi-type classical features selection</article-title>,&#x201D; <source>Computers, Material and Continua</source>, vol. <volume>73</volume>, no. <issue>2</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>16</lpage>, <year>2023</year>.</mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Masood</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Maham</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Javed</surname></string-name>, <string-name><given-names>U.</given-names> <surname>Tariq</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Kadry</surname></string-name></person-group>, &#x201C;<article-title>Brain MRI analysis using deep neural network for medical of internet things applications</article-title>,&#x201D; <source>Computers and Electrical Engineering</source>, vol. <volume>103</volume>, no. <issue>4</issue>, pp. <fpage>108386</fpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Hamza</surname></string-name>, <string-name><given-names>S. H.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Alqahtani</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Alsubai</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Binbusayyis</surname></string-name></person-group>, &#x201C;<article-title>COVID-19 classification using chest X-ray images: A framework of CNN-LSTM and improved max value moth flame optimization</article-title>,&#x201D; <source>Frontiers in Public Health</source>, vol. <volume>10</volume>, no. <issue>2</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>24</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>N.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Ruan</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Lebonvallet</surname></string-name>, <string-name><given-names>Q.</given-names> <surname>Liao</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Zhu</surname></string-name></person-group>, &#x201C;<article-title>Kernel feature selection to fuse multi-spectral MRI images for brain tumor segmentation</article-title>,&#x201D; <source>Computer Vision and Image Understanding</source>, vol. <volume>115</volume>, no. <issue>6</issue>, pp. <fpage>256</fpage>&#x2013;<lpage>269</lpage>, <year>2011</year>.</mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Botoca</surname></string-name></person-group>, &#x201C;<article-title>Cellular neural networks assisted automatic detection of elements in microscopic medical images. A preliminary study</article-title>,&#x201D; in <conf-name>2014 11th Int. Symp. on Electronics and Telecommunications (ISETC)</conf-name>, <publisher-loc>NY, USA</publisher-loc>, pp. <fpage>1</fpage>&#x2013;<lpage>4</lpage>, <year>2014</year>. </mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>W.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Li</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Deng</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Wang</surname></string-name> and <string-name><given-names>W.</given-names> <surname>Lin</surname></string-name></person-group>, &#x201C;<article-title>Deep convolutional neural networks for multi-modality isointense infant brain image segmentation</article-title>,&#x201D; <source>NeuroImage</source>, vol. <volume>108</volume>, no. <issue>6</issue>, pp. <fpage>214</fpage>&#x2013;<lpage>224</lpage>, <year>2015</year>; <pub-id pub-id-type="pmid">25562829</pub-id></mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>K. L.</given-names> <surname>Hua</surname></string-name>, <string-name><given-names>C. H.</given-names> <surname>Hsu</surname></string-name>, <string-name><given-names>S. C.</given-names> <surname>Hidayati</surname></string-name> and <string-name><given-names>Y. J.</given-names> <surname>Chen</surname></string-name></person-group>, &#x201C;<article-title>Computer-aided classification of lung nodules on computed tomography images via deep learning technique</article-title>,&#x201D; <source>OncoTargets and Therapy</source>, vol. <volume>8</volume>, no. <issue>6</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>21</lpage>, <year>2015</year>.</mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Pereira</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Pinto</surname></string-name>, <string-name><given-names>V.</given-names> <surname>Alves</surname></string-name> and <string-name><given-names>C. A.</given-names> <surname>Silva</surname></string-name></person-group>, &#x201C;<article-title>Brain tumor segmentation using convolutional neural networks in MRI images</article-title>,&#x201D; <source>IEEE Transactions on Medical Imaging</source>, vol. <volume>35</volume>, no. <issue>4</issue>, pp. <fpage>1240</fpage>&#x2013;<lpage>1251</lpage>, <year>2016</year>; <pub-id pub-id-type="pmid">26960222</pub-id></mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>O.</given-names> <surname>Ronneberger</surname></string-name>, <string-name><given-names>P.</given-names> <surname>Fischer</surname></string-name> and <string-name><given-names>T.</given-names> <surname>Brox</surname></string-name></person-group>, &#x201C;<article-title>U-net: Convolutional networks for biomedical image segmentation</article-title>,&#x201D; in <conf-name>Int. Conf. on Medical Image Computing and Computer-Assisted Intervention</conf-name>, <publisher-loc>NY, USA</publisher-loc>, pp. <fpage>234</fpage>&#x2013;<lpage>241</lpage>, <year>2015</year>. </mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>F.</given-names> <surname>Milletari</surname></string-name>, <string-name><given-names>N.</given-names> <surname>Navab</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Ahmadi</surname></string-name></person-group>, &#x201C;<article-title>V-Net: Fully convolutional neural networks for volumetric medical image segmentation</article-title>,&#x201D; in <conf-name>Proc. 2016 4th Int. Conf. 3D</conf-name>, <publisher-loc>Vision, NY, USA</publisher-loc>, pp. <fpage>565</fpage>&#x2013;<lpage>571</lpage>, <year>2016</year>. </mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Liu</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Tong</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Chen</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Jiang</surname></string-name> and <string-name><given-names>F.</given-names> <surname>Zhou</surname></string-name></person-group>, &#x201C;<article-title>Deep learning based brain tumor segmentation: A survey</article-title>,&#x201D; <source>Complex &#x0026; Intelligent Systems</source>, vol. <volume>6</volume>, no. <issue>2</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>26</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>F.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Jiang</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Qian</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Yang</surname></string-name> and <string-name><given-names>C.</given-names> <surname>Li</surname></string-name></person-group>, &#x201C;<article-title>Residual attention network for image classification</article-title>,&#x201D; in <conf-name>Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition</conf-name>, <publisher-loc>NY, USA</publisher-loc>, pp. <fpage>3156</fpage>&#x2013;<lpage>3164</lpage>, <year>2017</year>. </mixed-citation></ref>
<ref id="ref-26"><label>[26]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Chen</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Kalantidis</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Li</surname></string-name> and <string-name><given-names>J.</given-names> <surname>Feng</surname></string-name></person-group>, &#x201C;<article-title>A 2-nets: Double attention networks</article-title>,&#x201D; <source>Advances in Neural Information Processing Systems</source>, vol. <volume>31</volume>, no. <issue>5</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>22</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-27"><label>[27]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Nam</surname></string-name>, <string-name><given-names>J. W.</given-names> <surname>Ha</surname></string-name> and <string-name><given-names>J.</given-names> <surname>Kim</surname></string-name></person-group>, &#x201C;<article-title>Dual attention networks for multimodal reasoning and matching</article-title>,&#x201D; in <conf-name>Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition</conf-name>, <publisher-loc>NY, USA</publisher-loc>, pp. <fpage>299</fpage>&#x2013;<lpage>307</lpage>, <year>2017</year>. </mixed-citation></ref>
<ref id="ref-28"><label>[28]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Yu</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Peng</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Gao</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Yu</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Learning a discriminative feature network for semantic segmentation</article-title>,&#x201D; in <conf-name>Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition</conf-name>, <publisher-loc>NY, USA</publisher-loc>, pp. <fpage>1857</fpage>&#x2013;<lpage>1866</lpage>, <year>2018</year>. </mixed-citation></ref>
<ref id="ref-29"><label>[29]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>O.</given-names> <surname>Oktay</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Schlemper</surname></string-name>, <string-name><given-names>L. L.</given-names> <surname>Folgoc</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Heinrich</surname></string-name></person-group>, &#x201C;<article-title>Attention u-net: Learning where to look for the pancreas</article-title>,&#x201D; <source>ArXiv Preprint</source>, vol. <volume>4</volume>, no. <issue>2</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>11</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-30"><label>[30]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Q.</given-names> <surname>Jin</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Meng</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Sun</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Cui</surname></string-name> and <string-name><given-names>R.</given-names> <surname>Su</surname></string-name></person-group>, &#x201C;<article-title>RA-UNet: A hybrid deep attention-aware network to extract liver and tumor in CT scans</article-title>,&#x201D; <source>Frontiers in Bioengineering and Biotechnology</source>, vol. <volume>5</volume>, no. <issue>2</issue>, pp. <fpage>1471</fpage>, <year>2020</year>. <pub-id pub-id-type="doi">10.3389/fbioe.2020.605132</pub-id>; <pub-id pub-id-type="pmid">33425871</pub-id></mixed-citation></ref>
<ref id="ref-31"><label>[31]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Li</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Yang</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Ni</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Elazab</surname></string-name> and <string-name><given-names>J.</given-names> <surname>Wu</surname></string-name></person-group>, &#x201C;<article-title>TA-Net: Triple attention network for medical image segmentation</article-title>,&#x201D; <source>Computers in Biology and Medicine</source>, vol. <volume>137</volume>, no. <issue>2</issue>, pp. <fpage>104836</fpage>, <year>2021</year>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2021.104836</pub-id>; <pub-id pub-id-type="pmid">34507157</pub-id></mixed-citation></ref>
<ref id="ref-32"><label>[32]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Gao</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Zhao</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Luo</surname></string-name> and <string-name><given-names>C.</given-names> <surname>Liang</surname></string-name></person-group>, &#x201C;<article-title>Dense encoder-decoder network based on two-level context enhanced residual attention mechanism for segmentation of breast tumors in magnetic resonance imaging</article-title>,&#x201D; in <conf-name>2019 IEEE Int. Conf. on Bioinformatics and Biomedicine (BIBM)</conf-name>, <publisher-loc>NY, USA</publisher-loc>, pp. <fpage>1123</fpage>&#x2013;<lpage>1129</lpage>, <year>2019</year>. </mixed-citation></ref>
<ref id="ref-33"><label>[33]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Zou</surname></string-name> and <string-name><given-names>P. X.</given-names> <surname>Liu</surname></string-name></person-group>, &#x201C;<article-title>Hybrid dilation and attention residual U-Net for medical image segmentation</article-title>,&#x201D; <source>Computers in Biology and Medicine</source>, vol. <volume>134</volume>, no. <issue>13</issue>, pp. <fpage>104449</fpage>, <year>2021</year>; <pub-id pub-id-type="pmid">33993015</pub-id></mixed-citation></ref>
<ref id="ref-34"><label>[34]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Hu</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Shen</surname></string-name> and <string-name><given-names>G.</given-names> <surname>Sun</surname></string-name></person-group>, &#x201C;<article-title>Squeeze-and-excitation networks</article-title>,&#x201D; in <conf-name>Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition</conf-name>, <publisher-loc>NY, USA</publisher-loc>, pp. <fpage>7132</fpage>&#x2013;<lpage>7141</lpage>, <year>2018</year>. </mixed-citation></ref>
<ref id="ref-35"><label>[35]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>B.</given-names> <surname>Menze</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Jakab</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Bauer</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Farahani</surname></string-name> and <string-name><given-names>J.</given-names> <surname>Kirby</surname></string-name></person-group>, &#x201C;<article-title>The multimodal brain tumor image segmentation benchmark (brats) medical imaging</article-title>,&#x201D; <source>IEEE Transactions on Medical Imaging</source>, vol. <volume>24</volume>, no. <issue>1</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>32</lpage>, <year>2014</year>.</mixed-citation></ref>
<ref id="ref-36"><label>[36]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Shah</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Xiao</surname></string-name>, <string-name><given-names>N.</given-names> <surname>Subbanna</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Francis</surname></string-name> and <string-name><given-names>D. L.</given-names> <surname>Arnold</surname></string-name></person-group>, &#x201C;<article-title>Evaluating intensity normalization on MRIs of human brain with multiple sclerosis</article-title>,&#x201D; <source>Medical Image Analysis</source>, vol. <volume>15</volume>, no. <issue>7</issue>, pp. <fpage>267</fpage>&#x2013;<lpage>282</lpage>, <year>2011</year>; <pub-id pub-id-type="pmid">21233004</pub-id></mixed-citation></ref>
<ref id="ref-37"><label>[37]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Liu</surname></string-name></person-group>, &#x201C;<article-title>Deep learning based brain tumor segmentation: A survey</article-title>,&#x201D; <source>Sensors</source>, vol. <volume>2</volume>, no. <issue>4</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>21</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-38"><label>[38]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>D.</given-names> <surname>Ni</surname></string-name></person-group>, &#x201C;<article-title>A novel multi-modal MRI synthesis network with enhanced quality and controllability for brain tumor segmentation</article-title>,&#x201D; <source>Medical Image Analysis</source>, vol. <volume>66</volume>, no. <issue>2</issue>, pp. <fpage>101770</fpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-39"><label>[39]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Pan</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Chen</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Yin</surname></string-name></person-group>, &#x201C;<article-title>BRAINS: Boundary-aware attentive network for brain tumor segmentation</article-title>,&#x201D; <source>Medical Image Analysis</source>, vol. <volume>61</volume>, no. <issue>2</issue>, pp. <fpage>101670</fpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-40"><label>[40]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Li</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Yang</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Ni</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Elazab</surname></string-name></person-group>, &#x201C;<article-title>TA-Net: Triple attention network for medical image segmentation</article-title>,&#x201D; <source>Computer Biology in Medicine</source>, vol. <volume>137</volume>, no. <issue>3</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>21</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-41"><label>[41]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Zou</surname></string-name> and <string-name><given-names>P. X.</given-names> <surname>Liu</surname></string-name></person-group>, &#x201C;<article-title>Hybrid dilation and attention residual U-Net for medical image segmentation</article-title>,&#x201D; <source>Computer Biology in Medicine</source>, vol. <volume>134</volume>, no. <issue>12</issue>, pp. <fpage>104449</fpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-42"><label>[42]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Lu</surname></string-name></person-group>, &#x201C;<article-title>Multilevel volumetric attention network for brain tumor segmentation</article-title>,&#x201D; <source>IEEE Transactions on Neural Networks and Learning Systems</source>, vol. <volume>32</volume>, no. <issue>8</issue>, pp. <fpage>3295</fpage>&#x2013;<lpage>3307</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-43"><label>[43]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Zhang</surname></string-name></person-group>, &#x201C;<article-title>CBAM-Unet: A novel convolutional block attention module based Unet for accurate brain tumor segmentation</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>9</volume>, no. <issue>3</issue>, pp. <fpage>115399</fpage>&#x2013;<lpage>115409</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-44"><label>[44]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>X.</given-names> <surname>Li</surname></string-name></person-group>, &#x201C;<article-title>FCGAN: Fully convolutional generative adversarial network for brain tumor segmentation</article-title>,&#x201D; <source>Frontiers in Neuroscience</source>, vol. <volume>15</volume>, no. <issue>3</issue>, pp. <fpage>602391</fpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-45"><label>[45]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Yuan</surname></string-name></person-group>, &#x201C;<article-title>Automatic brain tumor segmentation with scale attention network</article-title>,&#x201D; in <conf-name>BrainLes@MICCAI Conf.</conf-name>, <publisher-loc>NY, USA</publisher-loc>, pp. <fpage>1</fpage>&#x2013;<lpage>7</lpage>, <year>2020</year>. </mixed-citation></ref>
</ref-list>
</back>
</article>