<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CMC</journal-id>
<journal-id journal-id-type="nlm-ta">CMC</journal-id>
<journal-id journal-id-type="publisher-id">CMC</journal-id>
<journal-title-group>
<journal-title>Computers, Materials &#x0026; Continua</journal-title>
</journal-title-group>
<issn pub-type="epub">1546-2226</issn>
<issn pub-type="ppub">1546-2218</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">17101</article-id>
<article-id pub-id-type="doi">10.32604/cmc.2021.017101</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Segmentation and Classification of Stomach Abnormalities Using Deep Learning</article-title>
<alt-title alt-title-type="left-running-head">Segmentation and Classification of Stomach Abnormalities Using Deep Learning</alt-title>
<alt-title alt-title-type="right-running-head">Segmentation and Classification of Stomach Abnormalities Using Deep Learning</alt-title>
</title-group>
<contrib-group content-type="authors">
<contrib id="author-1" contrib-type="author">
<name name-style="western">
<surname>Naz</surname>
<given-names>Javeria</given-names>
</name>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western">
<surname>Khan</surname>
<given-names>Muhammad Attique</given-names>
</name>
<xref ref-type="aff" rid="aff-1">1</xref></contrib>
<contrib id="author-3" contrib-type="author">
<name name-style="western">
<surname>Alhaisoni</surname>
<given-names>Majed</given-names>
</name>
<xref ref-type="aff" rid="aff-2">2</xref></contrib>
<contrib id="author-4" contrib-type="author" corresp="yes">
<name name-style="western">
<surname>Song</surname>
<given-names>Oh-Young</given-names>
</name>
<xref ref-type="aff" rid="aff-3">3</xref></contrib>
<contrib id="author-5" contrib-type="author">
<name name-style="western">
<surname>Tariq</surname>
<given-names>Usman</given-names>
</name>
<xref ref-type="aff" rid="aff-4">4</xref></contrib>
<contrib id="author-6" contrib-type="author">
<name name-style="western">
<surname>Kadry</surname>
<given-names>Seifedine</given-names>
</name>
<xref ref-type="aff" rid="aff-5">5</xref></contrib>
<aff id="aff-1"><label>1</label><institution>Department of Computer Science, HITEC University Taxila</institution>, <addr-line>Taxila</addr-line>, <country>Pakistan</country></aff>
<aff id="aff-2"><label>2</label><institution>College of Computer Science and Engineering, University of Ha&#x2019;il</institution>, <addr-line>Ha&#x2019;il</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff-3"><label>3</label><institution>Department of Software, Sejong University</institution>, <addr-line>Seoul, Korea, Gwangjin-gu, Korea</addr-line></aff>
<aff id="aff-4"><label>4</label><institution>College of Computer Engineering and Sciences, Prince Sattam Bin Abdulaziz University</institution>, <addr-line>Al-Khraj</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff-5"><label>5</label><institution>Faculty of Applied Computing and Technology, Noroff University College</institution>, <addr-line>Kristiansand</addr-line>, <country>Norway</country></aff>
</contrib-group>
<author-notes><corresp id="cor1">&#x002A;Corresponding Author: Oh-Young Song. Email: <email>Oysong@sejong.edu</email></corresp></author-notes>
<pub-date pub-type="epub" date-type="pub" iso-8601-date="2021-05-31"><day>31</day><month>05</month><year>2021</year></pub-date>
<volume>69</volume>
<issue>1</issue>
<fpage>607</fpage>
<lpage>625</lpage>
<history>
<date date-type="received"><day>18</day><month>01</month><year>2021</year></date>
<date date-type="accepted"><day>21</day><month>02</month><year>2021</year></date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2021 Naz et al.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Naz et al.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CMC_17101.pdf"></self-uri>
<abstract>
<p>An automated system is proposed for the detection and classification of GI abnormalities. The proposed method operates under two pipeline procedures: (a) segmentation of the bleeding infection region and (b) classification of GI abnormalities by deep learning. The first bleeding region is segmented using a hybrid approach. The threshold is applied to each channel extracted from the original RGB image. Later, all channels are merged through mutual information and pixel-based techniques. As a result, the image is segmented. Texture and deep learning features are extracted in the proposed classification task. The transfer learning (TL) approach is used for the extraction of deep features. The Local Binary Pattern (LBP) method is used for texture features. Later, an entropy-based feature selection approach is implemented to select the best features of both deep learning and texture vectors. The selected optimal features are combined with a serial-based technique and the resulting vector is fed to the Ensemble Learning Classifier. The experimental process is evaluated on the basis of two datasets: Private and KVASIR. The accuracy achieved is 99.8 per cent for the private data set and 86.4 percent for the KVASIR data set. It can be confirmed that the proposed method is effective in detecting and classifying GI abnormalities and exceeds other methods of comparison.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Gastrointestinal tract</kwd>
<kwd>contrast stretching</kwd>
<kwd>segmentation</kwd>
<kwd>deep learning</kwd>
<kwd>features selection</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>Computerized detection of human diseases is an emerging research domain for the last two decades [<xref ref-type="bibr" rid="ref-1">1</xref>,<xref ref-type="bibr" rid="ref-2">2</xref>]. In medical imaging, numerous remarkable methods were developed for automated medical diagnosis systems [<xref ref-type="bibr" rid="ref-3">3</xref>]. Stomach infection is one of the most common types that caused a large number of deaths every year [<xref ref-type="bibr" rid="ref-4">4</xref>]. The most common type of disease related to the stomach is colorectal cancer. It affects both women and men [<xref ref-type="bibr" rid="ref-5">5</xref>]. Colorectal cancer consists of three significant infections that are polyp, ulcer, and bleeding. In 2015, about 132,000 cases of colorectal cancer were recorded in the USA [<xref ref-type="bibr" rid="ref-6">6</xref>]. A survey conducted in 2017 showed that approximately 21% population of the USA is suffering from gastrointestinal infections, and 765,000 deaths are noticed due to maladies found in the stomach [<xref ref-type="bibr" rid="ref-7">7</xref>]. According to the global cancer report of 2018 for 20 regions of the world, total estimated cases of cancer are 18.1 million. Among them, 9.2% for deaths and 6.1% for new cases relate to colorectal cancer. According to the American cancer society, in 2019, approximately 27,510 new stomach cancer cases of both genders (consisting of 10,280 women&#x2019;s and 17,230 men&#x2019;s) are observed in the US. A total of 11,140 deaths (consisting of 4340 women&#x2019;s and 6800 men&#x2019;s) are noticed in 2019 due to colorectal cancer [<xref ref-type="bibr" rid="ref-8">8</xref>].</p>
<p>GI infections can be easily cured if they were diagnosed at an early stage. As small bowel has a complex structure, that is why push gastroscopy is not considered as the best choice for the diagnosis of small bowel infections like bleeding, polyp, and ulcer [<xref ref-type="bibr" rid="ref-9">9</xref>]. The traditional endoscopic method is an invasive method that is not utilized by endoscopists and is also not recommended to patients due to its high level of discomfort [<xref ref-type="bibr" rid="ref-10">10</xref>]. These problems were resolved by a new technology introduced in the year 2000, namely, Wireless Capsule Endoscopy (WCE) [<xref ref-type="bibr" rid="ref-11">11</xref>]. WCE is a small pill shaped device, consisting of batteries, a camera, and a light source [<xref ref-type="bibr" rid="ref-12">12</xref>]. While passing through the gastrointestinal tract (GIT), WCE captures almost 50,000 images and releases them through the anus. Mostly malignant diseases of GIT like bleeding, polyp, and ulcer are diagnosed through WCE. WCE is proved to be an authentic modality for painless investigation and examination of GIT [<xref ref-type="bibr" rid="ref-13">13</xref>]. This technique is more convenient to use than traditional endoscopies. Moreover, it provides better diagnostic accuracy for bleeding and tumor detection, specifically in the small intestine [<xref ref-type="bibr" rid="ref-14">14</xref>]. It is too difficult for an expert physician to thoroughly examine all of the captured images as it is a chaotic and time taking task. The manual analysis of WCE frames is not an easy process, and it takes much time [<xref ref-type="bibr" rid="ref-15">15</xref>]. To resolve this problem, researchers are working on introducing several computer-aided diagnosis (CAD) methods [<xref ref-type="bibr" rid="ref-16">16</xref>]. The suggested methods will help the doctors to detect the disease more accurately in less time. A typical CAD system consists of five major steps that are image pre-processing, feature extraction, optimal feature selection, feature fusion, and classification. The extraction of useful features is one of the most significant tasks. Numerous features are extracted in the CAD systems for accurate diagnosis of disease. These features include texture [<xref ref-type="bibr" rid="ref-17">17</xref>], color [<xref ref-type="bibr" rid="ref-18">18</xref>], and so on [<xref ref-type="bibr" rid="ref-19">19</xref>]. All of the extracted features are not useful; there may exist some irrelevant. Therefore, it is essential to reduce the feature vector to remove irrelevant features for better efficiency.</p>
<p>In this work, to detect and classify GI abnormalities, two pipeline procedures are considered: bleeding segmentation and GI abnormality classification. The significant contributions of the work are: i) in the bleeding detection step, a hybrid technique is implemented. In this method, the original image is split into three channels and thresholding is applied for each color channel. After that, pixel-by-pixel matching is performed, and a mask is generated for each channel. Finally, combining all mask images in one image for final bleeding segmentation; ii) in the classification procedure, transfer learning is ustilized for extracting deep learning features. The original images are enhanced by using a unified method, which is a combination of chrominance weight map, gamma correction, haze reduction, and YCbCr color conversion. Later, deep learning features are extracted by using a pre-trained CNN model. Further, the LBP features are also obtained for textural information of each abnormal regions; iii) a new method named entropy controlled ensemble learning is proposed and it selects the best learning features for correct prediction as well as fast execution. The selected features are ensemble in one vector by using a concatenation approach; iv) the performance of the proposed method is validated by several combinations of features. Further, many classification methods are also used for validation of selected features vector.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Related Work</title>
<p>Several machine learning and computer vision-based techniques are introduced for the diagnosis of human diseases like lung cancer, brain tumor, GIT infections from WCE images, and so on [<xref ref-type="bibr" rid="ref-20">20</xref>,<xref ref-type="bibr" rid="ref-21">21</xref>]. The stomach is one of the most significant organs of the human body. The most conspicuous diseases of the stomach are ulcer, bleeding, and polyp. In the study [<xref ref-type="bibr" rid="ref-22">22</xref>], the authors have utilized six features of different color spaces for the classification of non-ulcer and ulcerous regions. The used color spaces are CMYK, YUV, RGB, HSV, LAB, and XYZ. After the extraction of features, the cross-correlation method is utilized for the fusion of extracted features. Finally, 97.89% accuracy is obtained by utilizing the support vector machine (SVM) for the classification. Suman et al. [<xref ref-type="bibr" rid="ref-23">23</xref>] proposed a new method for the classification of bleeding images from non-bleeding ones. Their suggested method is mainly based on the statistical color features obtained from the RGB images. Charfi et al. [<xref ref-type="bibr" rid="ref-24">24</xref>] presented another methodology for colon irregularity detection from WCE images utilizing variance, LBP, and DWT features. They have been used a multilayer perceptron (MLP) and SVM for the classification. The suggested method performed well than existing methods and achieved 85.86% accuracy on linear SVM and 89.43% accuracy on MLP. In [<xref ref-type="bibr" rid="ref-25">25</xref>], authors have proposed a CAD method for bleeding classification. They have used unsupervised and supervised learning algorithms. Souaidi et al. [<xref ref-type="bibr" rid="ref-26">26</xref>] proposed a unique approach named multiscale complete LBP (MS-CLBP) for ulcer detection which is mainly based on the Laplacian pyramid and completed LBP (CLBP). In this method, ulcer detection is performed using two-color spaces (YCbCr and RGB). They have used the G channel of RGB and Cr of YCbCr for the detection of ulcers. Classification is performed using SVM and attained an average accuracy of 93.88% and 95.11% for both datasets. According to the survey conducted by Fan et al. [<xref ref-type="bibr" rid="ref-27">27</xref>], different pre-trained models of deep learning have covered numerous aspects of the medical imaging domain. Many researchers have utilized CNN models for the accurate classification and segmentation of disease or infections.</p>
<p>In contrast, the images that have the same category should share the same learned features. The overall achieved recognition accuracy of this method is 98%. Sharif et al. [<xref ref-type="bibr" rid="ref-28">28</xref>] have used the contrast-enhanced color features for the segmentation of the infected region from the image. After that, geometric features are pull-out from the resultant segmented portion of the image. Two deep CNN models VGG19 and VGG16, are also used in this method. Extracted deep features of both models are fused using the Euclidean fisher vector (EFV) that are later combined with the geometric characteristics to obtain strong features. Conditional entropy is employed on the resultant feature vector for optimal feature selection, which is classified using the KNN classifier and achieved the highest accuracy of 99.42%. Diamantis et al. [<xref ref-type="bibr" rid="ref-29">29</xref>] have proposed a novel method named as LB-FCN (Look Behind Fully Convolutional Neural Network). The proposed method has performed well than existing methods and achieved better GI abnormality detection results. Alaskar et al. [<xref ref-type="bibr" rid="ref-30">30</xref>] have utilized GoogleNet and AlexNet for the ulcer and non-ulcer images classification. Khan et al. [<xref ref-type="bibr" rid="ref-31">31</xref>] have suggested a technique for the ulcer detection and GI abnormality classification. ResNet101 is utilized by the authors for features extraction. Moreover, they optimized features by utilizing distance fitness function along with grasshopper optimization. They have utilized C-SVM and achieved 99.13% classification accuracy. The literature depicts that CAD systems for disease recognition mostly rely on handcrafted features (shape, color, and texture information). However, impressed by the performance of CNN in other domains, some researchers have employed CNN models in the medical field for disease segmentation and classification [<xref ref-type="bibr" rid="ref-32">32</xref>,<xref ref-type="bibr" rid="ref-33">33</xref>]. Inspired by these studies, we utilized deep learning for better classification accuracy.</p>
</sec>
<sec id="s3">
<label>3</label>
<title>Proposed Methodology</title>
<p>A hybrid architecture is proposed in this work for automated detection and classification of stomach abnormalities. The proposed architecture follows two pipeline procedures: Bleeding abnormality segmentation and GI infections classification. The proposed bleeding segmentation procedure is illustrated in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>. First, select the RGB bleeding images from the Database, then, extract all three channels and apply thresholding. Output images produced by a threshold function are compared by pixel-wise and used for generating a mask for each channel. Later, combined the mask of all three channels, as a result, a segmented image is obtained. The detail description of each step given as follows.</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>Flow diagram of proposed segmentation method</title>
</caption><graphic mimetype="image" mime-subtype="png" xlink:href="fig-1.png"/>
</fig>
<sec id="s3_1">
<label>3.1</label>
<title>Bleeding Abnormality Segmentation</title>
<p>Consider <inline-formula id="ieqn-1"><mml:math id="mml-ieqn-1"><mml:mi>X</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is an original RGB WCE image of dimension <inline-formula id="ieqn-2"><mml:math id="mml-ieqn-2"><mml:mn>256</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>256</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>3</mml:mn></mml:math></inline-formula>. Consider <inline-formula id="ieqn-3"><mml:math id="mml-ieqn-3"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-4"><mml:math id="mml-ieqn-4"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula id="ieqn-5"><mml:math id="mml-ieqn-5"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>B</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> denotes extracted channels, namely red channel, green channel, and blue channel, respectively, and illustrated in <xref ref-type="fig" rid="fig-2">Fig. 2</xref>. After splitting all channels, applying a threshold on each channel. The Otsu thresholding is used for each channel. Mathematically, the thresholding process is defined as follows.</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>Visualization of the original image and extracted channels: <inline-formula id="ieqn-6"><mml:math id="mml-ieqn-6"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mi>R</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-7"><mml:math id="mml-ieqn-7"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mi>G</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula id="ieqn-8"><mml:math id="mml-ieqn-8"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mi>B</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula></title>
</caption><graphic mimetype="image" mime-subtype="png" xlink:href="fig-2.png"/>
</fig>
<p>Considering three channels <inline-formula id="ieqn-9"><mml:math id="mml-ieqn-9"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-10"><mml:math id="mml-ieqn-10"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula id="ieqn-11"><mml:math id="mml-ieqn-11"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>B</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> having <inline-formula id="ieqn-12"><mml:math id="mml-ieqn-12"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> gray levels (<inline-formula id="ieqn-13"><mml:math id="mml-ieqn-13"><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x2026;</mml:mo></mml:math></inline-formula>), and maximum pixels are <inline-formula id="ieqn-14"><mml:math id="mml-ieqn-14"><mml:mstyle mathvariant="normal"><mml:mi>N</mml:mi></mml:mstyle><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mo>&#x2211;</mml:mo> </mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo lspace='0pt' rspace='0pt'>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mtext>&#x0394;</mml:mtext><mml:mo lspace='0pt' rspace='0pt'>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mi>&#x03BE;</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. Here, <inline-formula id="ieqn-15"><mml:math id="mml-ieqn-15"><mml:msub><mml:mrow><mml:mi>&#x03BE;</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is a specific grey-level pixel. The variance is using for segmentation process and define as it is a sum of the difference between two classes and denoted by <inline-formula id="ieqn-16"><mml:math id="mml-ieqn-16"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and can be expressed as:</p>
<p><disp-formula id="eqn-1">
<label>(1)</label>
<mml:math id="mml-eqn-1" display="block"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="script">M</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mi mathvariant="script">M</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="script">M</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mi mathvariant="script">M</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:math></disp-formula></p>
<p>where, <inline-formula id="ieqn-17"><mml:math id="mml-ieqn-17"><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-18"><mml:math id="mml-ieqn-18"><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> denotes pixels probabilities and it can be calculated through the following formulas:</p>
<p><disp-formula id="eqn-2">
<label>(2)</label>
<mml:math id="mml-eqn-2" display="block"><mml:mrow></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle='true'><mml:mstyle displaystyle='true'><mml:munderover><mml:mrow><mml:mo>&#x2211;</mml:mo> </mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo lspace='0pt' rspace='0pt'>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:munderover></mml:mstyle></mml:mstyle><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow></mml:mrow></mml:math>
</disp-formula></p>
<p><disp-formula id="eqn-3">
<label>(3)</label>
<mml:math id="mml-eqn-3" display="block"><mml:mrow></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>R</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:mspace width="1em"/><mml:mstyle mathvariant="italic"><mml:mi>w</mml:mi><mml:mi>h</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi></mml:mstyle><mml:mspace width=".3em" /><mml:mi>R</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mspace width="0.3em"/><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mrow></mml:mrow></mml:math>
</disp-formula></p>
<p>where, <inline-formula id="ieqn-19"><mml:math id="mml-ieqn-19"><mml:msub><mml:mrow><mml:mi mathvariant="script">M</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-20"><mml:math id="mml-ieqn-20"><mml:msub><mml:mrow><mml:mi mathvariant="script">M</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> denotes the mean value between the two classes. And can be calculated as:</p>
<p><disp-formula id="eqn-4">
<label>(4)</label>
<mml:math id="mml-eqn-4" display="block"><mml:mrow></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="script">M</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mstyle displaystyle='true'><mml:mstyle displaystyle='true'><mml:munderover><mml:mrow><mml:mo>&#x2211;</mml:mo> </mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo lspace='0pt' rspace='0pt'>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:munderover></mml:mstyle></mml:mstyle><mml:mi>j</mml:mi><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:mrow><mml:mrow></mml:mrow></mml:math>
</disp-formula></p>
<p><disp-formula id="eqn-5">
<label>(5)</label>
<mml:math id="mml-eqn-5" display="block"><mml:mrow></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="script">M</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mstyle displaystyle='true'><mml:mstyle displaystyle='true'><mml:munderover><mml:mrow><mml:mo>&#x2211;</mml:mo> </mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo lspace='0pt' rspace='0pt'>=</mml:mo><mml:mn>1</mml:mn><mml:mo lspace='0pt' rspace='0pt'>+</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mo lspace='0pt' rspace='0pt'>&#x0303;</mml:mo></mml:mover></mml:mrow></mml:munderover></mml:mstyle></mml:mstyle><mml:mi>j</mml:mi><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:mrow><mml:mrow></mml:mrow></mml:math>
</disp-formula></p>
<p>Bi-class variance can be expressed as:</p>
<p><disp-formula id="eqn-6">
<label>(6)</label>
<mml:math id="mml-eqn-6" display="block"><mml:mrow></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle='true'><mml:mstyle displaystyle='true'><mml:munderover><mml:mrow><mml:mo>&#x2211;</mml:mo> </mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo lspace='0pt' rspace='0pt'>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:munderover></mml:mstyle></mml:mstyle><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mspace width="0.3em"/><mml:mi>j</mml:mi><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="script">M</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mfrac><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mrow><mml:mrow></mml:mrow></mml:math>
</disp-formula></p>
<p><disp-formula id="eqn-7">
<label>(7)</label>
<mml:math id="mml-eqn-7" display="block"><mml:mrow></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle='true'><mml:mstyle displaystyle='true'><mml:munderover><mml:mrow><mml:mo>&#x2211;</mml:mo> </mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo lspace='0pt' rspace='0pt'>=</mml:mo><mml:mn>1</mml:mn><mml:mo lspace='0pt' rspace='0pt'>+</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mo lspace='0pt' rspace='0pt'>&#x0303;</mml:mo></mml:mover><mml:mo lspace='0pt' rspace='0pt'>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:munderover></mml:mstyle></mml:mstyle><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mspace width="0.3em"/><mml:mi>j</mml:mi><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="script">M</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mfrac><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mrow><mml:mrow></mml:mrow></mml:math>
</disp-formula></p>
<p>In Ostu segmentation, threshold selection is based on the cost function and can be calculated as:</p>
<p><disp-formula id="eqn-8">
<label>(8)</label>
<mml:math id="mml-eqn-8" display="block"><mml:mi>&#x03C6;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo> arg</mml:mo><mml:mn>0</mml:mn><mml:mo>&#x2264;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x003C;</mml:mo><mml:mtext>&#x0394;</mml:mtext><mml:mi>M</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:math></disp-formula></p>
<p>where, <inline-formula id="ieqn-21"><mml:math id="mml-ieqn-21"><mml:mi>&#x03C6;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is a thresholded image and <italic>i</italic> denotes three thresholded images for channels <inline-formula id="ieqn-22"><mml:math id="mml-ieqn-22"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-23"><mml:math id="mml-ieqn-23"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula id="ieqn-24"><mml:math id="mml-ieqn-24"><mml:msub><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>B</mml:mi></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:math></inline-formula> After that, each channel is compared with corresponding threshold channels. As a result, the mask is generated for each color channel. Later, it combines all channels information through the following equation. Consider <inline-formula id="ieqn-25"><mml:math id="mml-ieqn-25"><mml:mi>&#x03C6;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> denote a thresholded image of the red channel, <inline-formula id="ieqn-26"><mml:math id="mml-ieqn-26"><mml:mi>&#x03C6;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> denote a thresholded image of the green channel and <inline-formula id="ieqn-27"><mml:math id="mml-ieqn-27"><mml:mi>&#x03C6;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> denote a thresholded image of the blue channel, respectively.</p>
<p>Suppose, <inline-formula id="ieqn-28"><mml:math id="mml-ieqn-28"><mml:mi>&#x03C6;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-29"><mml:math id="mml-ieqn-29"><mml:mi>&#x03C6;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, and <inline-formula id="ieqn-30"><mml:math id="mml-ieqn-30"><mml:mi>&#x03C6;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> represent three threshold channels, as shown in <xref ref-type="fig" rid="fig-3">Fig. 3</xref>; then the fusion formula is defined as follows:</p>
<p><disp-formula id="eqn-9">
<label>(9)</label>
<mml:math id="mml-eqn-9" display="block"><mml:mrow></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:math>
</disp-formula></p>
<p><disp-formula id="eqn-10">
<label>(10)</label>
<mml:math id="mml-eqn-10" display="block"><mml:mrow></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2223;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow></mml:mrow></mml:math>
</disp-formula></p>
<disp-formula id="eqn-11">
<label>(11)</label>
<mml:math id="mml-eqn-11" display="block"><mml:mrow></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2223;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2223;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:math>
</disp-formula>
<disp-formula id="eqn-12">
<label>(12)</label>
<mml:math id="mml-eqn-12" display="block"><mml:mrow></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2223;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2223;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow></mml:mrow></mml:math>
</disp-formula>
<p>where <inline-formula id="ieqn-31"><mml:math id="mml-ieqn-31"><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2229;</mml:mo><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> denotes the final bleeding segmented image, and visible results are shown in <xref ref-type="fig" rid="fig-4">Fig. 4</xref>.</p>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>Extracted thresholding channels</title>
</caption><graphic mimetype="image" mime-subtype="png" xlink:href="fig-3.png"/>
</fig>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Abnormalities Classification</title>
<p>In the classification task, we presented a deep learning-based architecture for GI abnormalities classification such as bleeding, ulcer, and healthy. This task consists of three steps: enhancement of original images, deep feature extraction, and selection of robust features for classification. The flow diagram is presented in <xref ref-type="fig" rid="fig-5">Fig. 5</xref>. The mathematical description of the proposed classification task is given below.</p>
<fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>Segmentation output. (a) Original image, (b) detected red spots, (c) proposed binarized segmented image, (d) ground truth image</title>
</caption><graphic mimetype="image" mime-subtype="png" xlink:href="fig-4.png"/>
</fig>
<sec id="s3_2_1">
<label>3.2.1</label>
<title>Contrast Enhancement</title>
<p>WCE images may suffer from non-uniform lighting, low visibility, diminished colors, blurring, and low contrast of image characteristics [<xref ref-type="bibr" rid="ref-34">34</xref>]. In the very first step, we have applied a chromatic weight map to improve the saturation in the images. Thus, the color can be an essential indicator of image value. Let, <inline-formula id="ieqn-32"><mml:math id="mml-ieqn-32"><mml:mi>X</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> be the input image of dimension <inline-formula id="ieqn-33"><mml:math id="mml-ieqn-33"><mml:mn>256</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>256</mml:mn></mml:math></inline-formula>.  For each input image distance of each pixel saturation with most of the congestion, the series is calculated to get this weight map. Mathematically, the chromatic weight map is expressing as follows:</p>
<p><disp-formula id="eqn-13">
<label>(13)</label>
<mml:math id="mml-eqn-13" display="block"><mml:mrow><mml:msub><mml:mi>X</mml:mi><mml:mi>c</mml:mi></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>e</mml:mi><mml:mi>x</mml:mi><mml:mi>p</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>X</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:msup><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:msup><mml:mi>&#x03C3;</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:math></disp-formula></p>
<p>where, <inline-formula id="ieqn-34"><mml:math id="mml-ieqn-34"><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is the output image of the chromatic weight map, <inline-formula id="ieqn-35"><mml:math id="mml-ieqn-35"><mml:mi>&#x03C3;</mml:mi></mml:math></inline-formula> represents the standard deviation, and its default value is 0.3. The <italic>S<sub>max</sub></italic> denote is the highest value of the saturation range. Therefore the small amount of saturation is allocated to the pixels. After chromatic weight computation, gamma correction is applying to control the brightness and contrast ratio of an image <inline-formula id="ieqn-36"><mml:math id="mml-ieqn-36"><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>. Generally, the value of input and output ranges between 0 and 1. In the proposed preprocessing approach, we have used <inline-formula id="ieqn-37"><mml:math id="mml-ieqn-37"><mml:mi>&#x03B3;</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>4</mml:mn></mml:math></inline-formula> because, at this value, images are enhanced suitably without losing important information. Mathematically gamma correction is formulated as follows:</p>
<p><disp-formula id="eqn-14">
<label>(14)</label>
<mml:math id="mml-eqn-14" display="block"><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mspace width="0.3em"/><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>A</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mspace width="0.3em"/><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>&#x03B3;</mml:mi></mml:mrow></mml:msup></mml:math></disp-formula></p>
<p>where, <italic>A</italic> is a constant, <inline-formula id="ieqn-38"><mml:math id="mml-ieqn-38"><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> is the positive real value of chromatic weight map image, raised to the power of <inline-formula id="ieqn-39"><mml:math id="mml-ieqn-39"><mml:mi>&#x03B3;</mml:mi></mml:math></inline-formula> multiplied by a constant value <italic>A</italic> (generally <italic>A</italic> = 1). Input and output range typically from 0 to 1. Further, we are applying image dehazing. The main objective of Image dehazing is to recover the flawless image from a blurry image. However, it may be deficient in capturing the essential features of hazy images fully. Mathematically image dehazing is formulated as follows:</p>
<p><disp-formula id="eqn-15">
<label>(15)</label>
<mml:math id="mml-eqn-15" display="block"><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>H</mml:mi><mml:mi>Z</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mspace width="0.3em"/><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mstyle mathvariant="normal"><mml:mi>m</mml:mi></mml:mstyle><mml:mo>,</mml:mo><mml:mspace width="0.3em"/><mml:mstyle mathvariant="normal"><mml:mi>n</mml:mi></mml:mstyle></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mi>A</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mspace width="0.3em"/><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>t</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mspace width="0.3em"/><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mspace width="0.3em"/><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
<p>The foundation of a hazy image can be presented as a curved combination of the image radiance <italic>J</italic> and the light <italic>A</italic>, which can be formulated as where <inline-formula id="ieqn-40"><mml:math id="mml-ieqn-40"><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mstyle mathvariant="normal"><mml:mi>m</mml:mi></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="normal"><mml:mi>n</mml:mi></mml:mstyle></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is a pixel from the hazy image <italic>X<sub>GM</sub></italic> and <inline-formula id="ieqn-41"><mml:math id="mml-ieqn-41"><mml:mi>t</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is its transmission. As a consequence, by recovering the image radiance <inline-formula id="ieqn-42"><mml:math id="mml-ieqn-42"><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> from hazy pixel <inline-formula id="ieqn-43"><mml:math id="mml-ieqn-43"><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>m</mml:mi><mml:mo>,</mml:mo><mml:mi>n</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> the problem of image hazing is resolved. The visual effects of this formulation are illustrated in <xref ref-type="fig" rid="fig-6">Fig. 6</xref>. Finally, the YCbCr color transformations are applying and select the Y channel based on peak pixel values. This channel is further utilized for features extraction task.</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>Flow diagram of proposed classification method</title>
</caption><graphic mimetype="image" mime-subtype="png" xlink:href="fig-5.png"/>
</fig>
<fig id="fig-6">
<label>Figure 6</label>
<caption>
<title>(a) Output images of gamma correction, (b) output images of haze reduction 
 
</title>
</caption><graphic mimetype="image" mime-subtype="png" xlink:href="fig-6.png"/>
</fig>
</sec>
<sec id="s3_2_2">
<label>3.2.2</label>
<title>Features Extraction (FE)</title>
<p><bold>Deep Learning Features:</bold> For deep learning features, we have utilized a pre-trained inceptionV3 model. This model is trained on an enhanced WCE dataset. InceptionV3 model consists of 316 layers and <inline-formula id="ieqn-44"><mml:math id="mml-ieqn-44"><mml:mn>350</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula> connections. <xref ref-type="fig" rid="fig-7">Fig. 7</xref> gives the generalized Visual representation of the layers of inceptionV3 model. The input size for this model is <inline-formula id="ieqn-45"><mml:math id="mml-ieqn-45"><mml:mn>299</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>299</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula>; the output of the first Inceptionv3 layer is to the next layer. It consists of a series of layers, such as scaling layer, AveragePooling layer, convolutional layers, Depth Concatenation layer, ReLU layer, Batch normalization layer, MaxPooling layer, softmax layer, and a classification layer. Each layer in the inceptionV3 model has several parameters.</p>
<fig id="fig-7">
<label>Figure 7</label>
<caption>
<title>Detailed model of inceptionV3 
 
</title>
</caption><graphic mimetype="image" mime-subtype="png" xlink:href="fig-7.png"/>
</fig>
<p>Using this model, the ratio that we have used for training and testing is 70:30. This feature extraction process is conducted using the transfer learning method. In the TL, a pre-trained model was trained on GI WCE images. For this purpose, we required an input layer and an output layer. For input layer, we are using the first convolution layer, while in the output; we select the average pool layer. We have obtained a feature vector of dimension <inline-formula id="ieqn-46"><mml:math id="mml-ieqn-46"><mml:mn>1</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>2048</mml:mn></mml:math></inline-formula>, after activation of the average pool layer. For <italic>N</italic> images, the vector should be <inline-formula id="ieqn-47"><mml:math id="mml-ieqn-47"><mml:mi>N</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mn>2048</mml:mn></mml:math></inline-formula>.</p>
<p><bold>Texture Oriented Features&#x2013;-</bold>For the sake of texture oriented features, we are extracting Local Binary Patterns (LBP). LBP is a significant method used for object identification and detection [<xref ref-type="bibr" rid="ref-35">35</xref>]. Basically, LBP features consist of two bitwise transitions from 0 to 1 and 1 to 0. LBP use a greyscale image as input, then calculate mean and variance for each pixel intensity. Mathematical representation of LBP is formulated as follow:</p>
<p><disp-formula id="eqn-16">
<label>(16)</label>
<mml:math id="mml-eqn-16" display="block"><mml:mi>L</mml:mi><mml:mi>B</mml:mi><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>P</mml:mi><mml:mo>,</mml:mo><mml:mspace width="0.3em"/><mml:mi>R</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle='true'><mml:mstyle displaystyle='true'><mml:munderover><mml:mrow><mml:mo>&#x2211;</mml:mo> </mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mo lspace='0pt' rspace='0pt'>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mo lspace='0pt' rspace='0pt'>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:munderover></mml:mstyle></mml:mstyle><mml:mi>S</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mi>P</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mi>C</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mn>2</mml:mn><mml:mstyle mathvariant="normal"><mml:mi>P</mml:mi></mml:mstyle></mml:math></disp-formula></p>
<p>Here, the number of neighborhood intensities are represented by<italic>P</italic>, <italic>R</italic> is the radius, <italic>k<sub>p</sub></italic> is the variance of the neighboring pixel intensity, <italic>KC</italic> represents the contrast of intensity calculated from <inline-formula id="ieqn-48"><mml:math id="mml-ieqn-48"><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>P</mml:mi><mml:mo>,</mml:mo><mml:mi>R</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>.</p>
<p><disp-formula id="eqn-17">
<label>(17)</label>
<mml:math id="mml-eqn-17" display="block"><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mtable equalrows="false" columnlines="none" equalcolumns="false"><mml:mtr><mml:mtd columnalign="left"><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mspace width="1em"/></mml:mtd><mml:mtd columnalign="left"><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mspace width=".3em" /><mml:mi>P</mml:mi><mml:mo>&#x2265;</mml:mo><mml:mi>t</mml:mi></mml:mtd></mml:mtr><mml:mtr><mml:mtd columnalign="left"><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mspace width="1em"/></mml:mtd><mml:mtd columnalign="left"><mml:mstyle mathvariant="italic"><mml:mi>O</mml:mi><mml:mi>t</mml:mi><mml:mi>h</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mi>w</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mstyle></mml:mtd></mml:mtr> </mml:mtable></mml:mrow><mml:mo></mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>where neighboring pixels <italic>d<sub>n</sub></italic>(<italic>P</italic>) are compared with the central pixel <italic>t</italic>. It obtains a feature vector of <inline-formula id="ieqn-49"><mml:math id="mml-ieqn-49"><mml:mn>1</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>59</mml:mn></mml:math></inline-formula> dimension for one image and <inline-formula id="ieqn-50"><mml:math id="mml-ieqn-50"><mml:mstyle mathvariant="normal"><mml:mi>N</mml:mi></mml:mstyle><mml:mo>&#x00D7;</mml:mo><mml:mn>59</mml:mn></mml:math></inline-formula> for <inline-formula id="ieqn-51"><mml:math id="mml-ieqn-51"><mml:mstyle mathvariant="normal"><mml:mi>N</mml:mi></mml:mstyle></mml:math></inline-formula> number of images.</p>
</sec>
<sec id="s3_2_3">
<label>3.2.3</label>
<title>Features Selection</title>
<p>After the extraction of texture and deep learning features, the next phase involves the optimal features selection. In this work, we have utilized Shannon entropy along with an ensemble learning classifier for best features selection. A heuristic approach has opted for feature selection. The Shannon Entropy is computed from both vectors separately and set a target function based on the mean value of original entropy vectors. The features that are equal or higher than mean features are selects as robust features and passed to ensemble classifiers. However, this process is to continue until the error rate of the ensemble classifier is below 0.1. Mathematically Shannon entropy is ratified by the equation as follow:</p>
<p>Let <inline-formula id="ieqn-52"><mml:math id="mml-ieqn-52"><mml:msubsup><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> be the number of occurrences of <italic>t<sub>j</sub></italic> in the category <italic>c<sub>k</sub></italic> and, <inline-formula id="ieqn-53"><mml:math id="mml-ieqn-53"><mml:mi>t</mml:mi><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> the frequency of <italic>t<sub>j</sub></italic> in this category <italic>c<sub>k</sub></italic>:</p>
<p><disp-formula id="eqn-18">
<label>(18)</label>
<mml:math id="mml-eqn-18" display="block"><mml:mi>t</mml:mi><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msubsup><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mo>&#x2211;</mml:mo> </mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msubsup><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
<p>The Shannon entropy <italic>E</italic>(<italic>t<sub>j</sub></italic>) of the term <italic>t<sub>j</sub></italic> is given by:</p>
<p><disp-formula id="eqn-19">
<label>(19)</label>
<mml:math id="mml-eqn-19" display="block"><mml:mstyle><mml:mtext>E</mml:mtext></mml:mstyle><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mstyle displaystyle='true'><mml:mstyle displaystyle='true'><mml:munderover><mml:mrow><mml:mo>&#x2211;</mml:mo> </mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo lspace='0pt' rspace='0pt'>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:munderover></mml:mstyle></mml:mstyle><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mo>log</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>Through this process, approximately 50% of features are removed from both vectors-deep learning and texture oriented. Later on, these selected vectors are fused in one vector by simple concatenation approach as given as: Let <italic>A</italic> and <italic>B</italic> be two feature spaces that are defined on the sample space <inline-formula id="ieqn-54"><mml:math id="mml-ieqn-54"><mml:mi>&#x03A9;</mml:mi></mml:math></inline-formula>, where <italic>A</italic> represents selected LBP features and <italic>B</italic> represents selected inceptionV3 features. An arbitrary sample <inline-formula id="ieqn-55"><mml:math id="mml-ieqn-55"><mml:mi>&#x03BE;</mml:mi></mml:math></inline-formula> is selected from the sample space such that <inline-formula id="ieqn-56"><mml:math id="mml-ieqn-56"><mml:mi>&#x03BE;</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mi>&#x03A9;</mml:mi></mml:math></inline-formula>, the corresponding two FVs are <inline-formula id="ieqn-57"><mml:math id="mml-ieqn-57"><mml:msub><mml:mrow><mml:mi>&#x03B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mi>b</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2208;</mml:mo><mml:mi>A</mml:mi></mml:math></inline-formula> and <inline-formula id="ieqn-58"><mml:math id="mml-ieqn-58"><mml:msub><mml:mrow><mml:mi>&#x03B2;</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="italic"><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:mi>e</mml:mi><mml:mi>p</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mo>&#x2208;</mml:mo><mml:mi>B</mml:mi></mml:math></inline-formula>. Then, the fusion of feature <inline-formula id="ieqn-59"><mml:math id="mml-ieqn-59"><mml:mi>&#x03BE;</mml:mi></mml:math></inline-formula> is defined by <inline-formula id="ieqn-60"><mml:math id="mml-ieqn-60"><mml:mi>&#x03B3;</mml:mi><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x03B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mi>b</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>&#x03B2;</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="italic"><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:mi>e</mml:mi><mml:mi>p</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula>, if <inline-formula id="ieqn-61"><mml:math id="mml-ieqn-61"><mml:msub><mml:mrow><mml:mi>&#x03B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mi>b</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is <italic>n</italic> dimensional and <inline-formula id="ieqn-62"><mml:math id="mml-ieqn-62"><mml:msub><mml:mrow><mml:mi>&#x03B2;</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="italic"><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:mi>e</mml:mi><mml:mi>p</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula> is <italic>m</italic> dimensional then the fused FV <inline-formula id="ieqn-63"><mml:math id="mml-ieqn-63"><mml:mi>&#x03B3;</mml:mi></mml:math></inline-formula> is <inline-formula id="ieqn-64"><mml:math id="mml-ieqn-64"><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>+</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>-dimensional. All combined feature FVs form a <inline-formula id="ieqn-65"><mml:math id="mml-ieqn-65"><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>+</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>-dimensional feature space.</p>
</sec>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Experimental Results and Comparison</title>
<p>Two datasets are used in this work for the assessment of suggested GI infections detection and classification method. The description of each dataset is given as:</p>
<p><bold>KVASIR Dataset</bold> contains a total of 4000 images, which are confirmed by expert doctors [<xref ref-type="bibr" rid="ref-36">36</xref>]. Eight different types of infections are included in this dataset, such as Dyed-Lifted-Polyp (DLP), Dyed-Resection-Margin (DRM), Esophagitis (ESO), Normal-Cecum (NC), Normal-Pylorus (NP), Normal Z-Line (NZL), Polyps (P), and Ulcerative-Colities (UCE). Every class contains 500 images of different resolution-<inline-formula id="ieqn-66"><mml:math id="mml-ieqn-66"><mml:mn>720</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>576</mml:mn></mml:math></inline-formula> up to <inline-formula id="ieqn-67"><mml:math id="mml-ieqn-67"><mml:mn>1920</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>1072</mml:mn></mml:math></inline-formula> pixels. Sample images are shown in <xref ref-type="fig" rid="fig-8">Fig. 8a</xref>.</p>
<p><bold>Private Dataset</bold> was collected from COMSATS Computer Vision Lab [<xref ref-type="bibr" rid="ref-37">37</xref>], and it includes a total of 2326 clinical sample images. These images consist of three categories-ulcer, bleeding, and healthy. The image size is <inline-formula id="ieqn-68"><mml:math id="mml-ieqn-68"><mml:mn>512</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>512</mml:mn></mml:math></inline-formula>. Some sample images are presented in <xref ref-type="fig" rid="fig-8">Fig. 8b</xref>.</p>
<fig id="fig-8">
<label>Figure 8</label>
<caption>
<title>Sample images selected from the datasets: (a) KVASIR dataset [<xref ref-type="bibr" rid="ref-36">36</xref>], (b) Private dataset [<xref ref-type="bibr" rid="ref-37">37</xref>]</title>
</caption><graphic mimetype="image" mime-subtype="png" xlink:href="fig-8.png"/>
</fig>
<sec id="s4_1">
<label>4.1</label>
<title>Results</title>
<p>A detailed description of classification results in quantitative and graphical form is given in this section. For experimental results, seven different classifiers are used for the evaluation of suggested methods that are Linear Discriminant (L-Disc), Fine Tree (F-Tree), Cubic SVM (C-SVM), Medium Gaussian SVM (M-G-SVM), Linear SVM (L-SVM), Fine KNN (F-KNN), and ENSEM subspace discriminant (ESDA). In this research, we have utilized different performance measures for the evaluation of suggested methods, including Specificity (SPE), FNR, Precision (PRE), FPR, Sensitivity (SEN), Accuracy (ACC), F1-Score, Jaccard index, and Dice. All the tests are implemented on MATLAB 2019b using Core i5-7thGen, 4 GB RAM. Further, an 8 GB graphics card is also used for the evaluation of results.</p>
<p><bold>Bleeding Segmentation Results</bold>: To validate segmentation accuracy, 20 sample images from the Private dataset are randomly chosen for bleeding segmentation. In this research we have utilized the ground truth images designed by an expert physician for the segmentation accuracy calculation. The ground truth images are compared with the proposed segmented images pixel-by-pixel. Our proposed bleeding method has achieved the highest accuracy of 93.39%. Other calculated measures are Jack-index (96.58%) and FNR (6.61). The selected image&#x2019;s accuracy is given in <xref ref-type="table" rid="table-1">Tab. 1</xref>. In this table, it is shown that for all selected 20 images, Jack-Index, Dice, FNR, and Time (s) is presented. The average dice rate is 87.59%, which is good for bleeding segmentation.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>Bleeding segmentation results</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Image #</th>
<th colspan="4">Performance measures</th>
</tr>
<tr>
<th></th>
<th>Jack-index (%)</th>
<th>Dice (%)</th>
<th>FNR (%)</th>
<th>Time (s)</th>
</tr>
</thead>
<tbody>
<tr>
<td>1</td>
<td>95.32</td>
<td>91.05</td>
<td>8.95</td>
<td>5.865451</td>
</tr>
<tr>
<td>2</td>
<td>93.74</td>
<td>88.21</td>
<td>11.79</td>
<td>6.481893</td>
</tr>
<tr>
<td>3</td>
<td>88.25</td>
<td>78.98</td>
<td>21.02</td>
<td>6.109725</td>
</tr>
<tr>
<td>4</td>
<td>91.46</td>
<td>84.27</td>
<td>15.73</td>
<td>14.66179</td>
</tr>
<tr>
<td>5</td>
<td>95.02</td>
<td>90.51</td>
<td>9.49</td>
<td>5.103764</td>
</tr>
<tr>
<td>6</td>
<td>96.23</td>
<td>92.73</td>
<td>7.27</td>
<td>8.664483</td>
</tr>
<tr>
<td>7</td>
<td><bold>96.58</bold></td>
<td><bold>93.39</bold></td>
<td><bold>6.61</bold></td>
<td><bold>4.845475</bold></td>
</tr>
<tr>
<td>8</td>
<td>90.81</td>
<td>83.17</td>
<td>16.83</td>
<td>6.732851</td>
</tr>
<tr>
<td>9</td>
<td>94.34</td>
<td>89.29</td>
<td>5.66</td>
<td>9.608909</td>
</tr>
<tr>
<td>10</td>
<td>88.51</td>
<td>79.38</td>
<td>20.62</td>
<td>9.828089</td>
</tr>
<tr>
<td>11</td>
<td>93.66</td>
<td>88.08</td>
<td>11.92</td>
<td>9.818192</td>
</tr>
<tr>
<td>12</td>
<td>94.98</td>
<td>90.44</td>
<td>9.56</td>
<td>10.99683</td>
</tr>
<tr>
<td>13</td>
<td>93.62</td>
<td>88.01</td>
<td>11.99</td>
<td>8.654121</td>
</tr>
<tr>
<td>14</td>
<td>96.23</td>
<td>92.74</td>
<td>7.26</td>
<td>17.56766</td>
</tr>
<tr>
<td>15</td>
<td>90.38</td>
<td>82.44</td>
<td>17.56</td>
<td>13.99112</td>
</tr>
<tr>
<td>16</td>
<td>95.63</td>
<td>91.62</td>
<td>8.38</td>
<td>13.67921</td>
</tr>
<tr>
<td>17</td>
<td>93.46</td>
<td>87.72</td>
<td>12.28</td>
<td>7.435595</td>
</tr>
<tr>
<td>18</td>
<td>94.71</td>
<td>89.94</td>
<td>10.06</td>
<td>11.13178</td>
</tr>
<tr>
<td>19</td>
<td>95.25</td>
<td>90.94</td>
<td>9.06</td>
<td>8.380193</td>
</tr>
<tr>
<td>20</td>
<td>88.25</td>
<td>78.98</td>
<td>11.75</td>
<td>5.927833</td>
</tr>
</tbody>
</table>

</table-wrap>
<p><bold>Classification Results:</bold> For classification results, we have performed experiments for both selection and fusion processes separately. As mentioned above, two datasets named KVASIR and Private datasets are used for the evaluation of the proposed method. As the Private Dataset includes a total of 2326 RGB images consisting of three classes, namely ulcer, bleeding, and healthy. Initially, robust deep learning features are selected using the proposed selected approach. For robust deep learning features, maximum achieved accuracy of 99.7% on ESDA classifier, as numerical results are given in <xref ref-type="table" rid="table-2">Tab. 2</xref>. In this table, it is observed that the other calculated measures are SEN (99.7%), SPE (99.88%), PRE (99.70%), F1 score (99.70%), FPR (zero), and FNR (0.3%). Further, the accuracy of ESDA can also be validated by <xref ref-type="table" rid="table-3">Tab. 3</xref>. Besides, the execution time of each classifier is noted during the testing, as given in <xref ref-type="table" rid="table-2">Tab. 2</xref>. In this table, the best noted time is 16.818 s for the ESDA classifier. The worst execution time is 26.166 s for the M-G-SVM classifier.</p>
<table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>Propose classification results on selected deep learning features for private dataset. The maximum accuracy is bold in the table</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Classifier</th>
<th colspan="7">Performance measures</th>
<th></th>
</tr>
<tr>
<th></th>
<th>ACC (%)</th>
<th>SEN (%)</th>
<th>SPE (%)</th>
<th>PRE (%)</th>
<th>F1 score (%)</th>
<th>FPR</th>
<th>FNR (%)</th>
<th>Execution time (s)</th>
</tr>
</thead>
<tbody>
<tr>
<td>F-Tree</td>
<td>88.1</td>
<td>86.83</td>
<td>94.19</td>
<td>86.32</td>
<td>86.55</td>
<td>0.06</td>
<td>11.9</td>
<td>22.82</td>
</tr>
<tr>
<td>L-Disc</td>
<td>99.4</td>
<td>99.30</td>
<td>99.68</td>
<td>99.39</td>
<td>99.34</td>
<td>0.00</td>
<td>0.6</td>
<td>25.88</td>
</tr>
<tr>
<td>L-SVM</td>
<td>98.8</td>
<td>98.57</td>
<td>99.34</td>
<td>98.86</td>
<td>98.72</td>
<td>0.01</td>
<td>1.2</td>
<td>18.14</td>
</tr>
<tr>
<td>C-SVM</td>
<td>99.4</td>
<td>99.30</td>
<td>99.68</td>
<td>99.39</td>
<td>99.34</td>
<td>0.00</td>
<td>0.6</td>
<td>18.01</td>
</tr>
<tr>
<td>M-G-SVM</td>
<td>99.5</td>
<td>99.39</td>
<td>99.70</td>
<td>99.55</td>
<td>99.47</td>
<td>0.00</td>
<td>0.5</td>
<td>26.16</td>
</tr>
<tr>
<td>F-KNN</td>
<td>99.4</td>
<td>99.28</td>
<td>99.72</td>
<td>99.39</td>
<td>99.33</td>
<td>0.00</td>
<td>0.6</td>
<td>111.2</td>
</tr>
<tr>
<td>ESDA</td>
<td><bold>99.7</bold></td>
<td><bold>99.70</bold></td>
<td><bold>99.88</bold></td>
<td><bold>99.70</bold></td>
<td><bold>99.70</bold></td>
<td>0.00</td>
<td><bold>0.3</bold></td>
<td><bold>16.81</bold></td>
</tr>
</tbody>
</table>

</table-wrap>
 
<table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>Confusion matrix using selected inceptionV3 features for private dataset</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Classes</th>
<th colspan="3">Classification classes</th>
</tr>
<tr>
<th></th>
<th>Bleeding</th>
<th>Healthy</th>
<th>Ulcer</th>
</tr>
</thead>
<tbody>
<tr>
<td>Bleeding</td>
<td><inline-formula id="ieqn-69"><mml:math id="mml-ieqn-69"><mml:mo>&#x003E;</mml:mo></mml:math></inline-formula><bold>99%</bold></td>
<td>&#x003C; 1%</td>
</tr>
<tr>
<td>Healthy</td>
<td/>
<td><inline-formula id="ieqn-70"><mml:math id="mml-ieqn-70"><mml:mo>&#x003E;</mml:mo></mml:math></inline-formula><bold>99%</bold></td>
<td>&#x003C; 1%</td>
</tr>
<tr>
<td>Ulcer</td>
<td/>
<td>1%</td>
<td><bold>99%</bold></td>
</tr>
</tbody>
</table>

</table-wrap>
<p><xref ref-type="table" rid="table-4">Tab. 4</xref> shows the classification results using robust deep learning features for KVASIR dataset. The 10-fold cross-validation was used for testing results and achieved the best accuracy of 86.6% on the ESDA classifier, as numerical results can be seen in <xref ref-type="table" rid="table-4">Tab. 4</xref>. In this table, it is observed that the other calculated measures of ESDA classifier are SEN (86.62%), SPE (98.08%), PRE (87.08%), F1 score (86.60%), FPR (0.04) and FNR (13.4%). The accuracy of this classifier can also be validated by <xref ref-type="table" rid="table-5">Tab. 5</xref>. The minimum noted time is 27.67 s for F-Tree, while 43.01 s for ESDA classifier. However, the accuracy of the ESDA classifier is better as compared to all other methods.</p>
<table-wrap id="table-4">
<label>Table 4</label>
<caption>
<title>Propose classification results on selected deep learning features for KVASIR dataset. The maximum accuracy is bold in the table</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Classifier</th>
<th colspan="7">Performance measures</th>
<th></th>
</tr>
<tr>
<th></th>
<th>ACC (%)</th>
<th>SEN (%)</th>
<th>SPE (%)</th>
<th>PRE (%)</th>
<th>F1 Score (%)</th>
<th>FPR</th>
<th>FNR (%)</th>
<th>Execution time (s)</th>
</tr>
</thead>
<tbody>
<tr>
<td>F-Tree</td>
<td>56.3</td>
<td>56.33</td>
<td>93.76</td>
<td>56.22</td>
<td>56.24</td>
<td>0.06</td>
<td>43.7</td>
<td><bold>27.67</bold></td>
</tr>
<tr>
<td>L-Disc</td>
<td>84.9</td>
<td>84.93</td>
<td>97.85</td>
<td>85.26</td>
<td>85.05</td>
<td>0.02</td>
<td>15.1</td>
<td>27.84</td>
</tr>
<tr>
<td>L-SVM</td>
<td>85.7</td>
<td>85.65</td>
<td>97.95</td>
<td>85.63</td>
<td>85.59</td>
<td>0.02</td>
<td>14.3</td>
<td>46.46</td>
</tr>
<tr>
<td>C-SVM</td>
<td>85.8</td>
<td>85.23</td>
<td>97.93</td>
<td>85.23</td>
<td>85.22</td>
<td>0.02</td>
<td>14.2</td>
<td>53.72</td>
</tr>
<tr>
<td>M-G-SVM</td>
<td>84.7</td>
<td>84.70</td>
<td>97.81</td>
<td>84.82</td>
<td>84.67</td>
<td>0.02</td>
<td>15.3</td>
<td>29.36</td>
</tr>
<tr>
<td>F-KNN</td>
<td>74.9</td>
<td>72.18</td>
<td>96.03</td>
<td>73.09</td>
<td>72.12</td>
<td>0.04</td>
<td>25.1</td>
<td>50.30</td>
</tr>
<tr>
<td>ESDA</td>
<td><bold>86.6</bold></td>
<td><bold>86.62</bold></td>
<td><bold>98.08</bold></td>
<td><bold>87.08</bold></td>
<td><bold>86.80</bold></td>
<td>0.02</td>
<td><bold>13.4</bold></td>
<td>43.01</td>
</tr>
</tbody>
</table>

</table-wrap>
 
<table-wrap id="table-5">
<label>Table 5</label>
<caption>
<title>Confusion matrix using selected inception features for KVASIR dataset</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Classes</th>
<th colspan="8">Classification classes</th>
</tr>
<tr>
<th></th>
<th>DLP</th>
<th>DRM</th>
<th>ESO</th>
<th>NC</th>
<th>NP</th>
<th>NZL</th>
<th>P</th>
<th>UCE</th>
</tr>
</thead>
<tbody>
<tr>
<td>DLP</td>
<td><bold>77%</bold></td>
<td>19%</td>
<td/>
<td/>
<td/>
<td/>
<td>3%</td>
<td>1%</td>
</tr>
<tr>
<td>DRM</td>
<td>20%</td>
<td><bold>77%</bold></td>
<td/>
<td>&#x003C; 1%</td>
<td/>
<td/>
<td>1%</td>
<td>&#x003C; 1%</td>
</tr>
<tr>
<td>ESO</td>
<td/>
<td/>
<td><bold>77%</bold></td>
<td/>
<td>1%</td>
<td>22%</td>
<td>&#x003C; 1%</td>
<td>&#x003C; 1%</td>
</tr>
<tr>
<td>NC</td>
<td/>
<td/>
<td/>
<td><bold>97%</bold></td>
<td/>
<td/>
<td>2%</td>
<td>1%</td>
</tr>
<tr>
<td>NP</td>
<td/>
<td/>
<td>1%</td>
<td/>
<td><bold>98%</bold></td>
<td>1%</td>
<td/>
<td/>
</tr>
<tr>
<td>NZL</td>
<td/>
<td/>
<td>13%</td>
<td/>
<td>1%</td>
<td><bold>85%</bold></td>
<td>1%</td>
<td/>
</tr>
<tr>
<td>P</td>
<td/>
<td/>
<td>1%</td>
<td>4%</td>
<td>1%</td>
<td>&#x003C; 1%</td>
<td><bold>88%</bold></td>
<td>5%</td>
</tr>
<tr>
<td>UCE</td>
<td/>
<td/>
<td>1%</td>
<td>1%</td>
<td>1%</td>
<td/>
<td>3%</td>
<td><bold>94%</bold></td>
</tr>
</tbody>
</table>

</table-wrap>
<p>Later on, the selected deep learning and LBP features are fused and performed evaluation. Results are presented in <xref ref-type="table" rid="table-6">Tab. 6</xref> using a fused vector for Private Dataset. The maximum achieved accuracy after the fusion process is reached to 99.8% for the ESDA classifier. In this table, it is observed that the ESDA classifier shows better performance in terms of other performance measures as compared to other classifiers. Other calculated measures of ESDA are SEN (99.83%), SPE (99.92%), PRE (99.80%), F1 score (99.81%), FPR (zero), and FNR (0.2%). The accuracy of this classifier can also be validated by <xref ref-type="table" rid="table-7">Tab. 7</xref>. From the comparison with <xref ref-type="table" rid="table-2">Tab. 2</xref>, it is shown that after the fusion process, the results are a little bit improved. But on the other side, the computational time is increased up to an average of 4 to 5 sec. The best noted time for this experiment is 17.03 s, while the worst time is 42.39 s. In conclusion, ESDA performs better as compared to other methods.</p>
<table-wrap id="table-6">
<label>Table 6</label>
<caption>
<title>Classification results of proposed classification architecture for private dataset. The maximum accuracy is bold in the table</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Classifier</th>
<th colspan="7">Performance measures</th>
<th></th>
</tr>
<tr>
<th></th>
<th>ACC (%)</th>
<th>SEN (%)</th>
<th>SPE (%)</th>
<th>PRE (%)</th>
<th>F1 score (%)</th>
<th>FPR</th>
<th>FNR (%)</th>
<th>Execution time (s)</th>
</tr>
</thead>
<tbody>
<tr>
<td>F-Tree</td>
<td>87.7</td>
<td>86.23</td>
<td>94.02</td>
<td>85.85</td>
<td>86.03</td>
<td>0.06</td>
<td>12.3</td>
<td>27.373</td>
</tr>
<tr>
<td>L-Disc</td>
<td>99.5</td>
<td>99.47</td>
<td>99.75</td>
<td>99.52</td>
<td>99.49</td>
<td>0.00</td>
<td>0.5</td>
<td>19.944</td>
</tr>
<tr>
<td>L-SVM</td>
<td>98.8</td>
<td>98.65</td>
<td>99.36</td>
<td>98.89</td>
<td>98.77</td>
<td>0.01</td>
<td>1.2</td>
<td>20.426</td>
</tr>
<tr>
<td>C-SVM</td>
<td>99.6</td>
<td>99.65</td>
<td>99.69</td>
<td>99.67</td>
<td>99.66</td>
<td>0.00</td>
<td><bold>0.4</bold></td>
<td>18.217</td>
</tr>
<tr>
<td>M-G-SVM</td>
<td>99.4</td>
<td>99.34</td>
<td>99.68</td>
<td>99.52</td>
<td>99.43</td>
<td>0.00</td>
<td>0.6</td>
<td>25.439</td>
</tr>
<tr>
<td>F-KNN</td>
<td>99.6</td>
<td>99.50</td>
<td>99.81</td>
<td>99.58</td>
<td>99.54</td>
<td>0.00</td>
<td>0.4</td>
<td>42.39</td>
</tr>
<tr>
<td>ESDA</td>
<td><bold>99.8</bold></td>
<td><bold>99.83</bold></td>
<td><bold>99.92</bold></td>
<td><bold>99.80</bold></td>
<td><bold>99.81</bold></td>
<td>0.00</td>
<td><bold>0.2</bold></td>
<td><bold>17.031</bold></td>
</tr>
</tbody>
</table>

</table-wrap>
<p>The fused vector is also applied to the KVASIR dataset and achieved maximum accuracy of 87.8% for the ESDA classifier. <xref ref-type="table" rid="table-8">Tab. 8</xref> shows the results of the proposed architecture. The other calculated measures of this classifier are SEN (86.4%), SPE (98.06%), PRE (86.99%), F1 score (86.63%), FPR (0.02), and FNR (13.6%). Moreover, <xref ref-type="table" rid="table-9">Tab. 9</xref> shows a confusion matrix of this classifier, which can be used as the authenticity of proposed ESDA accuracy. The diagonal values show the correct prediction of each abnormality while the rest of the values are FNR. After fusion process accuracy is improved. However, a little bit of increase occurs in the execution time. After the fusion process, the best noted time is 26.06 s, while for ESDA classifier is 50.70 s. Overall, it can be easily concluded that the ESDA classifier shows better performance.</p>
<table-wrap id="table-7">
<label>Table 7</label>
<caption>
<title>Confusion matrix for proposed classification architecture for private dataset</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Classes</th>
<th colspan="3">Classification classes</th>
</tr>
<tr>
<th></th>
<th>Bleeding</th>
<th>Healthy</th>
<th>Ulcer</th>
</tr>
</thead>
<tbody>
<tr>
<td>Bleeding</td>
<td><bold>100%</bold></td>
<td/>
<td/>
</tr>
<tr>
<td>Healthy</td>
<td/>
<td><inline-formula id="ieqn-71"><mml:math id="mml-ieqn-71"><mml:mo>&#x003E;</mml:mo></mml:math></inline-formula><bold>99%</bold></td>
<td>&#x00A1;1%</td>
</tr>
<tr>
<td>Ulcer</td>
<td/>
<td>&#x003C; 1%</td>
<td><inline-formula id="ieqn-72"><mml:math id="mml-ieqn-72"><mml:mo>&#x003E;</mml:mo></mml:math></inline-formula><bold>99%</bold></td>
</tr>
</tbody>
</table>

</table-wrap>
 
<table-wrap id="table-8">
<label>Table 8</label>
<caption>
<title>Classification results of proposed classification architecture for KVASIR Dataset. The maximum accuracy is bold in the table</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Classifier</th>
<th colspan="7">Performance measures</th>
<th></th>
</tr>
<tr>
<th></th>
<th>ACC (%)</th>
<th>SEN (%)</th>
<th>SPE (%)</th>
<th>PRE (%)</th>
<th>F1 score (%)</th>
<th>FPR</th>
<th>FNR (%)</th>
<th>Execution time (s)</th>
</tr>
</thead>
<tbody>
<tr>
<td>F-Tree</td>
<td>60.4</td>
<td>60.40</td>
<td>94.34</td>
<td>60.29</td>
<td>60.34</td>
<td>0.06</td>
<td>39.6</td>
<td>72.29</td>
</tr>
<tr>
<td>L-Disc</td>
<td>84.8</td>
<td>84.80</td>
<td>97.83</td>
<td>85.23</td>
<td>84.98</td>
<td>0.02</td>
<td>15.2</td>
<td>32.34</td>
</tr>
<tr>
<td>L-SVM</td>
<td>86.1</td>
<td>86.05</td>
<td>98.01</td>
<td>86.19</td>
<td>86.05</td>
<td>0.02</td>
<td>13.9</td>
<td>94.90</td>
</tr>
<tr>
<td>C-SVM</td>
<td>85.7</td>
<td>85.73</td>
<td>97.96</td>
<td>85.79</td>
<td>85.73</td>
<td>0.02</td>
<td>14.3</td>
<td>79.03</td>
</tr>
<tr>
<td>M-G-SVM</td>
<td>84.6</td>
<td>84.60</td>
<td>97.80</td>
<td>84.82</td>
<td>84.64</td>
<td>0.02</td>
<td>15.4</td>
<td>46.95</td>
</tr>
<tr>
<td>F-KNN</td>
<td>75.3</td>
<td>72.83</td>
<td>96.12</td>
<td>73.8</td>
<td>72.83</td>
<td>0.04</td>
<td>24.7</td>
<td><bold>26.06</bold></td>
</tr>
<tr>
<td>ESDA</td>
<td><bold>87.8</bold></td>
<td><bold>87.40</bold></td>
<td><bold>98.06</bold></td>
<td><bold>87.99</bold></td>
<td><bold>87.63</bold></td>
<td>0.02</td>
<td><bold>12.2</bold></td>
<td>50.70</td>
</tr>
</tbody>
</table>

</table-wrap>
 
<table-wrap id="table-9">
<label>Table 9</label>
<caption>
<title>Confusion matrix using selected fused LBP and inceptionV3 features for KVASIR dataset</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Classes</th>
<th colspan="8">Classification classes</th>
</tr>
<tr>
<th></th>
<th>DLP</th>
<th>DRM</th>
<th>ESO</th>
<th>NC</th>
<th>NP</th>
<th>NZL</th>
<th>P</th>
<th>UCE</th>
</tr>
</thead>
<tbody>
<tr>
<td><bold>DLP</bold></td>
<td><bold>78%</bold></td>
<td>19%</td>
<td/>
<td>&#x003C; 1%</td>
<td/>
<td/>
<td>2%</td>
<td/>
</tr>
<tr>
<td><bold>DRM</bold></td>
<td>21%</td>
<td><bold>77%</bold></td>
<td/>
<td>1%</td>
<td/>
<td/>
<td/>
<td>&#x003C; 1%</td>
</tr>
<tr>
<td><bold>ESO</bold></td>
<td/>
<td/>
<td><bold>78%</bold></td>
<td/>
<td>1%</td>
<td>21%</td>
<td/>
<td/>
</tr>
<tr>
<td><bold>NC</bold></td>
<td/>
<td/>
<td/>
<td><bold>96%</bold></td>
<td/>
<td/>
<td>3%</td>
<td>1%</td>
</tr>
<tr>
<td><bold>NP</bold></td>
<td/>
<td/>
<td>1%</td>
<td/>
<td><bold>98%</bold></td>
<td/>
<td>1%</td>
<td/>
</tr>
<tr>
<td><bold>ZL</bold></td>
<td/>
<td/>
<td>12%</td>
<td/>
<td>1%</td>
<td><bold>86%</bold></td>
<td/>
<td>&#x00A1;1%</td>
</tr>
<tr>
<td><bold>P</bold></td>
<td/>
<td/>
<td>1%</td>
<td>4%</td>
<td>2%</td>
<td>1%</td>
<td><bold>87%</bold></td>
<td>5%</td>
</tr>
<tr>
<td><bold>UCE</bold></td>
<td/>
<td/>
<td>1%</td>
<td>1%</td>
<td>1%</td>
<td/>
<td>3%</td>
<td><bold>94%</bold></td>
</tr>
</tbody>
</table>

</table-wrap>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Comparison with Existing Methods</title>
<p>From the above review, it is noted that several methods are proposed by the researchers for GI disease abnormality detection and classification. Most of the existing CAD methods are based on traditional techniques and algorithms, such as most of them are based on only color information or texture information. Although there exists some methods in which authors have used a combination of features. Some methods are also based on deep CNN features. Despite too many existing CAD methods, there exist some limitations in the old approaches such as low contrast of captured frames, the same color of the infected and healthy region, the problem of proper color model selection, hazy image, redundant information, etc. These limitations forced us to develop a robust method for GI abnormality detection and classification with better accuracy. The proposed deep learning method is evaluated on two datasets-Kvasir and Private and achieved an accuracy of 99.80% and 87.80%. The comparison with existing techniques is given in <xref ref-type="table" rid="table-10">Tab. 10</xref>. In this table, the comparison is conducted based on the abnormality name or dataset. Because most of the GI datasets are private, therefore, we mainly focused on disease type. From this table, it is shown that the proposed architecture gives better accuracy as well as execution time.</p>
<table-wrap id="table-10">
<label>Table 10</label>
<caption>
<title>Comparison with existing methods</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Ref.</th>
<th>Year</th>
<th>Disease/dataset</th>
<th colspan="5">Performance measures</th>
</tr>
<tr>
<th></th>
<th></th>
<th></th>
<th>ACC (%)</th>
<th>SEN (%)</th>
<th>SPE (%)</th>
<th>PRE (%)</th>
<th>Execution time (s)</th>
</tr>
</thead>
<tbody>
<tr>
<td>[<xref ref-type="bibr" rid="ref-28">28</xref>]</td>
<td>2019</td>
<td>Bleeding, Ulcer and Normal</td>
<td>99.54</td>
<td>100</td>
<td>&#x2013;</td>
<td>99.51</td>
<td>12.59</td>
</tr>
<tr>
<td>[<xref ref-type="bibr" rid="ref-38">38</xref>]</td>
<td>2019</td>
<td>Ulcer, Cancer, Normal</td>
<td>96.49</td>
<td><bold>&#x2013;</bold></td>
<td><bold>&#x2013;</bold></td>
<td><bold>&#x2013;</bold></td>
<td><bold>&#x2013;</bold></td>
</tr>
<tr>
<td>[<xref ref-type="bibr" rid="ref-37">37</xref>]</td>
<td>2018</td>
<td>Ulcer, bleeding</td>
<td>98.49</td>
<td>98.00</td>
<td>98.00</td>
<td>99.00</td>
<td>17.393</td>
</tr>
<tr>
<td><bold>Proposed</bold></td>
<td><bold>2020</bold></td>
<td>Private dataset: Bleeding, Ulcer and Normal</td>
<td><bold>99.80</bold></td>
<td><bold>99.83</bold></td>
<td><bold>99.92</bold></td>
<td><bold>99.80</bold></td>
<td><bold>17.031</bold></td>
</tr>
<tr>
<td/>
<td><bold>2020</bold></td>
<td>KVASIR dataset</td>
<td><bold>87.80</bold></td>
<td><bold>87.40</bold></td>
<td><bold>98.06</bold></td>
<td><bold>87.99</bold></td>
<td><bold>50.7</bold></td>
</tr>
</tbody>
</table>

</table-wrap>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Conclusion</title>
<p>In this article, we proposed a deep learning architecture for the detection and classification of GI abnormalities. The proposed architecture consists of two procedures for pipeline detection and classification. In the detection task, the bleeding region is segmented by a fusion of three separate channels. In the classification task, deep learning features and texture-oriented features are extracted and the best features are selected using the Shanon Entropy controlled ESDA classifier. The selected features are concatenated and are classified. In the evaluation phase, the segmentation process achieves an average accuracy of over 87% for abnormal bleeding regions. For classification, the accuracy of the private data set is 99.80 percent, while for the Kvasir data set, the accuracy is 87.80 percent. It is concluded from the results that the proposed selection method shows better performance compared to the existing techniques. It also concludes that the merger process is effective for more classes, such as the Kvasir dataset classification. In addition, texture features also have a high impact on disease classification and deep learning fusion, addressing the issue of texture variation. In future studies, we will focus on ulcer segmentation through deep learning.</p>
</sec>
</body>
<back>
<fn-group><fn fn-type="other"><p><bold>Funding Statement:</bold> This research was financially supported in part by the Ministry of Trade, Industry and Energy (MOTIE) and Korea Institute for Advancement of Technology (KIAT) through the International Cooperative R&#x0026;D program. (Project No. P0016038) and in part by the MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2021-2016-0-00312) supervised by the IITP (Institute for Information &#x0026; communications Technology Planning &#x0026; Evaluation).</p></fn>
<fn fn-type="conflict"><p><bold>Conflicts of Interest:</bold> The authors declare that they have no conflicts of interest to report regarding the present study.</p></fn></fn-group>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.-D.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Sharif</surname></string-name> and <string-name><given-names>T.</given-names> <surname>Akram</surname></string-name></person-group>, &#x201C;<article-title>Pixels to classes: Intelligent learning framework for multiclass skin lesion localization and classification</article-title>,&#x201D; <source>Computers &#x0026; Electrical Engineering</source>, vol. <volume>90</volume>, pp. <fpage>106956</fpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>H. T.</given-names> <surname>Rauf</surname></string-name>, <string-name><given-names>M. I. U.</given-names> <surname>Lali</surname></string-name>, <string-name><given-names>M. A.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Kadry</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Alolaiyan</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Time series forecasting of COVID-19 transmission in Asia Pacific countries using deep neural networks</article-title>,&#x201D; <source>Personal and Ubiquitous Computing</source>, vol. <volume>2</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>18</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Sharif</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Akram</surname></string-name>, <string-name><given-names>S. A. C.</given-names> <surname>Bukhari</surname></string-name> and <string-name><given-names>R. S.</given-names> <surname>Nayak</surname></string-name></person-group>, &#x201C;<article-title>Developed newton-raphson based deep features selection framework for skin lesion recognition</article-title>,&#x201D; <source>Pattern Recognition Letters</source>, vol. <volume>129</volume>, no. <issue>4/5</issue>, pp. <fpage>293</fpage>&#x2013;<lpage>303</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M. A.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Sharif</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Akram</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Yasmin</surname></string-name> and <string-name><given-names>R. S.</given-names> <surname>Nayak</surname></string-name></person-group>, &#x201C;<article-title>Stomach deformities recognition using rank-based deep features selection</article-title>,&#x201D; <source>Journal of Medical Systems</source>, vol. <volume>43</volume>, no. <issue>12</issue>, pp. <fpage>329</fpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>R.</given-names> <surname>Siegel</surname></string-name>, <string-name><given-names>C.</given-names> <surname>DeSantis</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Jemal</surname></string-name></person-group>, &#x201C;<article-title>Colorectal cancer statistics</article-title>,&#x201D; <source>2014 A Cancer Journal for Clinicians</source>, vol. <volume>64</volume>, no. <issue>2</issue>, pp. <fpage>104</fpage>&#x2013;<lpage>117</lpage>, <year>2014</year>.</mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>R. L.</given-names> <surname>Siegel</surname></string-name>, <string-name><given-names>K. D.</given-names> <surname>Miller</surname></string-name>, <string-name><given-names>S. A.</given-names> <surname>Fedewa</surname></string-name>, <string-name><given-names>D. J.</given-names> <surname>Ahnen</surname></string-name>, <string-name><given-names>R. G.</given-names> <surname>Meester</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Colorectal cancer statistics, 2017</article-title>,&#x201D; <source>A Cancer Journal for Clinicians</source>, vol. <volume>67</volume>, no. <issue>3</issue>, pp. <fpage>177</fpage>&#x2013;<lpage>193</lpage>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Rashid</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Sharif</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Javed</surname></string-name> and <string-name><given-names>T.</given-names> <surname>Akram</surname></string-name></person-group>, &#x201C;<article-title>Classification of gastrointestinal diseases of stomach from WCE using improved saliency-based method and discriminant features selection</article-title>,&#x201D; <source>Multimedia Tools and Applications</source>, vol. <volume>78</volume>, no. <issue>19</issue>, pp. <fpage>27743</fpage>&#x2013;<lpage>27770</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>R. L.</given-names> <surname>Siegel</surname></string-name>, <string-name><given-names>K. D.</given-names> <surname>Miller</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Jemal</surname></string-name></person-group>, &#x201C;<article-title>Cancer statistics</article-title>,&#x201D; <source>2019 A Cancer Journal for Clinicians</source>, vol. <volume>69</volume>, no. <issue>1</issue>, pp. <fpage>7</fpage>&#x2013;<lpage>34</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M. A.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>M. S.</given-names> <surname>Sarfraz</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Alhaisoni</surname></string-name>, <string-name><given-names>A. A.</given-names> <surname>Albesher</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Wang</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>StomachNet: Optimal deep learning features fusion for stomach abnormalities classification</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>8</volume>, pp. <fpage>197969</fpage>&#x2013;<lpage>197981</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Mackiewicz</surname></string-name></person-group>, &#x201C;<article-title>Capsule endoscopy-state of the technology and computer vision tools after the first decade, new techniques in gastrointestinal endoscopy</article-title>,&#x201D; <source>InTech ISBN</source>, vol. <volume>4</volume>, pp. <fpage>978</fpage>&#x2013;<lpage>953</lpage>, <year>2011</year>.</mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>G.</given-names> <surname>Iddan</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Meron</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Glukhovsky</surname></string-name> and <string-name><given-names>P.</given-names> <surname>Swain</surname></string-name></person-group>, &#x201C;<article-title>Wireless capsule endoscopy</article-title>,&#x201D; <source>Nature</source>, vol. <volume>405</volume>, no. <issue>6785</issue>, pp. <fpage>417</fpage>, <year>2000</year>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Liedlgruber</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Uhl</surname></string-name></person-group>, &#x201C;<article-title>Computer-aided decision support systems for endoscopy in the gastrointestinal tract: A review</article-title>,&#x201D; <source>IEEE Reviews in Biomedical Engineering</source>, vol. <volume>4</volume>, pp. <fpage>73</fpage>&#x2013;<lpage>88</lpage>, <year>2011</year>.</mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>G. M.</given-names> <surname>Cobrin</surname></string-name>, <string-name><given-names>R. H.</given-names> <surname>Pittman</surname></string-name> and <string-name><given-names>B. S.</given-names> <surname>Lewis</surname></string-name></person-group>, &#x201C;<article-title>Increased diagnostic yield of small bowel tumors with capsule endoscopy</article-title>,&#x201D; <source>Cancer: Interdisciplinary International Journal of the American Cancer Society</source>, vol. <volume>107</volume>, pp. <fpage>22</fpage>&#x2013;<lpage>27</lpage>, <year>2006</year>.</mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>G.</given-names> <surname>Liu</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Yan</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Kuang</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Wang</surname></string-name></person-group>, &#x201C;<article-title>Detection of small bowel tumor based on multi-scale curvelet analysis and fractal technology in capsule endoscopy</article-title>,&#x201D; <source>Computers in Biology and Medicine</source>, vol. <volume>70</volume>, no. <issue>3</issue>, pp. <fpage>131</fpage>&#x2013;<lpage>138</lpage>, <year>2016</year>.</mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>T.</given-names> <surname>Aoki</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Yamada</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Aoyama</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Saito</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Tsuboi</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Automatic detection of erosions and ulcerations in wireless capsule endoscopy images based on a deep convolutional neural network</article-title>,&#x201D; <source>Gastrointestinal Endoscopy</source>, vol. <volume>89</volume>, no. <issue>2</issue>, pp. <fpage>357</fpage>&#x2013;<lpage>363</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M. A.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Kadry</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Alhaisoni</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Nam</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Zhang</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Computer-aided gastrointestinal diseases analysis from wireless capsule endoscopy: A framework of best features selection</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>8</volume>, pp. <fpage>132850</fpage>&#x2013;<lpage>132859</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>B.-P.</given-names> <surname>Li</surname></string-name> and <string-name><given-names>M. Q.-H.</given-names> <surname>Meng</surname></string-name></person-group>, &#x201C;<article-title>Comparison of several texture features for tumor detection in CE images</article-title>,&#x201D; <source>Journal of Medical Systems</source>, vol. <volume>36</volume>, no. <issue>4</issue>, pp. <fpage>2463</fpage>&#x2013;<lpage>2469</lpage>, <year>2012</year>.</mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>K.</given-names> <surname>Pogorelov</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Suman</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Azmadi Hussin</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Saeed Malik</surname></string-name>, <string-name><given-names>O.</given-names> <surname>Ostroukhova</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Bleeding detection in wireless capsule endoscopy videos&#x2013;-color versus texture features</article-title>,&#x201D; <source>Journal of Applied Clinical Medical Physics</source>, vol. <volume>20</volume>, no. <issue>8</issue>, pp. <fpage>141</fpage>&#x2013;<lpage>154</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>T.</given-names> <surname>Saba</surname></string-name>, <string-name><given-names>M. A.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Rehman</surname></string-name> and <string-name><given-names>S. L.</given-names> <surname>Marie-Sainte</surname></string-name></person-group>, &#x201C;<article-title>Region extraction and classification of skin cancer: A heterogeneous framework of deep CNN features fusion and reduction</article-title>,&#x201D; <source>Journal of Medical Systems</source>, vol. <volume>43</volume>, pp. <fpage>289</fpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M. A.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Rubab</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Kashif</surname></string-name>, <string-name><given-names>M. I.</given-names> <surname>Sharif</surname></string-name>, <string-name><given-names>N.</given-names> <surname>Muhammad</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Lungs cancer classification from CT images: An integrated design of contrast based classical features fusion and selection</article-title>,&#x201D; <source>Pattern Recognition Letters</source>, vol. <volume>129</volume>, pp. <fpage>77</fpage>&#x2013;<lpage>85</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M. A.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>I.</given-names> <surname>Ashraf</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Alhaisoni</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Dama&#353;evi&#x010D;ius</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Scherer</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Multimodal brain tumor classification using deep learning and robust feature selection: A machine learning application for radiologists</article-title>,&#x201D; <source>Diagnostics</source>, vol. <volume>10</volume>, pp. <fpage>565</fpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Suman</surname></string-name>, <string-name><given-names>F. A.</given-names> <surname>Hussin</surname></string-name>, <string-name><given-names>A. S.</given-names> <surname>Malik</surname></string-name>, <string-name><given-names>S. H.</given-names> <surname>Ho</surname></string-name>, <string-name><given-names>I.</given-names> <surname>Hilmi</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Feature selection and classification of ulcerated lesions using statistical analysis for WCE images</article-title>,&#x201D; <source>Applied Sciences</source>, vol. <volume>7</volume>, no. <issue>10</issue>, pp. <fpage>1097</fpage>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Suman</surname></string-name>, <string-name><given-names>F. A. B.</given-names> <surname>hussin</surname></string-name>, <string-name><given-names>A. S.</given-names> <surname>Malik</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Pogorelov</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Riegler</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Detection and classification of bleeding region in WCE images using color feature</article-title>,&#x201D; in <conf-name>Proc. of the 15th Int. Workshop on Content-Based Multimedia Indexing</conf-name>, <publisher-loc>NY, United States</publisher-loc>, pp. <fpage>1</fpage>&#x2013;<lpage>6</lpage>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Charfi</surname></string-name> and <string-name><given-names>M. El</given-names> <surname>Ansari</surname></string-name></person-group>, &#x201C;<article-title>Computer-aided diagnosis system for colon abnormalities detection in wireless capsule endoscopy images</article-title>,&#x201D; <source>Multimedia Tools and Applications</source>, vol. <volume>77</volume>, no. <issue>3</issue>, pp. <fpage>4047</fpage>&#x2013;<lpage>4064</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>O.</given-names> <surname>Bchir</surname></string-name>, <string-name><given-names>M. M. B.</given-names> <surname>Ismail</surname></string-name> and <string-name><given-names>N.</given-names> <surname>AlZahrani</surname></string-name></person-group>, &#x201C;<article-title>Multiple bleeding detection in wireless capsule endoscopy</article-title>,&#x201D; <source>Signal, Image and Video Processing</source>, vol. <volume>13</volume>, no. <issue>1</issue>, pp. <fpage>121</fpage>&#x2013;<lpage>126</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-26"><label>[26]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Souaidi</surname></string-name>, <string-name><given-names>A. A.</given-names> <surname>Abdelouahed</surname></string-name> and <string-name><given-names>M.</given-names> <surname>El Ansari</surname></string-name></person-group>, &#x201C;<article-title>Multi-scale completed local binary patterns for ulcer detection in wireless capsule endoscopy images</article-title>,&#x201D; <source>Multimedia Tools and Applications</source>, vol. <volume>78</volume>, no. <issue>10</issue>, pp. <fpage>13091</fpage>&#x2013;<lpage>13108</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-27"><label>[27]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Fan</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Xu</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Fan</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Wei</surname></string-name> and <string-name><given-names>L.</given-names> <surname>Li</surname></string-name></person-group>, &#x201C;<article-title>Computer-aided detection of small intestinal ulcer and erosion in wireless capsule endoscopy images</article-title>,&#x201D; <source>Physics in Medicine &#x0026; Biology</source>, vol. <volume>63</volume>, pp. <fpage>165001</fpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-28"><label>[28]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Sharif</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Attique Khan</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Rashid</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Yasmin</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Afza</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Deep CNN and geometric features-based gastrointestinal tract diseases detection and classification from wireless capsule endoscopy images</article-title>,&#x201D; <source>Journal of Experimental &#x0026; Theoretical Artificial Intelligence</source>, vol. <volume>2</volume>, no. <issue>7</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>23</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-29"><label>[29]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>D. E.</given-names> <surname>Diamantis</surname></string-name>, <string-name><given-names>D. K.</given-names> <surname>Iakovidis</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Koulaouzidis</surname></string-name></person-group>, &#x201C;<article-title>Look-behind fully convolutional neural network for computer-aided endoscopy</article-title>,&#x201D; <source>Biomedical Signal Processing and Control</source>, vol. <volume>49</volume>, no. <issue>2</issue>, pp. <fpage>192</fpage>&#x2013;<lpage>201</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-30"><label>[30]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Alaskar</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Hussain</surname></string-name>, <string-name><given-names>N.</given-names> <surname>Al-Aseem</surname></string-name>, <string-name><given-names>P.</given-names> <surname>Liatsis</surname></string-name> and <string-name><given-names>D.</given-names> <surname>Al-Jumeily</surname></string-name></person-group>, &#x201C;<article-title>Application of convolutional neural networks for automated ulcer detection in wireless capsule endoscopy images</article-title>,&#x201D; <source>Sensors</source>, vol. <volume>19</volume>, no. <issue>5</issue>, pp. <fpage>1265</fpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-31"><label>[31]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M. A.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>M. A.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Ahmed</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Mittal</surname></string-name>, <string-name><given-names>L. M.</given-names> <surname>Goyal</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Gastrointestinal diseases segmentation and classification based on duo-deep architectures</article-title>,&#x201D; <source>Pattern Recognition Letters</source>, vol. <volume>131</volume>, pp. <fpage>193</fpage>&#x2013;<lpage>204</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-32"><label>[32]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>V.</given-names> <surname>Rajinikanth</surname></string-name>, <string-name><given-names>A. N.</given-names> <surname>Joseph Raj</surname></string-name>, <string-name><given-names>K. P.</given-names> <surname>Thanaraj</surname></string-name> and <string-name><given-names>G. R.</given-names> <surname>Naik</surname></string-name></person-group>, &#x201C;<article-title>A customized VGG19 network with concatenation of deep and handcrafted features for brain tumor detection</article-title>,&#x201D; <source>Applied Sciences</source>, vol. <volume>10</volume>, no. <issue>10</issue>, pp. <fpage>3429</fpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-33"><label>[33]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>N.</given-names> <surname>Dey</surname></string-name>, <string-name><given-names>Y.-D.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>V.</given-names> <surname>Rajinikanth</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Pugalenthi</surname></string-name> and <string-name><given-names>N. S. M.</given-names> <surname>Raja</surname></string-name></person-group>, &#x201C;<article-title>Customized VGG19 architecture for pneumonia detection in chest X-rays</article-title>,&#x201D; <source>Pattern Recognition Letters</source>, vol. <volume>2</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>8</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-34"><label>[34]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>N.</given-names> <surname>Dey</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Shi</surname></string-name> and <string-name><given-names>V.</given-names> <surname>Rajinikanth</surname></string-name></person-group>, &#x201C;<article-title>Image examination system to detect gastric polyps from endoscopy images</article-title>,&#x201D; <source>Information Technology and Intelligent Transportation Systems</source>, vol. <volume>323</volume>, pp. <fpage>107</fpage>&#x2013;<lpage>116</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-35"><label>[35]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>T.</given-names> <surname>Ojala</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Pietik&#x00E4;inen</surname></string-name> and <string-name><given-names>D.</given-names> <surname>Harwood</surname></string-name></person-group>, &#x201C;<article-title>A comparative study of texture measures with classification based on featured distributions</article-title>,&#x201D; <source>Pattern Recognition</source>, vol. <volume>29</volume>, no. <issue>1</issue>, pp. <fpage>51</fpage>&#x2013;<lpage>59</lpage>, <year>1996</year>.</mixed-citation></ref>
<ref id="ref-36"><label>[36]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>K.</given-names> <surname>Pogorelov</surname></string-name>, <string-name><given-names>K. R.</given-names> <surname>Randel</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Griwodz</surname></string-name>, <string-name><given-names>S. L.</given-names> <surname>Eskeland</surname></string-name> and <string-name><given-names>T.</given-names> <surname>de Lange</surname></string-name></person-group>, &#x201C;<article-title>Kvasir: A multi-class image dataset for computer aided gastrointestinal disease detection</article-title>,&#x201D; in <conf-name>Proc. of the 8th ACM on Multimedia Systems Conf.</conf-name>, <publisher-loc>NY, United States</publisher-loc>, pp. <fpage>164</fpage>&#x2013;<lpage>169</lpage>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-37"><label>[37]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Liaqat</surname></string-name>, <string-name><given-names>J. H.</given-names> <surname>Shah</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Sharif</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Yasmin</surname></string-name> and <string-name><given-names>S. L.</given-names> <surname>Fernandes</surname></string-name></person-group>, &#x201C;<article-title>Automated ulcer and bleeding classification from WCE images using multiple features fusion and selection</article-title>,&#x201D; <source>Journal of Mechanics in Medicine and Biology</source>, vol. <volume>18</volume>, no. <issue>04</issue>, pp. <fpage>1850038</fpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-38"><label>[38]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>J. H.</given-names> <surname>Lee</surname></string-name>, <string-name><given-names>Y. J.</given-names> <surname>Kim</surname></string-name>, <string-name><given-names>Y. W.</given-names> <surname>Kim</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Park</surname></string-name>, <string-name><given-names>Y.-I.</given-names> <surname>Choi</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Spotting malignancies from gastric endoscopic images using deep learning</article-title>,&#x201D; <source>Surgical Endoscopy</source>, vol. <volume>11</volume>, no. <issue>2</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>8</lpage>, <year>2019</year>.</mixed-citation></ref>
</ref-list>
</back>
</article>