<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">IASC</journal-id>
<journal-id journal-id-type="nlm-ta">IASC</journal-id>
<journal-id journal-id-type="publisher-id">IASC</journal-id>
<journal-title-group>
<journal-title>Intelligent Automation &#x0026; Soft Computing</journal-title>
</journal-title-group>
<issn pub-type="epub">2326-005X</issn>
<issn pub-type="ppub">1079-8587</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">19513</article-id>
<article-id pub-id-type="doi">10.32604/iasc.2021.019513</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>CT Segmentation of Liver and Tumors Fused Multi-Scale Features</article-title>
<alt-title alt-title-type="left-running-head">CT Segmentation of Liver and Tumors Fused Multi-Scale Features</alt-title>
<alt-title alt-title-type="right-running-head">CT Segmentation of Liver and Tumors Fused Multi-Scale Features</alt-title>
</title-group>
<contrib-group content-type="authors">
<contrib id="author-1" contrib-type="author">
<name name-style="western">
<surname>Yu</surname>
<given-names>Aihong</given-names>
</name>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<contrib id="author-2" contrib-type="author" corresp="yes">
<name name-style="western">
<surname>Liu</surname>
<given-names>Zhe</given-names>
</name>
<xref ref-type="aff" rid="aff-1">1</xref>
<email>1000004088@ujs.edu.cn</email>
</contrib>
<contrib id="author-3" contrib-type="author">
<name name-style="western">
<surname>Sheng</surname>
<given-names>Victor S.</given-names>
</name>
<xref ref-type="aff" rid="aff-2">2</xref>
</contrib>
<contrib id="author-4" contrib-type="author">
<name name-style="western">
<surname>Song</surname>
<given-names>Yuqing</given-names>
</name>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<contrib id="author-5" contrib-type="author">
<name name-style="western">
<surname>Liu</surname>
<given-names>Xuesheng</given-names>
</name>
<xref ref-type="aff" rid="aff-3">3</xref>
</contrib>
<contrib id="author-6" contrib-type="author">
<name name-style="western">
<surname>Ma</surname>
<given-names>Chongya</given-names>
</name>
<xref ref-type="aff" rid="aff-4">4</xref>
</contrib>
<contrib id="author-7" contrib-type="author">
<name name-style="western">
<surname>Wang</surname>
<given-names>Wenqiang</given-names>
</name>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<contrib id="author-8" contrib-type="author">
<name name-style="western">
<surname>Ma</surname>
<given-names>Cong</given-names>
</name>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<aff id="aff-1">
<label>1</label><institution>School of Computer Science and Communications Engineering, Jiangsu University</institution>, <addr-line>Zhenjiang, 212013</addr-line>, <country>China</country></aff>
<aff id="aff-2">
<label>2</label><institution>Department of Computer Science, University of Central Arkansas</institution>, <addr-line>Conway, 72032</addr-line>, <country>USA</country></aff>
<aff id="aff-3">
<label>3</label><institution>Department of Anesthesiology, the First Affiliated Hospital of Anhui Medical University</institution>, <addr-line>Hefei, 230001</addr-line>, <country>China</country></aff>
<aff id="aff-4">
<label>4</label><institution>College of Tourism and Geographic Science, Jilin Normal University</institution>, <addr-line>Jilin, 136000</addr-line>, <country>China</country></aff>
</contrib-group><author-notes><corresp id="cor1">&#x002A;Corresponding Author: Zhe Liu. Email: <email>1000004088@ujs.edu.cn</email></corresp></author-notes>
<pub-date pub-type="epub" date-type="pub" iso-8601-date="2021-07-29">
<day>29</day>
<month>7</month>
<year>2021</year>
</pub-date>
<volume>30</volume>
<issue>2</issue>
<fpage>589</fpage>
<lpage>599</lpage>
<history>
<date date-type="received">
<day>14</day>
<month>4</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>15</day>
<month>5</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2021 Yu et al.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Yu et al.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_IASC_19513.pdf"></self-uri>
<abstract>
<p>Liver cancer is one of frequent causes of death from malignancy in the world. Owing to the outstanding advantages of computer-aided diagnosis and deep learning, fully automatic segmentation of computed tomography (CT) images turned into a research hotspot over the years. The liver has quite low contrast with the surrounding tissues, together with its lesion areas are thoroughly complex. To deal with these problems, we proposed effective methods for enhancing features and processed public datasets from Liver Tumor Segmentation Challenge (LITS) for the verification. In this experiment, data pre-processing based on the image enhancement and noise reduction. This study redesigned the original UNet with two novel modules and named it DResUNet which was applied deformable convolution. The first module aimed to recalibrate information by the channel and spatial dimension. The other module enriched deep information of these liver CT images through fusing multi-scale features. Besides, we used cross-entropy loss function for adaptive weights to solve the troubles of class imbalance in the dataset samples. These can improve the performance of the network in-depth and breadth feature learning to deal with many complex segmentation scenes in abdominal CT images. More importantly, the effect of predicted images fully proved that our methods are highly competitive among the segmentation of liver and liver tumors.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Computed tomography (CT)</kwd>
<kwd>liver tumors</kwd>
<kwd>deep learning</kwd>
<kwd>fully automatic segmentation</kwd>
<kwd>multi-scale features</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>The liver resembles a large chemical processing plant inside the human body, maintaining vital life activities. As the largest substantial tissue in the human abdomen, the liver has plenty of functions, such as metabolism, detoxification, and immune defense [<xref ref-type="bibr" rid="ref-1">1</xref>]. Nevertheless, the liver is extremely vulnerable to suffer trauma or physical injury, which leads to multiple liver diseases [<xref ref-type="bibr" rid="ref-2">2</xref>]. According to the latest statistics from the World Health Organization, liver cancer has a high mortality rate worldwide. Since liver tumors endanger people&#x2019;s health to a large extent, it is of great clinical significance to accurately diagnose diseases both and assist doctors in formulating suitable treatment plans. As the rapid development of computer technology, imaging technology has become one of the indispensable methods of clinical medical diagnosis [<xref ref-type="bibr" rid="ref-3">3</xref>]. Computed tomography (CT) has been widely used to reveal internal injury or bleeding, due to its fast speed of scanning and less harmful to the human body [<xref ref-type="bibr" rid="ref-4">4</xref>]. It is necessary to know the location of the liver, the size and number of tumors when the doctors diagnose liver cancer or monitor treatment response [<xref ref-type="bibr" rid="ref-5">5</xref>]. Therefore, segmentation of liver and lesions accurately on CT scan abdominal images became especially required.</p>
<p>Manual segmentation has been applied initially, but this is time-consuming and inefficient. Doctors are required to have a professional level and rich experience. Semi-automatic segmentation which combines manual operation and computer-assisted processing has been utilized afterward. Some researchers analyzed the gray histogram of the tumor and used the maximum between-class variance to obtain the optimal threshold for extracting the liver [<xref ref-type="bibr" rid="ref-6">6</xref>]. Then, they used the method of contrast stretching to enhance the contrast of the region of interest and manually segmented the tumor region at last. Other scholars [<xref ref-type="bibr" rid="ref-7">7</xref>] proposed an adaptive method that used traditional region-growing algorithms which achieved great results in CT images segmentation only if there is a single liver tumor. Certain experts [<xref ref-type="bibr" rid="ref-8">8</xref>] designed a morphological gradient correction level set method. Different sizes of structural elements in gray-level neighborhoods to perform corresponding local gradient corrections. Representative algorithms about semi-automatic segmentation mainly include region growth, threshold, level set and graph cut, etc [<xref ref-type="bibr" rid="ref-9">9</xref>]. Semi-automatic segmentation required manual design and extraction of features. As a result, the ability to express features was limited.</p>
<p>In recent years, artificial intelligence has developed rapidly [<xref ref-type="bibr" rid="ref-10">10</xref>]. Image segmentation technologies based on various neural networks have also been applied to the field of medical image processing [<xref ref-type="bibr" rid="ref-11">11</xref>]. This way can be named fully automatic segmentation. Convolutional neural network (CNN) is a globally artificial neural network model that can learn abstract, essential, and deep features. LeNet is mainly applied to recognize and classify handwritten characters with very high accuracy [<xref ref-type="bibr" rid="ref-12">12</xref>]. Many excellent networks included AlexNet [<xref ref-type="bibr" rid="ref-13">13</xref>], VggNet [<xref ref-type="bibr" rid="ref-14">14</xref>], and ResNet [<xref ref-type="bibr" rid="ref-15">15</xref>] have been produced on the basis of LeNet. Fully Convolutional Network (FCN) was proposed by Long et al to make the scale extracted by CNN reach the size of the original image [<xref ref-type="bibr" rid="ref-16">16</xref>]. The fully connected layers of the convolutional neural network are replaced with convolutional layers. This method not only can input data of any size, but also realizes pixel-level image segmentation. Researchers have proposed many improved fully convolutional neural networks such as UNet [<xref ref-type="bibr" rid="ref-17">17</xref>], DeepLab [<xref ref-type="bibr" rid="ref-18">18</xref>], and PSPNet [<xref ref-type="bibr" rid="ref-19">19</xref>] which have achieved great results in the image segmentation.</p>
<p>The article is mainly to solve the problem of missing segmentation targets and reducing segmentation errors, especially for small targets. Firstly, we proposed a hybrid enhanced attention module at the channel and spatial level. In addition, designed a high-resolution and low-resolution feature fusion module. Simultaneously, we analyzed the complexity of segmentation targets and the imbalance of target categories about samples. We applied deformable convolution and used an adaptive weighted cross-entropy loss function. Therefore, we have created a novel liver and tumors segmentation network DResUNet that based on UNet, which applied residual learning and multi-scale feature fusion. After training and testing on Liver Tumor Segmentation Challenge (LITS) dataset, it is proved that the segmentation results and efficiency of the network have been steadily improved.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Related Theories</title>
<sec id="s2_1">
<label>2.1</label>
<title>Deformable Convolution</title>
<p>The purpose of the convolution kernel is to extract more features [<xref ref-type="bibr" rid="ref-20">20</xref>]. The traditional convolution kernel has poor adaptability to unknown changes. Different positions need correspond to objects with different scales or deformations, the model needs to automatically adjust the changes. Therefore, the method of deformable convolution was presented.</p>
<p>The deviation can be obtained through an ordinary convolution kernel firstly. Generated channel dimension is 2N, corresponding to the original output feature and the offset feature respectively. These two convolution kernels learn together through the bilinear interpolation backward propagation algorithm. After that, the size and position of the deformable convolution kernel can be dynamically adjusted according to the current image content that needs to be recognized [<xref ref-type="bibr" rid="ref-21">21</xref>].</p>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Residual Block</title>
<p>Deeper networks are more difficult to be trained. For the purpose of addressing gradient explosion and gradient disappearance, residual blocks have been applied to the model. Shortcut connection and mapping are the main designs of residual block. Experiments proved that the residual block often requires more than two layers, and the residual block of a single layer could not work. The so-called residual network is actually a network composed of several residual blocks.</p>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>Windowing</title>
<p>Hu represents the true density of an object, independent of the device, and values within different ranges can represent different organs. The contrast of the image depends on the window width and the brightness of the image depends on the window level. window width is the range of ct values contained in a ct image. The window level is the midpoint of the ct value. As the window level decreases, the image will be brighter and the opposite will be darker. When the window width and the window level are given, the upper and lower boundaries of the window can be calculated. The target range of its windowing is [level &#x2212; width/2, level &#x002B; width/2].</p>
</sec>
<sec id="s2_4">
<label>2.4</label>
<title>Gaussian Filtering</title>
<p>The presence of noise will seriously affect the accuracy of segmentation, so noise reduction was needed on CT images. Gaussian filtering is a linear smoothing filter. Its basic principle is to use gaussian function to perform weighted average operation on the entire image. The gray value of each pixel and the value of all other pixels in the neighborhood are weighted and averaged to obtain a new image. This new gray value will replace the gray value of the original pixel [<xref ref-type="bibr" rid="ref-22">22</xref>].</p>
</sec>
<sec id="s2_5">
<label>2.5</label>
<title>Histogram Equalization</title>
<p>The histogram describes the distribution of gray levels in an image. Through the histogram operation to change the gray distribution in the original image to achieve the effect of enhancing the image. Histogram equalization is a common way of histogram operation. The central idea of histogram equalization was to redistribute relatively concentrated grayscale intervals in original histogram, so that pixels of images were distributed across entire grayscale range [<xref ref-type="bibr" rid="ref-23">23</xref>].</p>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Methods</title>
<sec id="s3_1">
<label>3.1</label>
<title>CT Data Pre-processing Based on Image Enhancement and Noise Reduction</title>
<p>The preprocessing process of this experiment can be shown in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>. It was necessary to read data at first and extracted the target slices after windowing operation. Then, we used gaussian filtering to reduce some noise of images. This research applied histogram equalization to enhance the presentation of features. Last but not least, normalization was an indispensable step.</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>Steps of Pre-processing</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19513-fig-1.png"/>
</fig>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>DResUNet Model Structure</title>
<p>The network structure is shown in <xref ref-type="fig" rid="fig-2">Fig. 2</xref>. The overall structure is a typical ResUnet structure. It was named as DResUNet due to the use of deformable convolution and feature extraction modules which we proposed. In detail, we passed images through firstblock for processing during the stage of down-sampling. Then, we used Max_Pooling for down-sampling. The processed data is transferred not only to the efficient channel and spatial attention residual module (EcsaResM), but also to the concat operation layer at the up-sampling stage. This module used deformable convolution and embed designed efficient channel and spatial attention into residual blocks.</p>
<p>After four rounds of iteration, the images entered the up-sampling stage. The same five sets of up-sampling operations which was named deconvblock restore the feature map size to the same size as the original image. Next, we designed the feature pyramid rearrangement module, the characteristics will be enhanced. In addition, after the processing of finalblock, we final got the segmentation results for liver and tumors prediction.</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>The structure of proposed DResUNet</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19513-fig-2.png"/>
</fig>
<sec id="s3_2_1">
<label>3.2.1</label>
<title>Efficient Channel and Spatial Attention</title>
<p>Being the core of convolutional neural network, convolution kernel aggregates spatial information and channel-wise information on the local receptive field. In order to obtain more key features of the target area in the images, we designed a hybrid attention block of channel and space. The block performed feature recalibration of CT images after comprehensively considering the information which on the channel together with spatial dimension level. One-dimensional convolution was used to quickly realize local cross-channel interaction without dimensionality reduction, and learn effective attention more effective. Therefore, we named it the efficient channel and spatial attention (ECSA), and the specific implementation process is shown in <xref ref-type="fig" rid="fig-3">Fig. 3</xref>.</p>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>Mechanism of proposed efficient channel and spatial attention</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19513-fig-3.png"/>
</fig>
<p>It was mainly divided into efficient channel attention mechanism and spatial attention mechanism. Efficient channel attention as follows: first, we changed the feature map from C &#x00D7; H &#x00D7; W to C &#x00D7; 1 &#x00D7; 1 through the global average pooling and max pooling method, and we used K &#x00D7; 1 &#x00D7; 1 convolution to process information and obtain a C-dimensional vector. We used the sigmoid function to normalize and got the corresponding weights. Finally, through channel-wise multiplication, the feature map after information calibration is obtained. Spatial attention mechanism part: firstly, we directly used 1 &#x00D7; 1 &#x00D7; 1 convolution on the feature map to change C &#x00D7; H &#x00D7; W into features of 1 &#x00D7; H &#x00D7; W. Secondly, sigmoid function for activation to get the spatial attention map. Then we directly apply it to the original feature map to complete the spatial information calibration.</p>
<p>In <xref ref-type="fig" rid="fig-3">Fig. 3</xref>, it can be showed that the ECSA is a parallel connection of the above two parts. Specifically, after the feature map passes through the two modules, the two parts are added together to obtain a more accurately calibrated feature map. It should be noted that without dimensionality reduction, K is adaptively determined through a function related to the channel dimension. After all, the interaction coverage rate is related to the channel dimension.</p>
<p>Here, we embed ECSA into residual blocks, applying deformable convolution before the residual blocks to form efficient channel and spatial attention residual module (EcsaResM). More details about the principle of EcsaResM can be obtained in <xref ref-type="fig" rid="fig-4">Fig. 4</xref>.</p>
<fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>The principle of EcsaResM</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19513-fig-4.png"/>
</fig>
</sec>
<sec id="s3_2_2">
<label>3.2.2</label>
<title>Feature Pyramid Rearrangement</title>
<p>Almost every parameter of the CNN network will affect the generated feature map. This implies that a network with a fixed structure will often learn features of a specific scale, and the network may also have a certain scale invariance. Mixing the feature map directly into a vector will lose spatial information and cause ambiguity. At the same time, these parameters were often closely related to the current task and cannot be easily modified, which made it difficult for the model to learn multi-scale information. Global context information with subregion information was helpful for distinguishing multiple categories. Furthermore, feature representation more powerful can fuse information from different sub-regions with multiple receptive fields.</p>
<p>For the purpose of making up for the insufficient feature learning generated by the model when extracting the deep-level information of the image, we designed another network block to supplement the information of the previous EcsaResM. It was like saying that we proposed feature pyramid rearrangement module (FPRM) to integrate the context of different regions to obtain global context information. The input image passed through a feature extraction network to extract features. This part used ResUnet with deformable convolution. The function of deformable convolution was mainly to increase the receptive field to adapt to the geometric deformation of targets, and the extracted features are used as the input of the FPRM. In <xref ref-type="fig" rid="fig-5">Fig. 5</xref>, a feature pyramid with a depth of 4 was constructed. Features of different depths are obtained through pooling operations of different scales based on input features, and the scale of pooling can be adjusted. Therefore, multiple levels of pooling kernels should be maintained a reasonable gap.</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>Process of feature pyramid rearrangement module</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19513-fig-5.png"/>
</fig>
<p>There are four levels of FPRM, and the pooling kernel size of each level is 1 &#x00D7; 1, 2 &#x00D7; 2, 3 &#x00D7; 3 and 6 &#x00D7; 6. Then, we used a 1 &#x00D7; 1 convolutional layer to reduce the feature dimension to a quarter of the original. Afterward, we resized feature maps about different scales to convert them into feature maps of four sizes. After rearranging the feature maps according to the size, we performed the operation of concat. The feature maps of different scales were sequentially upsampled and concatenated together with the feature maps of the higher level, and then the feature maps of each level were upsampled by bilinear interpolation to the size of the original feature map before being sent to the FPRM. Finally, we concatenated all the features on the pyramid levels as well as the original features together as the final output of this module.</p>
</sec>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Cross-entropy Loss Function for Adaptive Weights</title>
<p>We applied a weighted cross-entropy function as the loss function in experiments. In the case of unbalanced categories in the samples, different weights can be assigned to different categories. In this experiment, we needed to use the adaptive weighting approach. The equation is as follows:</p>
<p><disp-formula id="eqn-1">
<label>(1)</label>
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-1.png"/><tex-math id="tex-eqn-1"><![CDATA[$${\rm L}\left( {y,\hat y} \right) = - \displaystyle{1 \over N}\mathop \sum \nolimits_{i = 1}^N \mathop \sum \nolimits_{c = 1}^3 w_i^cy_i^clog\hat y_i^c$$]]></tex-math>--><mml:math id="mml-eqn-1" display="block"><mml:mrow><mml:mi mathvariant="normal">L</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>y</mml:mi><mml:mo>,</mml:mo><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x003D;</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mstyle scriptlevel="0" displaystyle="true"><mml:mrow><mml:mfrac><mml:mn>1</mml:mn><mml:mi>N</mml:mi></mml:mfrac></mml:mrow><mml:msubsup><mml:mrow><mml:mo movablelimits="false">&#x2211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>N</mml:mi></mml:msubsup><mml:mo>&#x2061;</mml:mo><mml:msubsup><mml:mrow><mml:mo movablelimits="false">&#x2211;</mml:mo></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mn>3</mml:mn></mml:msubsup><mml:mo>&#x2061;</mml:mo><mml:msubsup><mml:mi>w</mml:mi><mml:mi>i</mml:mi><mml:mi>c</mml:mi></mml:msubsup><mml:msubsup><mml:mi>y</mml:mi><mml:mi>i</mml:mi><mml:mi>c</mml:mi></mml:msubsup><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>g</mml:mi><mml:msubsup><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mi>i</mml:mi><mml:mi>c</mml:mi></mml:msubsup></mml:mstyle></mml:math>
<!--</alternatives>--></disp-formula></p>
<p>In this formula,<inline-formula id="ieqn-1">
<!--<alternatives><inline-graphic xlink:href="ieqn-1.tif"/><tex-math id="tex-ieqn-1"><![CDATA[$\; \; \hat y_i^c\; \;$]]></tex-math>--><mml:math id="mml-ieqn-1"><mml:mspace width="thickmathspace"></mml:mspace><mml:mspace width="thickmathspace"></mml:mspace><mml:msubsup><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mi>i</mml:mi><mml:mi>c</mml:mi></mml:msubsup><mml:mspace width="thickmathspace"></mml:mspace><mml:mspace width="thickmathspace"></mml:mspace></mml:math>
<!--</alternatives>--></inline-formula> denotes the probability that image i falls into category c which includes background, liver or liver lesions.<inline-formula id="ieqn-2">
<!--<alternatives><inline-graphic xlink:href="ieqn-2.tif"/><tex-math id="tex-ieqn-2"><![CDATA[$\; w_i^c\; \;$]]></tex-math>--><mml:math id="mml-ieqn-2"><mml:mspace width="thickmathspace"></mml:mspace><mml:msubsup><mml:mi>w</mml:mi><mml:mi>i</mml:mi><mml:mi>c</mml:mi></mml:msubsup><mml:mspace width="thickmathspace"></mml:mspace><mml:mspace width="thickmathspace"></mml:mspace></mml:math>
<!--</alternatives>--></inline-formula> indicates weights, <inline-formula id="ieqn-3">
<!--<alternatives><inline-graphic xlink:href="ieqn-3.tif"/><tex-math id="tex-ieqn-3"><![CDATA[$\; y_i^c\;$]]></tex-math>--><mml:math id="mml-ieqn-3"><mml:mspace width="thickmathspace"></mml:mspace><mml:msubsup><mml:mi>y</mml:mi><mml:mi>i</mml:mi><mml:mi>c</mml:mi></mml:msubsup><mml:mspace width="thickmathspace"></mml:mspace></mml:math>
<!--</alternatives>--></inline-formula> represents the truth tag for image i. In addition, the weighting of the categories is defined by the following formula:</p>
<p><disp-formula id="eqn-2">
<label>(2)</label>
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-2.png"/><tex-math id="tex-eqn-2"><![CDATA[$${{\rm \omega }_{{\rm class}}} = \displaystyle{1 \over {\ln \left( {t + {P_{class}}} \right)}}$$]]></tex-math>--><mml:math id="mml-eqn-2" display="block"><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x03C9;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">c</mml:mi><mml:mi mathvariant="normal">l</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">s</mml:mi><mml:mi mathvariant="normal">s</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x003D;</mml:mo><mml:mstyle scriptlevel="0" displaystyle="true"><mml:mrow><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mi>ln</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>&#x002B;</mml:mo><mml:mrow><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>a</mml:mi><mml:mi>s</mml:mi><mml:mi>s</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:math>
<!--</alternatives>--></disp-formula></p>
<p>We set t which is a hyper-parameter to 1.02. <inline-formula id="ieqn-4">
<!--<alternatives><inline-graphic xlink:href="ieqn-4.tif"/><tex-math id="tex-ieqn-4"><![CDATA[${P_{class}}\; \;$]]></tex-math>--><mml:math id="mml-ieqn-4"><mml:mrow><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>a</mml:mi><mml:mi>s</mml:mi><mml:mi>s</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mspace width="thickmathspace"></mml:mspace><mml:mspace width="thickmathspace"></mml:mspace></mml:math>
<!--</alternatives>--></inline-formula> is the proportion of the sample of that class. We also limited <inline-formula id="ieqn-5">
<!--<alternatives><inline-graphic xlink:href="ieqn-5.tif"/><tex-math id="tex-ieqn-5"><![CDATA[${{\rm \omega }_{{\rm class}}}$]]></tex-math>--><mml:math id="mml-ieqn-5"><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x03C9;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">c</mml:mi><mml:mi mathvariant="normal">l</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">s</mml:mi><mml:mi mathvariant="normal">s</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula> to [1.0, 50].</p>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Experiments</title>
<sec id="s4_1">
<label>4.1</label>
<title>Dataset and Equipment</title>
<p>We processed the public dataset about Liver Tumor Segmentation (LITS) provided by Coda Lab. There were hundreds of CT image slices for each patient. The data which have been Pre-processing can be shown in the <xref ref-type="fig" rid="fig-6">Fig. 6</xref>. The abdominal CT images and labels of cases used in this research all from LITS which were manually annotated by some radiologists. From these 131 groups of cases, we selected 118 groups of cases as training set and tumor labels and 25 groups of cases for testing. In addition, we set to extract 45 valid slices in each group of cases.</p>
<fig id="fig-6">
<label>Figure 6</label>
<caption>
<title>The slices used to train</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19513-fig-6.png"/>
</fig>
<p>The experimental environment was a computer running Linux Ubuntu 16.04 64-bit operating system, using Intel i7 4.20 GHz CPU (Intel Core i7-7700k), 16gb memory and NVIDIA GeForce GTX 1070 graphics card.</p>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Evaluation Metrics and Results</title>
<p>In order to prove that the two new modules designed in this network and the adaptive weighted cross-entropy loss function achieved superior results for liver and tumors segmentation, the ablation experiments were first performed and the trained model was tested on the test set, and the segmentation results are shown in the <xref ref-type="fig" rid="fig-6">Fig. 6</xref>. It indicated the test results of the proposed network for liver and tumors segmentation together with the results of manual annotation by experts.</p>
<p>From <xref ref-type="fig" rid="fig-7">Fig. 7</xref>, we can see that the segmentation results obtained from the model designed in this paper were very close to expert annotation images, and this method can effectively optimize the border details of liver and tumor and solve low contrast and complex areas problems.</p>
<fig id="fig-7">
<label>Figure 7</label>
<caption>
<title>Raw, ground truth and segmentation results on the test image</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19513-fig-7.png"/>
</fig>
<p>We wanted to verify the effectiveness of the model for completing segmentation tasks, so we will compare the test results about Dice. The details of the comparison can be seen in <xref ref-type="table" rid="table-1">Tab. 1</xref>. From the data in the table, we can find that the performance has been slightly improved after adding ECSA and FPRM separately; and after using cross-entropy loss function for adaptive weights, the performance has been better improved. This also verifies the efficiency of this model.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>Quantitative results of the Liver and Liver Tumors</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Models</th>
<th>Dice (Liver)</th>
<th>Dice (Liver Tumors)</th>
</tr>
</thead>
<tbody>
<tr>
<td>ResUNet</td>
<td>0.951</td>
<td>0.658</td>
</tr>
<tr>
<td>ResUNet &#x002B; ECSA</td>
<td>0.957</td>
<td>0.663</td>
</tr>
<tr>
<td>ResUNet &#x002B; FPRM</td>
<td>0.959</td>
<td>0.665</td>
</tr>
<tr>
<td>DResUNet</td>
<td>0.965</td>
<td>0.672</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>When we conducted comparative experiments, five metrics are used to measure the accuracy of segmentation results, including the volumetric overlap error (VOE), relative volume difference (RVD), average symmetric surface distance (ASSD), root mean square symmetric surface distance (MSD) and DICE. <xref ref-type="table" rid="table-2">Tabs. 2</xref> and <xref ref-type="table" rid="table-3">3</xref> made clear results of comparing the proposed method with other segmentation algorithms. It can be seen that the Dice of the proposed network model in segmenting the liver was 0.965, and the values of other indicators were also relatively high. The Dice of segmented tumor was 0.672, which stated clearly that the network structure was more stable.</p>
<table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>Results compared with other methods (Liver)</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Approach</th>
<th>Dice</th>
<th>VOE</th>
<th>RVD</th>
<th>ASSD</th>
<th>MSD</th>
</tr>
</thead>
<tbody>
<tr>
<td>Li et al. [<xref ref-type="bibr" rid="ref-24">24</xref>]<break/>Kaluva et al. [<xref ref-type="bibr" rid="ref-25">25</xref>]<break/>Yuan et al. [<xref ref-type="bibr" rid="ref-26">26</xref>]</td>
<td>0.961<break/> 0.912<break/> 0.963</td>
<td>0.074<break/> 0.150<break/> 0.071</td>
<td>&#x2212;0.018<break/>&#x2212;0.008<break/>&#x2212;0.010</td>
<td>1.450<break/>6.465<break/>1.104</td>
<td>27.118<break/>45.928<break/>23.847</td>
</tr>
<tr>
<td>Our method</td>
<td>0.965</td>
<td>0.078</td>
<td>0.003</td>
<td>1.126</td>
<td>26.958</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>Results compared with other methods (Liver Tumors)</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Approach</th>
<th>Dice</th>
<th>VOE</th>
<th>RVD</th>
<th>ASSD</th>
<th>MSD</th>
</tr>
</thead>
<tbody>
<tr>
<td>Li et al. [<xref ref-type="bibr" rid="ref-24">24</xref>]<break/>Kaluva et al. [<xref ref-type="bibr" rid="ref-25">25</xref>]<break/>Yuan et al. [<xref ref-type="bibr" rid="ref-26">26</xref>]</td>
<td>0.722<break/> 0.625<break/> 0.657</td>
<td>0.366<break/> 0.411<break/> 0.378</td>
<td>4.272<break/> 19.705<break/> 0.288</td>
<td>1.102<break/>1.441<break/>1.151</td>
<td>6.228<break/>7.515<break/>6.269</td>
</tr>
<tr>
<td>Our method</td>
<td>0.672</td>
<td>0.398</td>
<td>0.516</td>
<td>1.332</td>
<td>6.565</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Although the Dice of H-DenseUNet was slightly higher, it used 3DdenseUNet that required higher conditions for hardware acceleration operations. The model we proposed can complete the segmentation of liver and tumors under the condition of low hardware requirements. It also fully proved that the modules designed by us effectively optimized the segmentation results both and obtained more accurate predicted target areas of CT scan abdominal images.</p>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Conclusions</title>
<p>This paper proposed novel methods to segment liver and tumors, because the liver has the low contrast with surrounding tissues and its lesion complex areas. The DResUNet still achieved accurate segmentation in the case of insufficient data. Experimental results showed that insertion of residual blocks during the down-sampling process and the designed new two modules can significantly improve the accuracy of liver and tumors segmentation. In addition, deformable convolution was applied to enlarge the receptive field to adapt to targets deformations. The most important thing was new embedded modules alleviate the blurring of the liver and each tumor boundary and complex areas effectively.</p>
</sec>
</body>
<back>
<ack>
<p>The authors would like to thank Radiologists of the Medical Imaging department of Affiliated Hospital of Jiangsu University, University of Central Arkansas, the First Affiliated Hospital of Anhui Medical University, Jilin Normal University.</p>
</ack><fn-group>
<fn fn-type="other">
<p><bold>Funding Statement:</bold> This work was supported by Zhenjiang Key Deprogram &#x201C;Fire Early Warning Technology Based on Multimodal Data Analysis&#x201D; (SH2020011) and Jiangsu Emergency Management Science and Technology Project &#x201C;Research on Very Early Warning of Fire Based on Multi-modal Data Analysis and Multi-Intelligent Body Technology&#x201D; (YJGL-TG-2020-8).</p>
</fn>
<fn fn-type="conflict">
<p><bold>Conflicts of Interest:</bold> The authors declare no conflict of interest.</p>
</fn>
</fn-group>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1">
<label>1</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>Z.</given-names> 
<surname>Liu</surname></string-name>, <string-name>
<given-names>Y. Q.</given-names> 
<surname>Song</surname></string-name>, <string-name>
<given-names>V. S.</given-names> 
<surname>Sheng</surname></string-name>, <string-name>
<given-names>L.</given-names> 
<surname>Wang</surname></string-name>, <string-name>
<given-names>R.</given-names> 
<surname>Jiang</surname></string-name> <etal>et al.</etal>
</person-group><italic>,</italic> &#x201C;
<article-title>Liver CT sequence segmentation based with improved U-Net and graph cut</article-title>,&#x201D; 
<source>Expert Systems with Applications</source>, vol. 
<volume>126</volume>, pp. 
<fpage>54</fpage>&#x2013;
<lpage>63</lpage>, 
<year>2019</year>.</mixed-citation>
</ref>
<ref id="ref-2">
<label>2</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>Z.</given-names> 
<surname>Bai</surname></string-name>, <string-name>
<given-names>H.</given-names> 
<surname>Jiang</surname></string-name>, <string-name>
<given-names>S.</given-names> 
<surname>Li</surname></string-name> and <string-name>
<given-names>Y. D.</given-names> 
<surname>Yao</surname></string-name>
</person-group>, &#x201C;
<article-title>Liver tumor segmentation based on multi-scale candidate generation and fractal residual network</article-title>,&#x201D; 
<source>IEEE Access</source>, vol. 
<volume>7</volume>, pp. 
<fpage>82122</fpage>&#x2013;
<lpage>82133</lpage>, 
<year>2019</year>.</mixed-citation>
</ref>
<ref id="ref-3">
<label>3</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>H.</given-names> 
<surname>Wu</surname></string-name>, <string-name>
<given-names>Q.</given-names> 
<surname>Liu</surname></string-name> and <string-name>
<given-names>X.</given-names> 
<surname>Liu</surname></string-name>
</person-group>, &#x201C;
<article-title>A review on deep learning approaches to image classification and object segmentation</article-title>,&#x201D; 
<source>Computers, Materials &#x0026; Continua</source>, vol. 
<volume>60</volume>, no. 
<issue>2</issue>, pp. 
<fpage>575</fpage>&#x2013;
<lpage>597</lpage>, 
<year>2019</year>.</mixed-citation>
</ref>
<ref id="ref-4">
<label>4</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>B.</given-names> 
<surname>Hu</surname></string-name> and <string-name>
<given-names>J.</given-names> 
<surname>Wang</surname></string-name>
</person-group>, &#x201C;
<article-title>Deep learning for distinguishing computer-generated images and natural images: A survey</article-title>,&#x201D; 
<source>Journal of Information Hiding and Privacy Protection</source>, vol. 
<volume>2</volume>, no. 
<issue>2</issue>, pp. 
<fpage>95</fpage>, 
<year>2020</year>.</mixed-citation>
</ref>
<ref id="ref-5">
<label>5</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>E.</given-names> 
<surname>Vorontsov</surname></string-name>, <string-name>
<given-names>M.</given-names> 
<surname>Cerny</surname></string-name> and <string-name>
<given-names>P.</given-names> 
<surname>R&#x00E9;gnier</surname></string-name>
</person-group>, &#x201C;
<article-title>Deep learning for automated segmentation of liver lesions at CT in patients with colorectal cancer liver metastases</article-title>,&#x201D; 
<source>Radiology: Artificial Intelligence</source>, vol. 
<volume>1</volume>, no. 
<issue>2</issue>, pp. 
<fpage>180014</fpage>, 
<year>2020</year>.</mixed-citation>
</ref>
<ref id="ref-6">
<label>6</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>A.</given-names> 
<surname>Nugroho</surname></string-name>, <string-name>
<given-names>D.</given-names> 
<surname>Ihtatho</surname></string-name> and <string-name>
<given-names>H.</given-names> 
<surname>Nugroho</surname></string-name>
</person-group>, &#x201C;
<article-title>Contrast enhancement for liver tumor identification</article-title>,&#x201D; 
<source>MICCAI Workshop</source>, vol. 
<volume>41</volume>, no. 
<issue>43</issue>, pp. 
<fpage>201</fpage>, 
<year>2008</year>.</mixed-citation>
</ref>
<ref id="ref-7">
<label>7</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>C. C.</given-names> 
<surname>Chang</surname></string-name>, <string-name>
<given-names>H. H.</given-names> 
<surname>Chen</surname></string-name> and <string-name>
<given-names>Y. C.</given-names> 
<surname>Chang</surname></string-name>
</person-group>, &#x201C;
<article-title>Computer-aided diagnosis of liver tumors on computed tomography images</article-title>,&#x201D; 
<source>Computer methods and programs in biomedicine</source>, vol. 
<volume>145</volume>, pp. 
<fpage>45</fpage>&#x2013;
<lpage>51</lpage>, 
<year>2017</year>.</mixed-citation>
</ref>
<ref id="ref-8">
<label>8</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>M.</given-names> 
<surname>Cui</surname></string-name>, <string-name>
<given-names>H.</given-names> 
<surname>Chen</surname></string-name> and <string-name>
<given-names>J.</given-names> 
<surname>Zhou</surname></string-name>
</person-group>, &#x201C;
<article-title>A level-set based multi-material topology optimization method using a reaction diffusion equation</article-title>,&#x201D; 
<source>Computer-Aided Design</source>, vol. 
<volume>73</volume>, pp. 
<fpage>41</fpage>&#x2013;
<lpage>52</lpage>, 
<year>2016</year>.</mixed-citation>
</ref>
<ref id="ref-9">
<label>9</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>A.</given-names> 
<surname>Khaloo</surname></string-name> and <string-name>
<given-names>L.</given-names> 
<surname>David</surname></string-name>
</person-group>, &#x201C;
<article-title>Robust normal estimation and region growing segmentation of infrastructure 3D point cloud models</article-title>,&#x201D; 
<source>Advanced Engineering Informatics</source>, vol. 
<volume>34</volume>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>16</lpage>, 
<year>2017</year>.</mixed-citation>
</ref>
<ref id="ref-10">
<label>10</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>V. S.</given-names> 
<surname>Sheng</surname></string-name> and <string-name>
<given-names>J.</given-names> 
<surname>Zhang</surname></string-name>
</person-group>, &#x201C;
<article-title>Machine learning with crowdsourcing: A brief summary of the past research and future directions</article-title>,&#x201D; 
<source>Proc. of the AAAI Conf. on Artificial Intelligence</source>, vol. 
<volume>33</volume>, pp. 
<fpage>9837</fpage>&#x2013;
<lpage>9843</lpage>, 
<year>2019</year>.</mixed-citation>
</ref>
<ref id="ref-11">
<label>11</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>J.</given-names> 
<surname>Hu</surname></string-name>, <string-name>
<given-names>Z.</given-names> 
<surname>Luo</surname></string-name>, <string-name>
<given-names>X.</given-names> 
<surname>Wang</surname></string-name>, <string-name>
<given-names>S.</given-names> 
<surname>Sun</surname></string-name>, <string-name>
<given-names>Y.</given-names> 
<surname>Yin</surname></string-name> <etal>et al.</etal>
</person-group><italic>,</italic> &#x201C;
<article-title>End-to-end multimodal image registration via reinforcement learning</article-title>,&#x201D; 
<source>Medical Image Analysis</source>, vol. 
<volume>68</volume>, pp. 
<fpage>101878</fpage>, 
<year>2020</year>.</mixed-citation>
</ref>
<ref id="ref-12">
<label>12</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>Y.</given-names> 
<surname>LeCun</surname></string-name>, <string-name>
<given-names>L.</given-names> 
<surname>Bottou</surname></string-name>, <string-name>
<given-names>Y.</given-names> 
<surname>Bengio</surname></string-name> and <string-name>
<given-names>P.</given-names> 
<surname>Haffner</surname></string-name>
</person-group>, &#x201C;
<article-title>Gradient-based learning applied to document recognition</article-title>,&#x201D; 
<source>Proceedings of the IEEE</source>, vol. 
<volume>86</volume>, no. 
<issue>11</issue>, pp. 
<fpage>2278</fpage>&#x2013;
<lpage>2324</lpage>, 
<year>1998</year>.</mixed-citation>
</ref>
<ref id="ref-13">
<label>13</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>A.</given-names> 
<surname>Krizhevsky</surname></string-name>, <string-name>
<given-names>I.</given-names> 
<surname>Sutskever</surname></string-name> and <string-name>
<given-names>G. E.</given-names> 
<surname>Hinton</surname></string-name>
</person-group>, &#x201C;
<article-title>Imagenet classification with deep convolutional neural networks</article-title>,&#x201D; 
<source>Communications of the ACM</source>, vol. 
<volume>60</volume>, no. 
<issue>6</issue>, pp. 
<fpage>84</fpage>&#x2013;
<lpage>90</lpage>, 
<year>2017</year>.</mixed-citation>
</ref>
<ref id="ref-14">
<label>14</label><mixed-citation publication-type="other">
<person-group person-group-type="author"><string-name>
<given-names>K.</given-names> 
<surname>Simonyan</surname></string-name> and <string-name>
<given-names>A.</given-names> 
<surname>Zisserman</surname></string-name>
</person-group>, &#x201C;
<article-title>Very deep convolutional networks for large-scale image recognition</article-title>,&#x201D; 
<comment>arXiv: cs.CV, arXiv:2003.09424</comment>, 
<year>2020</year>. [Online]. Available at: <uri>https://arxiv.org/abs/1409.1556</uri>.</mixed-citation>
</ref>
<ref id="ref-15">
<label>15</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>K.</given-names> 
<surname>He</surname></string-name>, <string-name>
<given-names>X.</given-names> 
<surname>Zhang</surname></string-name>, <string-name>
<given-names>S.</given-names> 
<surname>Ren</surname></string-name> and <string-name>
<given-names>J.</given-names> 
<surname>Sun</surname></string-name>
</person-group>, &#x201C;
<article-title>Deep residual learning for image recognition</article-title>,&#x201D; in <conf-name>Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition</conf-name>, pp. 
<fpage>770</fpage>&#x2013;
<lpage>778</lpage>, 
<year>2016</year>. </mixed-citation>
</ref>
<ref id="ref-16">
<label>16</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>J.</given-names> 
<surname>Long</surname></string-name>, <string-name>
<given-names>E.</given-names> 
<surname>Shelhamer</surname></string-name> and <string-name>
<given-names>T.</given-names> 
<surname>Darrell</surname></string-name>
</person-group>, &#x201C;
<article-title>Fully convolutional networks for semantic segmentation</article-title>,&#x201D; in <conf-name>Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition</conf-name>, pp. 
<fpage>3431</fpage>&#x2013;
<lpage>3440</lpage>, 
<year>2015</year>. </mixed-citation>
</ref>
<ref id="ref-17">
<label>17</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>O.</given-names> 
<surname>Ronneberger</surname></string-name>, <string-name>
<given-names>P.</given-names> 
<surname>Fischer</surname></string-name> and <string-name>
<given-names>T.</given-names> 
<surname>Brox</surname></string-name>
</person-group>, &#x201C;
<article-title>U-net: Convolutional networks for biomedical image segmentation</article-title>,&#x201D; in <conf-name>Int. Conf. on Medical Image Computing and Computer-Assisted Intervention</conf-name>, 
<publisher-loc>Cham</publisher-loc>: 
<publisher-name>Springer</publisher-name>, pp. 
<fpage>234</fpage>&#x2013;
<lpage>241</lpage>, 
<year>2015</year>. </mixed-citation>
</ref>
<ref id="ref-18">
<label>18</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>L. C.</given-names> 
<surname>Chen</surname></string-name>, <string-name>
<given-names>G.</given-names> 
<surname>Papandreou</surname></string-name>, <string-name>
<given-names>I.</given-names> 
<surname>Kokkinos</surname></string-name>, <string-name>
<given-names>K.</given-names> 
<surname>Murphy</surname></string-name> and <string-name>
<given-names>A. L.</given-names> 
<surname>Yuille</surname></string-name>
</person-group>, &#x201C;
<article-title>Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs</article-title>,&#x201D; 
<source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>, vol. 
<volume>40</volume>, no. 
<issue>4</issue>, pp. 
<fpage>834</fpage>&#x2013;
<lpage>848</lpage>, 
<year>2017</year>.</mixed-citation>
</ref>
<ref id="ref-19">
<label>19</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>H.</given-names> 
<surname>Zhao</surname></string-name>, <string-name>
<given-names>J.</given-names> 
<surname>Shi</surname></string-name>, <string-name>
<given-names>X.</given-names> 
<surname>Qi</surname></string-name>, <string-name>
<given-names>X.</given-names> 
<surname>Wang</surname></string-name> and <string-name>
<given-names>J.</given-names> 
<surname>Jia</surname></string-name>
</person-group>, &#x201C;
<article-title>Pyramid scene parsing network</article-title>,&#x201D; in <conf-name>Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition</conf-name>, pp. 
<fpage>2881</fpage>&#x2013;
<lpage>2890</lpage>, 
<year>2017</year>. </mixed-citation>
</ref>
<ref id="ref-20">
<label>20</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>J.</given-names> 
<surname>Chen</surname></string-name>, <string-name>
<given-names>Z.</given-names> 
<surname>Zhou</surname></string-name>, <string-name>
<given-names>Z.</given-names> 
<surname>Pan</surname></string-name> and <string-name>
<given-names>C. N.</given-names> 
<surname>Yang</surname></string-name>
</person-group>, &#x201C;
<article-title>Instance Retrieval Using Region of Interest Based CNN Features</article-title>,&#x201D; 
<source>Journal of New Media</source>, vol. 
<volume>1</volume>, no. 
<issue>2</issue>, pp. 
<fpage>87</fpage>, 
<year>2019</year>.</mixed-citation>
</ref>
<ref id="ref-21">
<label>21</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>J.</given-names> 
<surname>Dai</surname></string-name>, <string-name>
<given-names>H.</given-names> 
<surname>Qi</surname></string-name>, <string-name>
<given-names>Y.</given-names> 
<surname>Xiong</surname></string-name>, <string-name>
<given-names>Y.</given-names> 
<surname>Li</surname></string-name>, <string-name>
<given-names>G.</given-names> 
<surname>Zhang</surname></string-name> <etal>et al.</etal>
</person-group><italic>,</italic> &#x201C;
<article-title>Deformable convolutional networks</article-title>,&#x201D; in <conf-name>Proc. of the IEEE Int. Conf. on Computer Vision</conf-name>, pp. 
<fpage>764</fpage>&#x2013;
<lpage>773</lpage>, 
<year>2017</year>. </mixed-citation>
</ref>
<ref id="ref-22">
<label>22</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>Z.</given-names> 
<surname>Liu</surname></string-name>, <string-name>
<given-names>Y.</given-names> 
<surname>Song</surname></string-name> and <string-name>
<given-names>Z.</given-names> 
<surname>Tang</surname></string-name>
</person-group>, &#x201C;
<article-title>Noised image segmentation based on rough set and orthogonal polynomial density model</article-title>,&#x201D; 
<source>Journal of Electronic Imaging</source>, vol. 
<volume>24</volume>, no. 
<issue>2</issue>, pp. 
<fpage>023010</fpage>, 
<year>2015</year>.</mixed-citation>
</ref>
<ref id="ref-23">
<label>23</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>M.</given-names> 
<surname>Raghavendra</surname></string-name>, <string-name>
<given-names>M. V.</given-names> 
<surname>Lakshmaiah</surname></string-name> and <string-name>
<given-names>S.</given-names> 
<surname>Dastagiri</surname></string-name>
</person-group>, &#x201C;
<article-title>Image enhancement using histogram equalization and spatial filtering</article-title>,&#x201D; 
<source>International Journal of Science and Research (IJSR)</source>, vol. 
<volume>1</volume>, no. 
<issue>3</issue>, pp. 
<fpage>105</fpage>&#x2013;
<lpage>107</lpage>, 
<year>2012</year>.</mixed-citation>
</ref>
<ref id="ref-24">
<label>24</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>X.</given-names> 
<surname>Li</surname></string-name>, <string-name>
<given-names>H.</given-names> 
<surname>Chen</surname></string-name>, <string-name>
<given-names>X.</given-names> 
<surname>Qi</surname></string-name>, <string-name>
<given-names>Q.</given-names> 
<surname>Dou</surname></string-name>, <string-name>
<given-names>C. W.</given-names> 
<surname>Fu</surname></string-name> <etal>et al.</etal>
</person-group><italic>,</italic> &#x201C;
<article-title>H-DenseUNet: Hybrid densely connected UNet for liver and tumor segmentation from CT volumes</article-title>,&#x201D; 
<source>IEEE Transactions on Medical Imaging</source>, vol. 
<volume>37</volume>, no. 
<issue>12</issue>, pp. 
<fpage>2663</fpage>&#x2013;
<lpage>2674</lpage>, 
<year>2018</year>.</mixed-citation>
</ref>
<ref id="ref-25">
<label>25</label><mixed-citation publication-type="other">
<person-group person-group-type="author"><string-name>
<given-names>K. C.</given-names> 
<surname>Kaluva</surname></string-name>, <string-name>
<given-names>M.</given-names> 
<surname>Khened</surname></string-name>, <string-name>
<given-names>A.</given-names> 
<surname>Kori</surname></string-name> and <string-name>
<given-names>K.</given-names> 
<surname>Ganapathy</surname></string-name>
</person-group>, &#x201C;
<article-title>2D-densely connected convolution neural networks for automatic liver and tumor segmentation</article-title>.&#x201D; 
<comment>arXiv: cs.CV, arXiv: 1802.02182</comment>, 
<year>2018</year>. [Online]. Available at: <uri>https://arxiv.org/abs/1802.02182</uri>.</mixed-citation>
</ref>
<ref id="ref-26">
<label>26</label><mixed-citation publication-type="other">
<person-group person-group-type="author"><string-name>
<given-names>Y.</given-names> 
<surname>Yuan</surname></string-name>
</person-group>, &#x201C;
<article-title>Hierarchical convolutional-deconvolutional neural networks for automatic liver and tumor segmentation</article-title>.&#x201D; 
<comment>arXiv: cs.CV, arXiv:1710.04540</comment>, 
<year>2017</year>. [Online]. Available at: <uri>https://arxiv.org/abs/1710.04540</uri>.</mixed-citation>
</ref>
</ref-list>
</back>
</article>