<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">IASC</journal-id>
<journal-id journal-id-type="nlm-ta">IASC</journal-id>
<journal-id journal-id-type="publisher-id">IASC</journal-id>
<journal-title-group>
<journal-title>Intelligent Automation &#x0026; Soft Computing</journal-title>
</journal-title-group>
<issn pub-type="epub">2326-005X</issn>
<issn pub-type="ppub">1079-8587</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">30017</article-id>
<article-id pub-id-type="doi">10.32604/iasc.2023.030017</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Cephalopods Classification Using Fine Tuned Lightweight Transfer Learning Models</article-title><alt-title alt-title-type="left-running-head">Cephalopods Classification Approach Using Fine Tuned Lightweight Transfer Learning</alt-title><alt-title alt-title-type="right-running-head">Cephalopods Classification Approach Using Fine Tuned Lightweight Transfer Learning</alt-title>
</title-group>
<contrib-group content-type="authors">
<contrib id="author-1" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Prabha</surname><given-names>P. Anantha</given-names></name>
<xref ref-type="aff" rid="aff-1">1</xref><email>pananthaprabha@outlook.com</email>
</contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western"><surname>Suchitra</surname><given-names>G.</given-names></name>
<xref ref-type="aff" rid="aff-2">2</xref>
</contrib>
<contrib id="author-3" contrib-type="author">
<name name-style="western"><surname>Saravanan</surname><given-names>R.</given-names></name>
<xref ref-type="aff" rid="aff-3">3</xref>
</contrib>
<aff id="aff-1"><label>1</label><institution>Department of Computer Science &#x0026; Engineering, Sri Krishna College of Technology</institution>, <addr-line>Coimbatore, 641042, Tamil Nadu</addr-line>, <country>India</country></aff>
<aff id="aff-2"><label>2</label><institution>Department of Electronics and Communication Engineering, Government College of Technology</institution>, <addr-line>Coimbatore, 641013, Tamil Nadu</addr-line>, <country>India</country></aff>
<aff id="aff-3"><label>3</label><institution>Department of Marine Pharmacology, Faculty of Allied Health Sciences, Chettinad Academy of Research and Education</institution>, <addr-line>Kelambakkam, Chengalpattu, 603103, Tamil Nadu</addr-line>, <country>India</country></aff>
</contrib-group><author-notes><corresp id="cor1"><label>&#x002A;</label>Corresponding Author: P. Anantha Prabha. Email: <email>pananthaprabha@outlook.com</email></corresp></author-notes>
<pub-date pub-type="epub" date-type="pub" iso-8601-date="2022-08-08"><day>08</day>
<month>08</month>
<year>2022</year></pub-date>
<volume>35</volume>
<issue>3</issue>
<fpage>3065</fpage>
<lpage>3079</lpage>
<history>
<date date-type="received"><day>16</day><month>3</month><year>2022</year></date>
<date date-type="accepted"><day>26</day><month>4</month><year>2022</year></date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2023 Prabha, Suchitra and Saravanan</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Prabha, Suchitra and Saravanan</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_IASC_30017.pdf"></self-uri>
<abstract>
<p>Cephalopods identification is a formidable task that involves hand inspection and close observation by a malacologist. Manual observation and identification take time and are always contingent on the involvement of experts. A system is proposed to alleviate this challenge that uses transfer learning techniques to classify the cephalopods automatically. In the proposed method, only the Lightweight pre-trained networks are chosen to enable IoT in the task of cephalopod recognition. First, the efficiency of the chosen models is determined by evaluating their performance and comparing the findings. Second, the models are fine-tuned by adding dense layers and tweaking hyperparameters to improve the classification of accuracy. The models also employ a well-tuned Rectified Adam optimizer to increase the accuracy rates. Third, Adam with Gradient Centralisation (RAdamGC) is proposed and used in fine-tuned models to reduce the training time. The framework enables an Internet of Things (IoT) or embedded device to perform the classification tasks by embedding a suitable lightweight pre-trained network. The fine-tuned models, MobileNetV2, InceptionV3, and NASNet Mobile have achieved a classification accuracy of 89.74%, 87.12%, and 89.74%, respectively. The findings have indicated that the fine-tuned models can classify different kinds of cephalopods. The results have also demonstrated that there is a significant reduction in the training time with RAdamGC.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Cephalopods</kwd>
<kwd>transfer learning</kwd>
<kwd>lightweight models</kwd>
<kwd>classification</kwd>
<kwd>deep learning</kwd>
<kwd>fish</kwd>
<kwd>IoT</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>Marine molluscs are a distinct class of creatures with more than 100,000 diverse classes presently. Cephalopods are the members of the mollusc phylum encompassing fewer than 1000 classes scattered in 43 families [<xref ref-type="bibr" rid="ref-1">1</xref>]. Cephalopod has seized to grow steadily in the last four decades. According to food and agricultural organization, the catching of cephalopods has increased from 1 million metric tonnes to 3.6 million metric tonnes.</p>
<p>The most extravagant species in the Indian market are those rapports to cuttlefish and squid [<xref ref-type="bibr" rid="ref-2">2</xref>,<xref ref-type="bibr" rid="ref-3">3</xref>]. This experience makes to have approaches for classifying cephalopod species, despite morphological characteristics that have been evacuated. Earlier cephalopods are classified using biochemical and molecular taxonomic markers [<xref ref-type="bibr" rid="ref-4">4</xref>,<xref ref-type="bibr" rid="ref-5">5</xref>]. The molecular taxonomy of cephalopods is highly expensive and time-consuming, so there is a need for an alternative tool. This study focuses on developing an automated system that classifies and separates the cephalopod species into different categories.</p>
<p>Deep Learning (DL) techniques have recently been used to solve various classification problems in robotics, sports, and medicine [<xref ref-type="bibr" rid="ref-6">6</xref>&#x2013;<xref ref-type="bibr" rid="ref-9">9</xref>]. The most prevalent deep learning models [<xref ref-type="bibr" rid="ref-10">10</xref>] are Convolutional Neural Networks (CNNs), which offer a lot of potential for the classification of tasks [<xref ref-type="bibr" rid="ref-11">11</xref>]. Salman et al. [<xref ref-type="bibr" rid="ref-12">12</xref>] had created a deep network to classify undersea species of fish. Deep learning involves an extensive dataset and a lot of computing power, leading to overfitting in small datasets [<xref ref-type="bibr" rid="ref-13">13</xref>].</p>
<p>Allken et al. [<xref ref-type="bibr" rid="ref-14">14</xref>] suggested a strategy for increasing the training sample based on a rational deep vision picture simulation to compensate for the lack of labeled data, reaching a categorization rate of 94% for Atlantic herring, blue whiting, and Atlantic mackerel. A method based on cross-convolutional layer pooling on a pre-trained CNN was presented in [<xref ref-type="bibr" rid="ref-15">15</xref>] to avoid the requirement for a significant quantity of training data. A validation accuracy of 98.03% was found after a thorough analysis of 27,370 species of images.</p>
<p>Xu et al. [<xref ref-type="bibr" rid="ref-16">16</xref>] also reported the species recognition based on SE-ResNet152 and class-balanced focal loss. The marine fish Tuna was similarly classified using a region-based CNN with ensembling [<xref ref-type="bibr" rid="ref-17">17</xref>]. When tested with an independent test dataset, the proposed approach achieved a precision of 93.91%. Based on this, Iqbal et al. [<xref ref-type="bibr" rid="ref-18">18</xref>] presented a model which employed a simplified AlexNet model with four convolutional layers and two fully connected layers.</p>
<p>It is demonstrated to be advantageous that use of sample expansion and transfer learning methodologies to improve the learning model generalization and mitigate the problem of limited sample data overfitting. Furthermore, the optimal model structure for various target datasets usually is different. An automatic cephalopods identification method is proposed using the transfer learning approach based on the issues. A network with very little weight is necessary to enable an IoT device or any embedded devices [<xref ref-type="bibr" rid="ref-19">19</xref>,<xref ref-type="bibr" rid="ref-20">20</xref>] to recognize cephalopod images automatically. The following is a list of contributions to the present research.<list list-type="bullet"><list-item>
<p>For the novel cephalopod image classification, in this study, lightweight pre-trained models (Slim Networks) are being used. Furthermore, the models are fine-tuned for improving the accuracy of classification.</p></list-item><list-item>
<p>An optimizer is proposed to boost the accuracy rates further while reducing the training time.</p></list-item><list-item>
<p>The models are tested, and the performance of the models is compared to select the ideal candidates for integration in embedded or IoT devices for cephalopod species classification.</p></list-item></list></p>
<p>The following sections are laid out as follows; Data sources for this article and the pre-trained models are discussed in Section 2. Section 3 presents the results of the tests, the specifics of the suggested optimizer, and interpretations of the findings. Section 4 wraps up the results presented in this paper.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Materials and Methods</title>
<sec id="s2_1">
<label>2.1</label>
<title>Image Collection</title>
<p>A Canon EOS 1000D was used to capture the cephalopod images in this dataset with a resolution of 10.1 effective megapixels, 3,888 &#x00D7; 2,592, from Kasimedu and Tuticorin coastal areas of Bay of Bengal. Six common species of cephalopods, namely <italic>Sepia pharaonis</italic>, <italic>Sepia prashad</italic>, <italic>Sepia aculeata</italic>, <italic>Sepioteuthis lessoniana</italic>, <italic>Uroteuthis duvauceli</italic>, and <italic>Doryteuthis siboga</italic> are collected and used in the proposed work. Field and sample images are shown in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>. The selected species of cephalopods are found all over India, and their reproduction speed is also high.</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>(a) Field images of cephalopod species (b) Sample images of cephalopod species</title></caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_30017-fig-1.png"/>
</fig>
<p>These six species are noteworthy in terms of classification study. Totally, 492 images (<italic>D. sibogae</italic>, 84; <italic>S. aculeata</italic>, 88; <italic>S. pharaonis</italic>, 80; <italic>S. Prashad</italic>, 88; <italic>S. lessoniana</italic>, 78; <italic>U. duvauceli</italic>, 74) have been collected from the coastal areas. With these photographs, a new cephalopod image dataset is created. The newly created dataset is used for the first time in the proposed work.</p>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Image Augmentation and Scaling</title>
<p>Data augmentation was used in the proposed study to address the overfitting issues by increasing the base data set and implementing the label-preserving transformations [<xref ref-type="bibr" rid="ref-21">21</xref>]. The following transformation operations were used to expand the dataset. The methods uniform distribution, rotating images at random within a precise range were used. Images were shifted vertically and horizontally by a random percentage of their total size. A part of the pictures was flipped horizontally, then arbitrary scaling and shearing changes were done. Image data generator [<xref ref-type="bibr" rid="ref-22">22</xref>] accepts batches of images and transforms them into a new set of images at random. Before being applied to any model, the image of size 4608 &#x00D7; 3456 is scaled down to size 224 &#x00D7; 224.</p>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>Proposed System for Cephalopod Classification</title>
<sec id="s2_3_1">
<label>2.3.1</label>
<title>Fine Tuned MobileNetV2 Architecture</title>
<p>The MobileNet model focuses on depthwise separable convolutions, a factorized convolution that splits a standard convolution into a depth-wise and a pointwise convolution. The depth-wise convolution used by Mobile Nets utilizes a single filter for each input channel. The outputs of depth-wise convolution are then combined using an 1 &#x00D7; 1 convolution by pointwise convolution. This is divided into two layers by the depthwise separable convolutions; one for filtering and the other for connecting. This factorization has the benefit of significantly lowering the computing time. An improved module with an inverted residual structure has been added to MobileNetV2 [<xref ref-type="bibr" rid="ref-23">23</xref>]. The first layer is 1 &#x00D7; 1 convolution with Rectified Linear Unit (ReLU). The depthwise convolution is the second layer. Another 1 &#x00D7; 1 convolution is used in the third layer with no non-linearity. Deep networks will only have the power of a linear classifier on the non-zero volume output domain if ReLU is applied again.</p>
<p>The projection layer, a pointwise convolutional layer in V2, converts the data with many channels into a tensor with much fewer channels. Before going into the depth-wise convolutional layer, a 1 &#x00D7; 1 expansion of the convolutional layer extends the number of channels based on the expansion factor in the results. Each output of the block is a holdup in the bottleneck residual block. The residual connection exists to aid gradient flow through the network. Batch normalization and the ReLU6 activation features are included in each layer of MobileNetV2.</p>
<p>MobileNet V2 architecture includes 17 bottleneck residual blocks in a row, a standard 1 &#x00D7; 1 convolution, and a classification layer. The pre-trained MobileNetV2 model was adapted for the cephalopod classification task in the proposed work, and its architecture is shown in <xref ref-type="fig" rid="fig-2">Fig. 2</xref>. In the final classification section, fully connected dense layers with dropouts are added accordingly to achieve higher accuracy of classification. A softmax layer, an output classification layer, replaces the final classification layers of the model. The proposed RAdamGC optimizer is used in the fine-tuned architecture, and it is explained in Section 3.4 to increase the training speed of the network. The learning rate factor is adjusted in fully connected layers to expedite learning in the new ending layers. Training parameters such as learning rate, maximum epochs, mini-batch size, and validation data are being set to train the network using the cephalopod data.</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>Fine-tunedMobileNetV2 model</title></caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_30017-fig-2.png"/>
</fig>
</sec>
<sec id="s2_3_2">
<label>2.3.2</label>
<title>Fine Tuned NASNet Mobile Architecture</title>
<p>The brain team of google has created the Neural Architecture Search Network (NASNet), which has two main cell functions; (1) Normal cell and (2) Reduction Cell. NASNet [<xref ref-type="bibr" rid="ref-24">24</xref>] has performed its operations on a small dataset before transferring its block to a larger dataset to achieve a higher mean Average Precision (mAP). The output of NASNet is enhanced by a tweaked drop path called Scheduled drop path for successful regularization. In the first NASNet, the number of cells in the architecture is not pre-determined. Normal and reduction cells are used in particular. The size of the feature map is described by normal cells, and the reduction cell returns the reduced feature map in terms of height and width by a factor of two.</p>
<p>The controlled architecture of NASNet, which is built on a Recurrent Neural Network (RNN), predicts the entire structure of the network based on the two initial hidden states. The Controller Architecture employs an RNN-based LSTM model, with softmax prediction for convolutional cells prediction and recursive construction of network motifs. In NASNet Mobile, the input image size used is 224 &#x00D7; 224. For the transfer learning process, NASNetMobile employs ImageNet pre-trained network weights to classify the cephalopods. The classification layers of the pre-trained NASNet Mobile model are fine-tuned as shown in <xref ref-type="fig" rid="fig-3">Fig. 3</xref> for the cephalopod classification task. The model is fine-tuned with three fully connected layers with dropouts, a softmax layer, and an output classification layer to achieve a high categorization rate. The learning rate factor with fully connected layers are increased to accelerate the learning in the new final layers with the proposed optimizer. The performance of the model is evaluated by tweaking higher parameters.</p>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>Fine-tuned NASNet Mobile model</title></caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_30017-fig-3.png"/>
</fig>
</sec>
<sec id="s2_3_3">
<label>2.3.3</label>
<title>Fine Tuned InceptionV3 Architecture</title>
<p>The advent of Inception models inhibits the stacking layer mode, allowing us to expand the depth and width in a new way. Because the variety of modules in an inception model is more complex to design than a Visual Geometry Group (VGG). On the other hand, inception models can be built more extensively and broadly under the same computational budget, resulting in greater accuracy than a VGG. The Inception paradigm has gone through four generations; Inception-V1, Inception-V2, Inception-V3, and Inception-V4. Inception-V2 replaces the 5 &#x00D7; 5 convolutional layers of Inception v1with two 3 &#x00D7; 3 convolutional layers in a row. Asymmetric convolutions are used in Inception-V3 and Inception-V4 to expand the depth, although the former has fewer model parameters. The classification layers of the pre-trained InceptionV3 model have been fine-tuned for our cephalopod classification task. The fully connected layers with dropouts, a softmax layer, and an output classification layer replace the final layers of the model. The learning rate factor with fully connected layers is varied to accelerate the learning in the new final layers. The performance is observed by tweaking the other hyperparameters.</p>
</sec>
<sec id="s2_3_4">
<label>2.3.4</label>
<title>DenseNet201</title>
<p>A feed-forward network, the Dense Convolutional Network (DenseNet), in which each layer is linked to every other layer. Direct connections L(L &#x002B; 1)/2 are used in this network. All the feature maps of previous levels are used as inputs into each layer, and the feature maps of the layer are being used as inputs in all the following layers. DenseNets have several compelling advantages, including eliminating the vanishing-gradient problem, improved feature propagation, reuse of features, and a considerable reduction in the number of parameters. For Cephalopod classification, DenseNet201 is used. The DenseNet201 is a 201-layer deep CNN. It is a member of the DenseNet family of image classification models.</p>
</sec>
<sec id="s2_3_5">
<label>2.3.5</label>
<title>Xception</title>
<p>Xception is a more advanced variant of the Inception model that independently performs channel-wise and cross-channel convolutions. The primary operator in Xception is a separable convolution, which significantly reduces the parameters of regular convolutions. Xception uses a smaller window (3 &#x00D7; 3) for channel-wise convolution, which is similar to VGG networks. 1 &#x00D7; 1 standard convolutions are performed for cross-channel convolution. Inception models outperform the other types of networks in the classification of accuracy because of the multi-scale processing modules within them.</p>
</sec>
<sec id="s2_3_6">
<label>2.3.6</label>
<title>ResNet50</title>
<p>The ResNet-50 model and aggregation of ResNet models of various depths gained recognition for image classification. ResNet-50 is a residual network of 50 layers that aims to resolve the issue of vanishing gradients in CNN during back-propagation. Increasing the depth of the network can enhance its accuracy if over-fitting is taken into consideration. As the depth of the network rises, the signal needed to modify the weights, which originates from comparing the ground truth and prediction (seen <italic>vs</italic>. forecasted) at the end of the network, becomes extremely tiny at the early layers. Essentially, it implies that the layers before it is nearly unlearned. The ResNet-50 model relies heavily on the convolution blocks. Multiple filters are applied on these networks to distinguish the images, and the filters are passed over the original image of the strides. The values from the learned filters are multiplied by the values from the images. The outputs of the filters are pooled, thereby down sampling while preserving the essential characteristics.</p>
</sec>
<sec id="s2_3_7">
<label>2.3.7</label>
<title>Proposed Optimization Approach- Rectified Adam with Gradient Centralisation (RAdamGC)</title>
<p>Rectified Adam, often known as Adam, is a stochastic optimizer variation that adds a term to correct the variance of adaptive learning rate. In order to improve the accuracy in cephalopod classification, the optimizer Rectified Adam proposed in [<xref ref-type="bibr" rid="ref-25">25</xref>] is adopted and fine-tuned.</p>
<p>The values of <inline-formula id="ieqn-1">
<mml:math id="mml-ieqn-1"><mml:mi>&#x03D5;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mo>.</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math>
</inline-formula> and <inline-formula id="ieqn-2">
<mml:math id="mml-ieqn-2"><mml:mi>&#x03C8;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mo>.</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math>
</inline-formula>, may be changed to provide different optimization techniques, where <inline-formula id="ieqn-3">
<mml:math id="mml-ieqn-3"><mml:mi>&#x03D5;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mo>.</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math>
</inline-formula> indicates how the momentum is determined at time step t and <inline-formula id="ieqn-4">
<mml:math id="mml-ieqn-4"><mml:mi>&#x03C8;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mo>.</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math>
</inline-formula> specifies how the adaptive rate of learning is calculated. For example, in the Adam algorithm,</p>
<p><disp-formula id="eqn-1"><label>(1)</label>
<mml:math id="mml-eqn-1" display="block"><mml:mi>&#x03D5;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>&#x03B2;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mi>&#x03B2;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:mfrac></mml:mrow><mml:mspace width="thickmathspace" /><mml:mrow><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">n</mml:mi><mml:mi mathvariant="normal">d</mml:mi></mml:mrow><mml:mspace width="thickmathspace" /><mml:mi>&#x03C8;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:msqrt><mml:mfrac><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mi>&#x03B2;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>&#x03B2;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>g</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:mfrac></mml:msqrt></mml:mrow></mml:math>
</disp-formula></p>
<p>The function <inline-formula id="ieqn-5">
<mml:math id="mml-ieqn-5"><mml:mi>&#x03C8;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mo>.</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math>
</inline-formula> in <xref ref-type="disp-formula" rid="eqn-1">Eq. (1)</xref> is commonly computed as in <xref ref-type="disp-formula" rid="eqn-2">Eq. (2)</xref>.</p>
<p><disp-formula id="eqn-2"><label>(2)</label>
<mml:math id="mml-eqn-2" display="block"><mml:mrow><mml:mover><mml:mi>&#x03C8;</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:msqrt><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:msqrt></mml:mrow><mml:mrow><mml:mrow><mml:mi>&#x03F5;</mml:mi></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:msqrt><mml:mo stretchy="false">(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>&#x03B2;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>t</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>g</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:msqrt></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:math>
</disp-formula></p>
<p>In generic adaptive optimisation method, calculation of gradients with respect to stochastic objective at time step t, momentum, adaptive learning rate, updating parameters are given as follows (<xref ref-type="disp-formula" rid="eqn-3">Eqs. (3)</xref>&#x2013;<xref ref-type="disp-formula" rid="eqn-6">(6)</xref>).</p>
<p><disp-formula id="eqn-3"><label>(3)</label>
<mml:math id="mml-eqn-3" display="block"><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mi>&#x03B8;</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>f</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math>
</disp-formula></p>
<p><disp-formula id="eqn-4"><label>(4)</label>
<mml:math id="mml-eqn-4" display="block"><mml:mrow><mml:msub><mml:mi>m</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03D5;</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math>
</disp-formula></p>
<p><disp-formula id="eqn-5"><label>(5)</label>
<mml:math id="mml-eqn-5" display="block"><mml:mrow><mml:msub><mml:mi>l</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03C8;</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math>
</disp-formula></p>
<p><disp-formula id="eqn-6"><label>(6)</label>
<mml:math id="mml-eqn-6" display="block"><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>m</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>l</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:mspace width="thickmathspace" /><mml:mspace width="thickmathspace" /><mml:mspace width="thickmathspace" /><mml:mspace width="thickmathspace" /></mml:mrow></mml:math>
</disp-formula></p>
<p>In real-world applications, the Exponential Moving Average (EMA) can be interpreted as a close approximation to the Simple Moving Average (SMA), which is given in <xref ref-type="disp-formula" rid="eqn-7">Eq. (7)</xref>.</p>
<p><disp-formula id="eqn-7"><label>(7)</label>
<mml:math id="mml-eqn-7" display="block"><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mfrac><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>&#x03B2;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>g</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mi>&#x03B2;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:mfrac></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x2248;</mml:mo><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>t</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>g</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>t</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mfrac></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math>
</disp-formula></p>
<p>where <inline-formula id="ieqn-6">
<mml:math id="mml-ieqn-6"><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math>
</inline-formula> is the length of SMA, allowing the SMA and the EMA to share the same &#x201C;center of mass.&#x201D; In other words, <inline-formula id="ieqn-7">
<mml:math id="mml-ieqn-7"><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math>
</inline-formula> satisfies the following condition which is given in <xref ref-type="disp-formula" rid="eqn-8">Eq. (8)</xref>.</p>
<p><disp-formula id="eqn-8"><label>(8)</label>
<mml:math id="mml-eqn-8" display="block"><mml:mrow><mml:mrow><mml:mrow><mml:mfrac><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>&#x03B2;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x22C5;</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mi>&#x03B2;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:mfrac></mml:mrow></mml:mrow></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mrow><mml:mfrac><mml:mrow><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>t</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mi>i</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>t</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mfrac></mml:mrow></mml:mrow></mml:math>
</disp-formula></p>
<p>By solving <xref ref-type="disp-formula" rid="eqn-8">Eq. (8)</xref>, we get the <xref ref-type="disp-formula" rid="eqn-9">Eq. (9)</xref> as follows:</p>
<p><disp-formula id="eqn-9"><label>(9)</label>
<mml:math id="mml-eqn-9" display="block"><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mn>2</mml:mn><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:mfrac></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:mi>t</mml:mi><mml:msubsup><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn><mml:mi>t</mml:mi></mml:msubsup></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn><mml:mi>t</mml:mi></mml:msubsup></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:mstyle></mml:math>
</disp-formula></p>
<p>For ease of notation, <inline-formula id="ieqn-8">
<mml:math id="mml-ieqn-8"><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math>
</inline-formula> is marked as <inline-formula id="ieqn-9">
<mml:math id="mml-ieqn-9"><mml:mrow><mml:msub><mml:mi>&#x03C1;</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula>. Also, we refer <inline-formula id="ieqn-10">
<mml:math id="mml-ieqn-10"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mn>2</mml:mn><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:mfrac></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:mrow><mml:mspace width="thickmathspace" /><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">s</mml:mi><mml:mspace width="thickmathspace" /></mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03C1;</mml:mi><mml:mi mathvariant="normal">&#x221E;</mml:mi></mml:msub></mml:mrow></mml:mstyle></mml:math>
</inline-formula> (maximum length of the approximated SMA), due to the inequality <inline-formula id="ieqn-10a">
<mml:math id="mml-ieqn-10a"><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2264;</mml:mo><mml:munder><mml:mrow><mml:mrow><mml:mi mathvariant="normal">l</mml:mi><mml:mi mathvariant="normal">i</mml:mi><mml:mi mathvariant="normal">m</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mi mathvariant="normal">&#x221E;</mml:mi></mml:mrow></mml:munder><mml:mo>&#x2061;</mml:mo><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mn>2</mml:mn><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:mfrac></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mstyle></mml:math>
</inline-formula></p>
<p>The rectification term can be calculated as in <xref ref-type="disp-formula" rid="eqn-10">Eq. (10)</xref>:</p>
<p><disp-formula id="eqn-10"><label>(10)</label>
<mml:math id="mml-eqn-10" display="block"><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow><mml:mo>=</mml:mo><mml:msqrt><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03C1;</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>4</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03C1;</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03C1;</mml:mi><mml:mi mathvariant="normal">&#x221E;</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03C1;</mml:mi><mml:mi mathvariant="normal">&#x221E;</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>4</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03C1;</mml:mi><mml:mi mathvariant="normal">&#x221E;</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03C1;</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:msqrt></mml:math>
</disp-formula></p>
<p>Gradient Centralization (GC) is a technique [<xref ref-type="bibr" rid="ref-26">26</xref>] for bringing gradient vectors to a zero mean. GC improves the gradient of the loss function, enhancing the efficiency and consistency of the training process.</p>
<p>It is assumed that the gradient of an FC or convolution layer via propagating backwards is obtained, then, for a weight vector wi whose gradient is <inline-formula id="ieqn-11">
<mml:math id="mml-ieqn-11"><mml:mrow><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="script">L</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x2026;</mml:mo><mml:mo>,</mml:mo><mml:mi mathvariant="bold-italic">N</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math>
</inline-formula>, the GC operator, denoted by <inline-formula id="ieqn-12">
<mml:math id="mml-ieqn-12"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="normal">&#x03A6;</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">G</mml:mi><mml:mi mathvariant="bold-italic">C</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:math>
</inline-formula>, is given in <xref ref-type="disp-formula" rid="eqn-11">Eq. (11)</xref>.</p>
<p><disp-formula id="eqn-11"><label>(11)</label>
<mml:math id="mml-eqn-11" display="block"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="normal">&#x03A6;</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">G</mml:mi><mml:mi mathvariant="bold-italic">C</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="script">L</mml:mi></mml:mrow></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="script">L</mml:mi></mml:mrow></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03BC;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="script">L</mml:mi></mml:mrow></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:math>
</disp-formula></p>
<p>where <inline-formula id="ieqn-13">

</inline-formula> represent the gradient of <inline-formula id="ieqn-14">
<mml:math id="mml-ieqn-14"><mml:mrow><mml:mrow><mml:mi mathvariant="script">L</mml:mi></mml:mrow></mml:mrow></mml:math>
</inline-formula> w.r.t. the weight matrix W and weight vector wi, respectively.</p>
<p>The GC formula is pretty straight forward. The mean of the weight matrix&#x0027;s column vectors must be computed, and then the mean must be removed from each column vector.</p>
<p>In optimising the cephalopod classification, GC is embedded with RAdam to boost the performance of the network model and to reduce the training time extensively. It is proposed to have an algorithm RAdamGC to raise the performance of the model and save time for training. To embed GC with RAdam, the parameter updates are carried out as shown in <xref ref-type="disp-formula" rid="eqn-12">Eqs. (12)</xref>&#x2013;<xref ref-type="disp-formula" rid="eqn-16">(16)</xref> and embedded with RAdam algorithm.</p>
<p><disp-formula id="eqn-12"><label>(12)</label>
<mml:math id="mml-eqn-12" display="block"><mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">g</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mi mathvariant="bold-italic">&#x03B8;</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">f</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03B8;</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">t</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mspace width="thickmathspace" /></mml:math>
</disp-formula></p>
<p><disp-formula id="eqn-13"><label>(13)</label>
<mml:math id="mml-eqn-13" display="block"><mml:mrow><mml:mover><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">g</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">&#x2205;</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">G</mml:mi><mml:mi mathvariant="bold-italic">C</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">g</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math>
</disp-formula></p>
<p><disp-formula id="eqn-14"><label>(14)</label>
<mml:math id="mml-eqn-14" display="block"><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">t</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mover><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">g</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow></mml:math>
</disp-formula></p>
<p><disp-formula id="eqn-15"><label>(15)</label>
<mml:math id="mml-eqn-15" display="block"><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">v</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">v</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">t</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:mspace width="thickmathspace" /></mml:mrow></mml:msub></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mover><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">g</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mo mathvariant="bold">&#x2299;</mml:mo></mml:mrow><mml:mrow><mml:mover><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">g</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow></mml:math>
</disp-formula></p>
<p><disp-formula id="eqn-16"><label>(16)</label>
<mml:math id="mml-eqn-16" display="block"><mml:mrow><mml:mover><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mn>1</mml:mn><mml:mrow><mml:mi mathvariant="bold-italic">t</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:math>
</disp-formula></p>
<p>With the equation, <inline-formula id="ieqn-15">
<mml:math id="mml-ieqn-15"><mml:mrow><mml:mover><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">g</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">&#x2205;</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">G</mml:mi><mml:mi mathvariant="bold-italic">C</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">g</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math>
</inline-formula> gradient centralisation is applied for calculating the gradients at timestamp t with respect to stochastic objective. Then exponential moving 1st moment <inline-formula id="ieqn-16">
<mml:math id="mml-ieqn-16"><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">t</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mover><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">g</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow></mml:math>
</inline-formula> is calculated with GC. Bias corrected moving average is calculated with <inline-formula id="ieqn-17">
<mml:math id="mml-ieqn-17"><mml:mrow><mml:mover><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mn>1</mml:mn><mml:mrow><mml:mi mathvariant="bold-italic">t</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math>
</inline-formula>. In the classification task, the proposed algorithm is utilized using initial values of <inline-formula id="ieqn-18">
<mml:math id="mml-ieqn-18"><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03C1;</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mn>7</mml:mn><mml:mspace width="thickmathspace" /></mml:math>
</inline-formula> for MobileNetV2 and <inline-formula id="ieqn-19">
<mml:math id="mml-ieqn-19"><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03C1;</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mn>6</mml:mn></mml:math>
</inline-formula> for InceptionV3 and NASNet Mobile. For MobileNetV2, InceptionV3, and NASNet Mobile, the initial learning rate is 1e&#x2212;8, 1e&#x2212;6, and le&#x2212;5, respectively.</p>
</sec>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Experimentation and Results</title>
<p>The experiments were carried out in Google Colab Pro, a computational tool that included a Jupyter notebook environment with 25 GB RAM and a GPU interface. Before being applied to pre-trained models, the images were resized to 224 &#x00D7; 224 pixels. Image augmentation was performed while training the model with the Image Data Generator class available in the Keras deep learning neural network library. Initially, the lightweight models namely MobileNetV2, NASNet Mobile, Densenet201, Xception, InceptionV3, ResNet50, were considered for evaluation. Each model was under 100 MB and was considered to be a lightweight model.</p>
<sec id="s3_1">
<label>3.1</label>
<title>Evaluation of Pre-Trained Models</title>
<p>Adaptive Moment Estimation is a technique for optimizing the gradient descent algorithms. The models were evaluated for the pre-processed dataset. The pre-trained models with a dense layer at their output were trained with an Adam optimizer. Beta one and beta two values were set to values 0.9 and 0.999 correspondingly. The learning rate used with Adam was 0.001. The experiment had used several epochs up to 100. Of the whole cephalopod images, 70% of the total images were used for training, 15%were used for validation, and 15% for testing.</p>
<p>Six different pre-trained models that were considered in the work were evaluated for the classification of accuracy. From the results shown in <xref ref-type="fig" rid="fig-4">Fig. 4</xref>, it can be observed that InceptionV3 gives the highest accuracy of 82.69%. NASNet Mobile and DenseNet201 have provided an accuracy of 82.05%. The very lightweight model MobileNetV2 has an accuracy value of only 80.77%.</p>
<fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>Classification accuracy of the models</title></caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_30017-fig-4.png"/>
</fig>
<p>Based on the accuracy and size of the models, three are chosen for fine-tuning from the results above. When the outputs of Xception and ResNet 50 are observed, both of them provide less than 80% of accuracy. Despite its 92 MB size, Inception V3 is selected for further fine-tuning to assess the accuracy. Following that, the NASNet Mobile is selected since it is, exceptionally light in weight. Due to its tiny size compared to the other models, MobileNetV2 is also picked for additional fine-tuning.</p>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Evaluation of Fine Tuned Pre-Trained Models</title>
<p>The evaluation of model is performed by varying dense layers and by employing fine-tuned optimisers to reduce the loss and to improve the performance of the models</p>
<sec id="s3_2_1">
<label>3.2.1</label>
<title>Effect on Adding Additional Dense Layers</title>
<p>An improved classification model with dense layers is used to improve the accuracy of classification. Initially, the optimizer Adam is used with the default learning rate. According to the results of the fine-tuned models, there is a significant improvement in the inaccuracy. MobileNetV2 has achieved an accuracy of 84.62%, while the other lightweight model NASNet Mobile has achieved only 84.08%. The classification accuracy of InceptionV3 is 84.61%. <xref ref-type="fig" rid="fig-5">Fig. 5</xref> compares the models that have and do not have fine-tuning. Fine-tuning has improved the performance of all models.</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>Comparison of models with and without fine tuning (FT)</title></caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_30017-fig-5.png"/>
</fig>
<p>The performance of MobileNetV2 has significantly improved since the models are fine-tuned. The accuracy of the models NASNET Mobile and InceptionV3 fine-tuned with dense layers have increased by 2%. The accuracy of MobileNetV2 has increased by 4%.</p>
</sec>
<sec id="s3_2_2">
<label>3.2.2</label>
<title>Effect of Learning Rate (Lr)</title>
<p>In an optimization algorithm, the Lr is a tweaking parameter that controls the step size at each iteration while progressing towards a minimum of loss function. A higher learning rate causes the loss function to rise, whereas a slower learning rate causes the loss function to decline progressively.</p>
<p>The best learning rate must be determined to minimize the loss function in the cephalopod classification task. To assess the performance of the model, the models are now trained using learning rates of 1e&#x2212;1, 1e&#x2212;2, 1e&#x2212;3, 1e&#x2212;4, and 1e&#x2212;5. <xref ref-type="table" rid="table-1">Tab. 1</xref> gives the accuracy values when the model is tested with varied learning rates. The MobileNetV2 model performs the best for the Lr of 1e&#x2212;3. It gives an accuracy of 84.71% with Lr of 1e&#x2212;3. NASNet Mobile gives the accuracy of 84.62% with the Lr of 1e&#x2212;5. With the Lr, 1e&#x2212;4, InceptionV3 gives the accuracy of 84.63%.</p>
<table-wrap id="table-1"><label>Table 1</label>
<caption>
<title>Variation of accuracy with learning rate. Classification accuracy (%)</title></caption>
<table><colgroup>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th style="background:#FFFFFF;">Learning rate (Lr)</th>
<th style="background:#FFFFFF;">MobileNetV2</th>
<th style="background:#FFFFFF;">NASNet mobile</th>
<th style="background:#FFFFFF;">InceptionV3</th>
</tr>
</thead>
<tbody>
<tr>
<td style="background:#FFFFFF;">Lr &#x003D; 1e&#x2212;1</td>
<td style="background:#FFFFFF;">25.64</td>
<td style="background:#FFFFFF;">26.62</td>
<td style="background:#FFFFFF;">25.61</td>
</tr>
<tr>
<td style="background:#FFFFFF;">Lr &#x003D; 1e&#x2212;2</td>
<td style="background:#FFFFFF;">25.64</td>
<td style="background:#FFFFFF;">71.79</td>
<td style="background:#FFFFFF;">53.85</td>
</tr>
<tr>
<td style="background:#FFFFFF;">Lr &#x003D; 1e&#x2212;3</td>
<td style="background:#FFFFFF;">84.71</td>
<td style="background:#FFFFFF;">82.05</td>
<td style="background:#FFFFFF;">84.62</td>
</tr>
<tr>
<td style="background:#FFFFFF;">Lr &#x003D; 1e&#x2212;4</td>
<td style="background:#FFFFFF;">84.62</td>
<td style="background:#FFFFFF;">84.61</td>
<td style="background:#FFFFFF;">84.63</td>
</tr>
<tr>
<td style="background:#FFFFFF;">Lr &#x003D; 1e&#x2212;5</td>
<td style="background:#FFFFFF;">84.61</td>
<td style="background:#FFFFFF;">84.62</td>
<td style="background:#FFFFFF;">84.62</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Effect of Optimiser-Improved Rectified Adam</title>
<p>When the approximated SMA length is less than or equal to 4, the variance of adaptive learning rate is intractable, and the adaptive learning rate is inactivated. Otherwise, the variance rectification term is calculated, and the adaptive learning rate is employed to update the parameters. In order to enhance the performance, the fine-tuned models are tested with different values. As shown in <xref ref-type="table" rid="table-2">Tab. 2</xref>, the variation has significant improvement in the performance of the models.</p>
<table-wrap id="table-2"><label>Table 2</label>
<caption>
<title>Variation of accuracy with <inline-formula id="ieqn-20">
<mml:math id="mml-ieqn-20"><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03C1;</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula></title></caption>
<table><colgroup>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th style="background:#FFFFFF;">Length of SMA, <inline-formula id="ieqn-21">
<mml:math id="mml-ieqn-21"><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03C1;</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula></th>
<th style="background:#FFFFFF;">MobileNetV2</th>
<th style="background:#FFFFFF;">NASNet mobile</th>
<th style="background:#FFFFFF;">InceptionV3</th>
</tr>
</thead>
<tbody>
<tr>
<td style="background:#FFFFFF;">4</td>
<td style="background:#FFFFFF;">79.49</td>
<td style="background:#FFFFFF;">76.92</td>
<td style="background:#FFFFFF;">84.61</td>
</tr>
<tr>
<td style="background:#FFFFFF;">5</td>
<td style="background:#FFFFFF;">85.58</td>
<td style="background:#FFFFFF;">82.05</td>
<td style="background:#FFFFFF;">84.62</td>
</tr>
<tr>
<td style="background:#FFFFFF;">6</td>
<td style="background:#FFFFFF;">87.14</td>
<td style="background:#FFFFFF;">82.05</td>
<td style="background:#FFFFFF;">87.18</td>
</tr>
<tr>
<td style="background:#FFFFFF;">7</td>
<td style="background:#FFFFFF;">87.18</td>
<td style="background:#FFFFFF;">89.49</td>
<td style="background:#FFFFFF;">84.60</td>
</tr>
<tr>
<td style="background:#FFFFFF;">8</td>
<td style="background:#FFFFFF;">87.15</td>
<td style="background:#FFFFFF;">85.19</td>
<td style="background:#FFFFFF;">82.05</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The accuracy of MobileNetV2 is 87.18% with SMA length of 7. NASNet Mobile produces 89.49% for the SMA value of 7. InceptionV3 gives the accuracy of 87.18% with the SMA length of 6.</p>
<p>In order to find the optimal learning rate, Learning rate (Lr) is varied from 1e&#x2212;4 to 1e&#x2212;8 by setting the optimal values of <inline-formula id="ieqn-22">
<mml:math id="mml-ieqn-22"><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03C1;</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> obtained. MobileNetV2 has attained the maximum accuracy of 89.74% with an Lr of 1e&#x2212;8, as shown in <xref ref-type="table" rid="table-3">Tab. 3</xref>. The performance of NASNet is increased apparently when compared to the Adam and it also gives the accuracy of 89.74%. InceptionV3 has achieved the accuracy of 87.18% with the Lr of 1e&#x2212;4 and 1e&#x2212;5.</p>
<table-wrap id="table-3"><label>Table 3</label>
<caption>
<title>Variation of accuracy with learning rate with RAdam</title></caption>
<table><colgroup>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th style="background:#FFFFFF;">Learning rate (Initial)</th>
<th style="background:#FFFFFF;">MobileNetV2<break/> <inline-formula id="ieqn-23">
<mml:math id="mml-ieqn-23"><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03C1;</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mn>7</mml:mn></mml:math>
</inline-formula></th>
<th style="background:#FFFFFF;">NASNet mobile<break/> <inline-formula id="ieqn-24">
<mml:math id="mml-ieqn-24"><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03C1;</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mn>6</mml:mn></mml:math>
</inline-formula></th>
<th style="background:#FFFFFF;">InceptionV3<break/> <inline-formula id="ieqn-25">
<mml:math id="mml-ieqn-25"><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03C1;</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mn>6</mml:mn></mml:math>
</inline-formula></th>
</tr>
</thead>
<tbody>
<tr>
<td style="background:#FFFFFF;">Lr &#x003D; 1e&#x2212;4</td>
<td style="background:#FFFFFF;">83.35</td>
<td style="background:#FFFFFF;">84.62</td>
<td style="background:#FFFFFF;">87.18</td>
</tr>
<tr>
<td style="background:#FFFFFF;">Lr &#x003D; 1e&#x2212;5</td>
<td style="background:#FFFFFF;">87.18</td>
<td style="background:#FFFFFF;">89.74</td>
<td style="background:#FFFFFF;">87.18</td>
</tr>
<tr>
<td style="background:#FFFFFF;">Lr &#x003D; 1e&#x2212;6</td>
<td style="background:#FFFFFF;">87.2</td>
<td style="background:#FFFFFF;">87.18</td>
<td style="background:#FFFFFF;">84.62</td>
</tr>
<tr>
<td style="background:#FFFFFF;">Lr &#x003D; 1e&#x2212;7</td>
<td style="background:#FFFFFF;">87.21</td>
<td style="background:#FFFFFF;">87.19</td>
<td style="background:#FFFFFF;">84.12</td>
</tr>
<tr>
<td style="background:#FFFFFF;">Lr &#x003D; 1e&#x2212;8</td>
<td style="background:#FFFFFF;">89.74</td>
<td style="background:#FFFFFF;">84.62</td>
<td style="background:#FFFFFF;">82.05</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The performance measure values of various models for categorizing the species are shown in <xref ref-type="table" rid="table-4">Tab. 4</xref>. The MobileNetV2 Model, which has a precision of 1.00 for four different classes, provides the best results. The accuracy of the <italic>Sephia aculeata</italic> class is 0.83 with NASNet Mobile and InceptionV3.</p>
<table-wrap id="table-4"><label>Table 4</label>
<caption>
<title>Performance metric comparison of models in classifying cephalopods</title></caption>
<table><colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th style="background:#FFFFFF;">Model</th>
<th style="background:#FFFFFF;">Species</th>
<th style="background:#FFFFFF;">Precision</th>
<th style="background:#FFFFFF;">Recall</th>
<th style="background:#FFFFFF;">F1-score</th>
</tr>
</thead>
<tbody>
<tr>
<td rowspan="6" style="background:#FFFFFF;">MoblieNetV2</td>
<td style="background:#FFFFFF;"><italic>D. sibogae</italic></td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">0.800000</td>
<td style="background:#FFFFFF;">0.888889</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>S. aculeata</italic></td>
<td style="background:#FFFFFF;">0.714286</td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">0.833333</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>S. pharaonis</italic></td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">0.666667</td>
<td style="background:#FFFFFF;">0.800000</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>S. prashadi</italic></td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">1.000000</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>S. lessoniana</italic></td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">0.800000</td>
<td style="background:#FFFFFF;">0.888889</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>U. duvauceli</italic></td>
<td style="background:#FFFFFF;">0.833333</td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">0.909091</td>
</tr>
<tr>
<td rowspan="6" style="background:#FFFFFF;">NASNet Mobile</td>
<td style="background:#FFFFFF;"><italic>D. sibogae</italic></td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">0.600000</td>
<td style="background:#FFFFFF;">0.750000</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>S. aculeata</italic></td>
<td style="background:#FFFFFF;">0.833333</td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">0.909091</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>S. pharaonis</italic></td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">0.833333</td>
<td style="background:#FFFFFF;">0.909091</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>S. prashadi</italic></td>
<td style="background:#FFFFFF;">0.888889</td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">0.941176</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>S. lessoniana</italic></td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">1.000000</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>U. duvauceli</italic></td>
<td style="background:#FFFFFF;">0.818182</td>
<td style="background:#FFFFFF;">0.900000</td>
<td style="background:#FFFFFF;">0.857143</td>
</tr>
<tr>
<td rowspan="6" style="background:#FFFFFF;">InceptionV3</td>
<td style="background:#FFFFFF;"><italic>D. sibogae</italic></td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">0.600000</td>
<td style="background:#FFFFFF;">0.750000</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>S. aculeata</italic></td>
<td style="background:#FFFFFF;">0.833333</td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">0.909091</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>S. pharaonis</italic></td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">0.666667</td>
<td style="background:#FFFFFF;">0.800000</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>S. prashadi</italic></td>
<td style="background:#FFFFFF;">0.875000</td>
<td style="background:#FFFFFF;">0.875000</td>
<td style="background:#FFFFFF;">0.875000</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>S. lessoniana</italic></td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">1.000000</td>
</tr>
<tr>
<td style="background:#FFFFFF;"><italic>U. duvauceli</italic></td>
<td style="background:#FFFFFF;">0.769231</td>
<td style="background:#FFFFFF;">1.000000</td>
<td style="background:#FFFFFF;">0.869565</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The recall is the measure of the proposed model identifying the cephalopods in a right way. MobileNetV2 gives a reasonable recall rate of 1.00 for the classes <italic>S. aculeata</italic>, <italic>S. Prashadi</italic>, &#x0026; <italic>U. duvauceli</italic>. In comparison with the other models, NASNet Mobile predicts <italic>S. pharaonis</italic> in an appropriate manner. MobileNetV2 has a higher recall rate for <italic>D. sibogae</italic> compared to the other two models.</p>
<p>When compared to the other species, the F1 score shows that the MobileNetV2 model is accurate in classifying <italic>S. prashadi</italic>, while NASNet Mobile and InceptionV3 accurately categorize <italic>S. lessoniana</italic></p>
<p>The performance of various models on a basis of species-by-species is depicted in <xref ref-type="fig" rid="fig-6">Fig. 6</xref>. MobileNetV2 outperforms in predicting the species <italic>S. pharaonis</italic>, <italic>D. sibogae</italic> performs better. <italic>S. pharaonis</italic>, <italic>U. duvauceli,</italic> and <italic>S. aculata</italic> are accurately predicted by NASNet Mobile which outperforms the other models. InceptionV3 predicts <italic>S. lessoniana</italic> in an accurate way when compared to the other models. When it comes to classifying the other species, it falls short.</p>
<fig id="fig-6">
<label>Figure 6</label>
<caption>
<title>Performance of different models on species wise</title></caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_30017-fig-6.png"/>
</fig>
<p>The classification accuracy of all models with and without fine-tuning is shown in <xref ref-type="fig" rid="fig-7">Fig. 7</xref>. After fine-tuning with dense layers and by varying values of parameter change in Rectified Adam, both MobileNetV2 and NASNet Mobile have produced good results.</p>
<fig id="fig-7">
<label>Figure 7</label>
<caption>
<title>Performance comparison of all models (with and without Fine Tuning)</title></caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_30017-fig-7.png"/>
</fig>
<p>There is a significant improvement in Inceptionv3 after fine tuning but it is comparatively less when compared to the other fine-tuned models.</p>
</sec>
<sec id="s3_4">
<label>3.4</label>
<title>Computation Time with RAdamGC</title>
<p>The Proposed RAdamGC optimiser is used with the models MobileNetV2, NASNetMobile, InceptionV3.</p>
<p>The training time with the proposed optimizer is significantly reduced, as seen in <xref ref-type="table" rid="table-5">Tab. 5</xref>. MobileNetV2 requires 301.87 s of training time with RAdamGC and 23.43 s of additional training time with Adam. With Adam, NASNet Mobile takes the training time of 396.95 s without GC, but there is a reduction in time of 37.98 s with GC. InceptionV3 requires significantly less training time of 300.288 s with GC, and it is 97.08 s less when compared to Adam without GC. There is a very slight difference in training accuracy with and without GC in cephalopod classification.</p>
<table-wrap id="table-5"><label>Table 5</label>
<caption>
<title>Computation time for each model with the proposed RAdamGC</title></caption>
<table><colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th style="background:#FFFFFF;">Optimiser</th>
<th colspan="2" style="background:#FFFFFF;">Radam</th>
<th colspan="2" style="background:#FFFFFF;">RAdamGC (Proposed)</th>
</tr>
<tr>
<th style="background:#FFFFFF;">Parameter</th>
<th style="background:#FFFFFF;">Training time<break/>(In s)</th>
<th style="background:#FFFFFF;">Training accuracy</th>
<th style="background:#FFFFFF;">Training time<break/>(In s)</th>
<th style="background:#FFFFFF;">Training accuracy</th>
</tr>
</thead>
<tbody>
<tr>
<td style="background:#FFFFFF;">MobileNetV2</td>
<td style="background:#FFFFFF;">325.301</td>
<td style="background:#FFFFFF;">99.35</td>
<td style="background:#FFFFFF;">301.87</td>
<td style="background:#FFFFFF;">99.25</td>
</tr>
<tr>
<td style="background:#FFFFFF;">NASNet Mobile</td>
<td style="background:#FFFFFF;">396.95</td>
<td style="background:#FFFFFF;">100.0</td>
<td style="background:#FFFFFF;">358.97</td>
<td style="background:#FFFFFF;">99.25</td>
</tr>
<tr>
<td style="background:#FFFFFF;">InceptionV3</td>
<td style="background:#FFFFFF;">393.15</td>
<td style="background:#FFFFFF;">97.08</td>
<td style="background:#FFFFFF;">300.288</td>
<td style="background:#FFFFFF;">96.20</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Discussion</title>
<p>It is evident from the findings that NASNet Mobile, MobileNetV2, and InceptionV3 models had outperformed than the other lightweight models. Fine-tuning of the models with dense layers had given the accuracy of 84.62% with MobileNetV2, 84.08% with NASNet and 84.61% with InceptionV3. The performance of the models was evaluated with Adam by adjusting the learning rates. For learning rates of 1e&#x2212;3 and 1e&#x2212;4, the performance of the models was extraordinary. The improved rectified adam was utilized for training to reduce further loss and improve its performance. The models had performed much better for SMA thresholds of 6 and 7. NASNET mobile and MobileNetV2 had produced an accuracy of 89.74% with a learning rate of 1e-8 and 1e-7, respectively. Fine Tuned MobileNetV2 and NASNET Mobile could be utilized to classify cephalopods with extreme certainty. These networks could also be used as an &#x2018;ON DEVICE NETWORK&#x2019; for the classification of cephalopods with IoT devices or embedded devices. In order to reduce the training time, gradient centralization was embedded with RAdam. The proposed RAdamGC was applied to evaluate the performance of the model. The findings of the study had indicated that there is a much reduction in the training time with RAdamGC in comparision with all three models.</p>
</sec>
<sec id="s5">
<label>5</label>
<title>Conclusions</title>
<p>Without the manual and direct observation of cephalopods by marine scientists, the classification of cephalopod is quite a challenging task. This paper has proposed a system to alleviate the manual and the direct process of species classification. The proposed method enables an IoT device or a mobile application to perform the classification tasks by embedding a suitable lightweight pre-trained network. In this paper, six lightweight models are chosen and evaluated. Furthermore, three models are selected based on their performances and fine-tuned to improve the classification rate. As per the evaluation results of the three models, MobileNetV2 and NASNet Mobile have attained a higher degree of accuracy with 89.74%. Because of their light weight, these models are ideal for IoT and mobile devices. An IoT device embedded with such networks can be installed in boats and ships to identify the cephalopods that live beneath the surface of the water. The approach that is used in this study is the first step towards identifying hundreds or thousands of cephalopods. Future investigation is required to expand the range of situations in which the network is adequate for most of the species. In future, the proposed work can not only be developed to categorize the cephalopods into species but also to locate and count them, as well as to estimate their size using images. Detection and segmentation of cephalopods&#x0027; can also be accomplished using well-known object detection and segmentation techniques.</p>
</sec>
</body>
<back><fn-group>
<fn fn-type="other">
<p><bold>Funding Statement:</bold> The authors received no specific funding for this study.</p>
</fn>
<fn fn-type="conflict">
<p><bold>Conflicts of Interest:</bold> The authors declare that they have no conflicts of interest to report regarding the present study.</p>
</fn>
</fn-group>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>P. G.</given-names> <surname>Rodhouse</surname></string-name></person-group>, &#x201C;<article-title>Role of squid in the southern ocean pelagic ecosystem and the possible consequences of climate change</article-title>,&#x201D; <source>Deep Sea Research Part II: Topical Studies in Oceanography</source>, vol. <volume>95</volume>, pp. <fpage>129</fpage>&#x2013;<lpage>138</lpage>, <year>2013</year>.</mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>W. J.</given-names> <surname>Overholtz</surname></string-name>, <string-name><given-names>L. D.</given-names> <surname>Jacobson</surname></string-name> and <string-name><given-names>J. S.</given-names> <surname>Link</surname></string-name></person-group>, &#x201C;<article-title>An ecosystem approach for assessment advice and biological reference points for the gulf of maine georges bank atlantic herring complex</article-title>,&#x201D; <source>North American Journal of Fisheries Management</source>, vol. <volume>28</volume>, no. <issue>1</issue>, pp. <fpage>247</fpage>&#x2013;<lpage>257</lpage>, <year>2008</year>.</mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M. C.</given-names> <surname>Tyrrell</surname></string-name>, <string-name><given-names>J. S.</given-names> <surname>Link</surname></string-name> and <string-name><given-names>H.</given-names> <surname>Moustahfid</surname></string-name></person-group>, &#x201C;<article-title>The importance of including predation in fish population models: Implications for biological reference points</article-title>,&#x201D; <source>Fisheries Research</source>, vol. <volume>108</volume>, no. <issue>1</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>8</lpage>, <year>2011</year>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Quinteiro</surname></string-name>, <string-name><given-names>C. G.</given-names> <surname>Sotelo</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Rehbein</surname></string-name>, <string-name><given-names>S. E.</given-names> <surname>Pryde</surname></string-name>, <string-name><given-names>I.</given-names> <surname>Medina</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Use of mtDNA direct polymerase chain reaction (PCR) sequencing and PCR restriction fragment length polymorphism methodologies in species identification of canned tuna</article-title>,&#x201D; <source>Journal of Agricultural and Food Chemistry</source>, vol. <volume>46</volume>, no. <issue>4</issue>, pp. <fpage>1662</fpage>&#x2013;<lpage>1669</lpage>, <year>1998</year>.</mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>I. M.</given-names> <surname>Mackie</surname></string-name>, <string-name><given-names>S. E.</given-names> <surname>Pryde</surname></string-name>, <string-name><given-names>C.</given-names> <surname>GonzalesSotelo</surname></string-name>, <string-name><given-names>I.</given-names> <surname>Medina</surname></string-name>, <string-name><given-names>R.</given-names> <surname>P&#x00E9;rezMart&#x0131;n</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Challenges in the identification of species of canned fish</article-title>,&#x201D; <source>Trends in Food Science &#x0026; Technology</source>, vol. <volume>10</volume>, no. <issue>1</issue>, pp. <fpage>9</fpage>&#x2013;<lpage>14</lpage>, <year>1999</year>.</mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>W&#x00E4;ldchen</surname></string-name> and <string-name><given-names>P.</given-names> <surname>M&#x00E4;der</surname></string-name></person-group>, &#x201C;<article-title>Machine learning for image based species identification</article-title>,&#x201D; <source>Methods in Ecology and Evolution</source>, vol. <volume>9</volume>, no. <issue>11</issue>, pp. <fpage>2216</fpage>&#x2013;<lpage>2225</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>W&#x00E4;ldchen</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Rzanny</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Seeland</surname></string-name> and <string-name><given-names>P.</given-names> <surname>M&#x00E4;der</surname></string-name></person-group>, &#x201C;<article-title>Automated plant species identification&#x2014;Trends and future directions</article-title>,&#x201D; <source>PLOS Computational Biology</source>, vol. <volume>14</volume>, no. <issue>4</issue>, pp. <fpage>e1005993</fpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>F.</given-names> <surname>EmmertStreib</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Yang</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Feng</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Tripathi</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Dehmer</surname></string-name></person-group>, &#x201C;<article-title>An introductory review of deep learning for prediction models with big data</article-title>,&#x201D; <source>Frontiers in Artificial Intelligence</source>, vol. <volume>3</volume>, pp. <fpage>4</fpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S. R.</given-names> <surname>Jena</surname></string-name>, <string-name><given-names>S. T.</given-names> <surname>George</surname></string-name> and <string-name><given-names>D. N.</given-names> <surname>Ponraj</surname></string-name></person-group>, &#x201C;<article-title>Modeling an effectual multi-section you only look once for enhancing lung cancer prediction</article-title>,&#x201D; <source>International Journal of Imaging Systems and Technology</source>, vol. <volume>31</volume>, no. <issue>4</issue>, pp. <fpage>2144</fpage>&#x2013;<lpage>2157</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>LeCun</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Bengio</surname></string-name> and <string-name><given-names>G.</given-names> <surname>Hinton</surname></string-name></person-group>, &#x201C;<article-title>Deep learning</article-title>,&#x201D; <source>Nature</source>, vol. <volume>521</volume>, no. <issue>7553</issue>, pp. <fpage>436</fpage>&#x2013;<lpage>444</lpage>, <year>2015</year>.</mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Khan</surname></string-name> and <string-name><given-names>T.</given-names> <surname>Yairi</surname></string-name></person-group>, &#x201C;<article-title>A review on the application of deep learning in system health management</article-title>,&#x201D; <source>Mechanical Systems and Signal Processing</source>, vol. <volume>107</volume>, no. <issue>2</issue>, pp. <fpage>241</fpage>&#x2013;<lpage>265</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Salman</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Jalal</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Shafait</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Mian</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Shortis</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Fish species classification in unconstrained underwater environments based on deep learning</article-title>,&#x201D; <source>Limnology and Oceanography: Methods</source>, vol. <volume>14</volume>, no. <issue>9</issue>, pp. <fpage>570</fpage>&#x2013;<lpage>585</lpage>, <year>2016</year>.</mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S. A.</given-names> <surname>Siddiqui</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Salman</surname></string-name>, <string-name><given-names>M. I.</given-names> <surname>Malik</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Shafait</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Mian</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Automatic fish species classification in underwater videos: Exploiting pre-trained deep neural network models to compensate for limited labelled data</article-title>,&#x201D; <source>ICES Journal of Marine Science</source>, vol. <volume>75</volume>, no. <issue>1</issue>, pp. <fpage>374</fpage>&#x2013;<lpage>389</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>V.</given-names> <surname>Allken</surname></string-name>, <string-name><given-names>N. O.</given-names> <surname>Handegard</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Rosen</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Schreyeck</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Mahiout</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Fish species identification using a convolutional neural network trained on synthetic data</article-title>,&#x201D; <source>ICES Journal of Marine Science</source>, vol. <volume>76</volume>, no. <issue>1</issue>, pp. <fpage>342</fpage>&#x2013;<lpage>349</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Mathur</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Vasudev</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Sahoo</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Jain</surname></string-name> and <string-name><given-names>N.</given-names> <surname>Goel</surname></string-name></person-group>, &#x201C;<article-title>Crosspooled fishnet: Transfer learning based fish species classification model</article-title>,&#x201D; <source>Multimedia Tools and Applications</source>, vol. <volume>79</volume>, no. <issue>41</issue>, pp. <fpage>31625</fpage>&#x2013;<lpage>31643</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>X.</given-names> <surname>Xu</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Li</surname></string-name> and <string-name><given-names>Q.</given-names> <surname>Duan</surname></string-name></person-group>, &#x201C;<article-title>Transfer learning and SE-ResNet152 networks-based for small-scale unbalanced fish species identification</article-title>,&#x201D; <source>Computers and Electronics in Agriculture</source>, vol. <volume>180</volume>, pp. <fpage>105878</fpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>J. A.</given-names> <surname>Jose</surname></string-name>, <string-name><given-names>C. S.</given-names> <surname>Kumar</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Sureshkumar</surname></string-name></person-group>, &#x201C;<article-title>Tuna classification using super learner ensemble of region based CNN-grouped 2D-LBP models</article-title>,&#x201D; <source>Information Processing in Agriculture</source>, vol. <volume>9</volume>, no. <issue>1</issue>, pp. <fpage>68</fpage>&#x2013;<lpage>79</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M. A.</given-names> <surname>Iqbal</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>Z. A.</given-names> <surname>Ali</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Riaz</surname></string-name></person-group>, &#x201C;<article-title>Automatic fish species classification using deep convolutional neural networks</article-title>,&#x201D; <source>Wireless Personal Communications</source>, vol. <volume>116</volume>, no. <issue>2</issue>, pp. <fpage>1043</fpage>&#x2013;<lpage>1053</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A. S.</given-names> <surname>Winoto</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Kristianus</surname></string-name> and <string-name><given-names>C.</given-names> <surname>Premachandra</surname></string-name></person-group>, &#x201C;<article-title>Small and slim deep convolutional neural network for mobile device</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>8</volume>, pp. <fpage>125210</fpage>&#x2013;<lpage>125222</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>X.</given-names> <surname>Liu</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Jia</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Hou</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Fu</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Ma</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Realtime marine animal images classification by embedded system based on mobilenet and transfer learning</article-title>,&#x201D; <source>OCEANS 2019-Marseille</source>, vol. <volume>2019</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>5</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Liu</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Cao</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Li</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Xiao</surname></string-name>, <string-name><given-names>Q.</given-names> <surname>Qiu</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Automatic diagnosis of fungal keratitis using data augmentation and image fusion with deep convolutional neural network</article-title>,&#x201D; <source>Computer Methods and Programs in Biomedicine</source>, vol. <volume>187</volume>, no. <issue>11</issue>, pp. <fpage>105019</fpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>D. M.</given-names> <surname>Montserrat</surname></string-name>, <string-name><given-names>Q.</given-names> <surname>Lin</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Allebach</surname></string-name> and <string-name><given-names>E. J.</given-names> <surname>Delp</surname></string-name></person-group>, &#x201C;<article-title>Training object detection and recognition CNN models using data augmentation</article-title>,&#x201D; <source>Journal Electronic Imaging</source>, vol. <volume>10</volume>, no. <issue>10</issue>, pp. <fpage>27</fpage>&#x2013;<lpage>36</lpage>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Sandler</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Howard</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Zhu</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Zhmoginov</surname></string-name> and <string-name><given-names>L. C.</given-names> <surname>Chen</surname></string-name></person-group>, &#x201C;<article-title>Mobilenetv2: Inverted residuals and linear bottlenecks</article-title>,&#x201D; in <conf-name>Proc. IEEE Conf. on Computer Vision and Pattern Recognition</conf-name>, <publisher-loc>San Juan, USA</publisher-loc>, pp. <fpage>4510</fpage>&#x2013;<lpage>4520</lpage>, <year>2018</year>. </mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>B.</given-names> <surname>Zoph</surname></string-name>, <string-name><given-names>V.</given-names> <surname>Vasudevan</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Shlens</surname></string-name> and <string-name><given-names>Q. V.</given-names> <surname>Le</surname></string-name></person-group>, &#x201C;<article-title>Learning transferable architectures for scalable image recognition</article-title>,&#x201D; in <conf-name>Proc. IEEE Conf. on Computer Vision and Pattern Recognition</conf-name>, <publisher-loc>San Juan, USA</publisher-loc>, pp. <fpage>8697</fpage>&#x2013;<lpage>8710</lpage>, <year>2018</year>. </mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>L.</given-names> <surname>Liu</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Jiang</surname></string-name>, <string-name><given-names>P.</given-names> <surname>He</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Chen</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Liu</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Gao</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Han</surname></string-name></person-group>, &#x201C;<article-title>On the variance of the adaptive learning rate and beyond</article-title>&#x201D;. <comment>arXiv preprint arXiv:1908.03265, 2019 Aug 8</comment>.</mixed-citation></ref>
<ref id="ref-26"><label>[26]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Yong</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Huang</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Hua</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Zhang</surname></string-name></person-group>, &#x201C;<article-title>Gradient centralization: A new optimization technique for deep neural networks</article-title>,&#x201D; In <conf-name>European Conference on Computer Vision</conf-name>. pp. <fpage>635</fpage>&#x2013;<lpage>652</lpage>, <year>2020</year>.</mixed-citation></ref>
</ref-list>
</back>
</article>