<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CMC</journal-id>
<journal-id journal-id-type="nlm-ta">CMC</journal-id>
<journal-id journal-id-type="publisher-id">CMC</journal-id>
<journal-title-group>
<journal-title>Computers, Materials &#x0026; Continua</journal-title>
</journal-title-group>
<issn pub-type="epub">1546-2226</issn>
<issn pub-type="ppub">1546-2218</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">21449</article-id>
<article-id pub-id-type="doi">10.32604/cmc.2022.021449</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Arabic Fake News Detection Using Deep Learning</article-title>
<alt-title alt-title-type="left-running-head">Arabic Fake News Detection Using Deep Learning</alt-title>
<alt-title alt-title-type="right-running-head">Arabic Fake News Detection Using Deep Learning</alt-title>
</title-group>
<contrib-group content-type="authors">
<contrib id="author-1" contrib-type="author">
<name name-style="western"><surname>Fouad</surname><given-names>Khaled M.</given-names></name><xref ref-type="aff" rid="aff-1">1</xref>
<xref ref-type="aff" rid="aff-3">3</xref>
</contrib>
<contrib id="author-2" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Sabbeh</surname><given-names>Sahar F.</given-names></name><xref ref-type="aff" rid="aff-1">1</xref>
<xref ref-type="aff" rid="aff-2">2</xref><email>sfsabbeh@uj.edu.sa</email>
</contrib>
<contrib id="author-3" contrib-type="author">
<name name-style="western"><surname>Medhat</surname><given-names>Walaa</given-names></name><xref ref-type="aff" rid="aff-1">1</xref>
<xref ref-type="aff" rid="aff-3">3</xref>
</contrib>
<aff id="aff-1"><label>1</label><institution>Faculty of Computers &#x0026; Artificial Intelligence, Benha University</institution>, <country>Egypt</country></aff>
<aff id="aff-2"><label>2</label><institution>University of Jeddah, College of Computer Science and Engineering</institution>, <addr-line>Jeddah, 21493</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff-3"><label>3</label><institution>Information Technology and Computer Science, Nile University</institution>, <country>Egypt</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label>Corresponding Author: Sahar F. Sabbeh. Email: <email>sfsabbeh@uj.edu.sa</email></corresp>
</author-notes>
<pub-date pub-type="epub" date-type="pub" iso-8601-date="2021-11-29"><day>29</day>
<month>11</month>
<year>2021</year></pub-date>
<volume>71</volume>
<issue>2</issue>
<fpage>3647</fpage>
<lpage>3665</lpage>
<history>
<date date-type="received"><day>03</day><month>7</month><year>2021</year></date>
<date date-type="accepted"><day>14</day><month>10</month><year>2021</year></date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2022 Fouad et al.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>Fouad et al.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CMC_21449.pdf"></self-uri>
<abstract>
<p>Nowadays, an unprecedented number of users interact through social media platforms and generate a massive amount of content due to the explosion of online communication. However, because user-generated content is unregulated, it may contain offensive content such as fake news, insults, and harassment phrases. The identification of fake news and rumors and their dissemination on social media has become a critical requirement. They have adverse effects on users, businesses, enterprises, and even political regimes and governments. State of the art has tackled the English language for news and used feature-based algorithms. This paper proposes a model architecture to detect fake news in the Arabic language by using only textual features. Machine learning and deep learning algorithms were used. The deep learning models are used depending on conventional neural nets (CNN), long short-term memory (LSTM), bidirectional LSTM (BiLSTM), CNN&#x002B;LSTM, and CNN &#x002B; BiLSTM. Three datasets were used in the experiments, each containing the textual content of Arabic news articles; one of them is real-life data. The results indicate that the BiLSTM model outperforms the other models regarding accuracy rate when both simple data split and recursive training modes are used in the training process.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Fake news detection</kwd>
<kwd>deep learning</kwd>
<kwd>machine learning</kwd>
<kwd>natural language processing</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1"><label>1</label><title>Introduction</title>
<p>The rise of social networks has considerably changed the way users around the world communicate. Social networks and user-generated content (UGC) are examples of platforms that allow users to generate, share, and exchange their thoughts and opinions via posts, tweets, and comments. Thus, social media platforms (i.e., Twitter, Facebook, etc.) are considered powerful tools through which news and information can be rapidly transmitted and propagated. These platforms empowered their significance to be the essence of information and news source for individuals through the WWW [<xref ref-type="bibr" rid="ref-1">1</xref>]. However, social media and UGC platforms are a double-edged sword. On the one hand, they allow users to share their experiences which enriches the web content. On the other hand, the absence of content supervision may lead to the spread of false information intentionally or unintentionally [<xref ref-type="bibr" rid="ref-2">2</xref>], threatening the reliability of information and news on such platforms.</p>
<p>False information can be classified as intention-based or knowledge-based [<xref ref-type="bibr" rid="ref-3">3</xref>]. Intention-based can be further classified into misinformation and disinformation. Misinformation is an unintentional share of false information based on the user&#x0027;s beliefs, thoughts, and point of view. Whereas the intentional spread of false information to deceive, mislead, and harm users are called disinformation. Fake news is considered disinformation as they include news articles that are confirmed to be false/deceptive published intentionally to mislead people. Another categorization of false information [<xref ref-type="bibr" rid="ref-4">4</xref>] was based on the severity of its impact on users. Based on this study, false information is classified as a) fake news, b) biased/inaccurate news, and c) misleading/ambiguous news. Fake news has the highest impact and uses tools such as content fabrication, propaganda, and conspiracy theories [<xref ref-type="bibr" rid="ref-5">5</xref>]. Finally, biased content is considered to be less dangerous and mainly uses hoaxes and fallacies. The last group is misleading news, which has a minor impact on users. Misleading content usually comes in the forms of rumors, clickbait, and satire/sarcasm news. Disinformation results in biased, deceptive, and decontextualized information based upon which emotional decisions are made, impulsive reactions, or stopping actions in progress. Disinformation results negatively impact users&#x2019; experience and decisions, such as online shopping and stock markets [<xref ref-type="bibr" rid="ref-6">6</xref>].</p>
<p>The bulk of researches for fake news detection are based on machine learning techniques [<xref ref-type="bibr" rid="ref-7">7</xref>]. Those techniques are feature-based, as they require identifying and selecting features that can help identify any piece of information/text&#x0027;s fakeness. Those features are then fed into the chosen machine learning model for classification. In various languages, deep learning models [<xref ref-type="bibr" rid="ref-8">8</xref>] have recently proven efficiency in text classification tasks and fake news detection [<xref ref-type="bibr" rid="ref-9">9</xref>]. They have the advantage that they can automatically adjust their internal parameters until they identify the best features to differentiate between different labels on their own. However, no researches use deep learning models [<xref ref-type="bibr" rid="ref-10">10</xref>] for fake news detection for the Arabic language, as far as we know from the literature.</p>
<p>The problem of Fake news detection can have harmful consequences on social and political life. Detecting fake news is very challenging, mainly when applied in different languages than English. The Arabic language is one of the most spoken languages over the globe. There are a lot of sources for news in Arabic, including official news websites. These sources are considered the primary source of Arabic datasets. Our goal is to detect rumors and measure the effect of fake news detection in the middle east region. We have evaluated many algorithms to achieve the best results.</p>
<p>The work's main objective is exploring and evaluating the performance of different deep learning models in improving fake news detection for the Arabic language. Additionally, compare deep learning performance with the traditional machine learning techniques. Eight machine learning algorithms with cross-fold validation are evaluated, including probabilistic and vector space algorithms. We have also tested five combinations of deep learning algorithms, including CNN and LSTM.</p>
<p>The paper is organized as follows. Section 2 tackles the literature review in some detail. The proposed model architecture is presented in Section 3. Section 4 presents the experiments and the results with discussion. The paper is concluded in Section 5.</p>
</sec>
<sec id="s2"><label>2</label><title>Literature Review</title>
<p>There are many methods used for fake news detection and rumor detection. The methods include machine learning and deep learning algorithms, as illustrated in the following subsections.</p>
<sec id="s2_1"><label>2.1</label><title>Fake News Detection Methods</title>
<p>Fake news detection has been investigated from different perspectives; each utilized different features for information classification. These features included linguistic, visual, user, post, and network-based features [<xref ref-type="bibr" rid="ref-5">5</xref>,<xref ref-type="bibr" rid="ref-11">11</xref>]. The linguistic-based methods tried to find irregular styles within text based on a set of features such as the number of words, word length, multiple words frequencies; unique word count, psycho-linguistic features, syntactic features (i.e., TF-IDF, question marks, exclamation marks, hash-tags &#x2026;etc.) to discriminate natural and fake news [<xref ref-type="bibr" rid="ref-11">11</xref>]. Visual-based systems attempted to identify and extract visual elements from fictitious photos and movies [<xref ref-type="bibr" rid="ref-12">12</xref>] by using deep learning approaches. The user-based methods analyzed user-level features to identify likely fake accounts. It is believed that fake news can probably be created and shared by fake accounts or automatic pots created for this sake. User-based features were used to evaluate source/author credibility. Those features include, among others: the number of tweets, tweet repetition, number of followers, account age, account verifiability, user photo, demographics, user sentiment, topically relevancy, and physical proximity [<xref ref-type="bibr" rid="ref-13">13</xref>,<xref ref-type="bibr" rid="ref-14">14</xref>]. The post-based methods analyzed users&#x2019; feedback, opinions, and reactions as indicators of fake news. These features included comments, opinions, sentiment, user rating, tagging, likes, and emotional reactions [<xref ref-type="bibr" rid="ref-14">14</xref>]. The network-based methods of social networks enabled the rapid spread of fake news. These methods tried to construct and analyze networks from different perspectives. Friendship networks, for instance, explored the user followers relationship. In comparison, stance networks represent post-to-post similarities. Another type is the co-occurrence networks, which evaluate user&#x2013;topic relevancy [<xref ref-type="bibr" rid="ref-14">14</xref>].</p>
<p>Many methods considered rumors or fake news detection as a classification issue. These methods aim to associate attributes&#x2019; values, like a rumor or non-rumor, true or false, or fake or genuine, with a specific piece of text. Researchers had utilized machine-learning methods, accomplishing optimistic results. Substitutionary researchers utilized other methods based on data mining techniques. They depended on extrinsic resources, like knowledge bases, to forecast either the included class of social media content or to examine their truthfulness. Many methods that detect rumors have concentrated on utilizing content features for classification. Few methods for rumor detection have depended on social context. Otherwise, rumor detection and verification methods predominantly utilized a combination of content and context features. Using this combination is since using the social context of rumors may significantly enhance detection performance [<xref ref-type="bibr" rid="ref-14">14</xref>]. Some different method categories of rumor detection that may be considered in the works of rumor detection analysis are shown in <?A3B2 "tbl1",5,"anchor"?><xref ref-type="table" rid="table-1">Tab. 1</xref>. These methods can be categorized into classification methods and other methods.</p>
<table-wrap id="table-1"><label>Table 1</label><caption><title>Different attributes that participate in the start and growth of rumors on social media</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Categories</th>
<th align="left">Methods</th>
<th align="left">Samples</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">User&#x2013;related attributes</td>
<td align="left">Machine learning</td>
<td align="left">SVM, Random Forest, Decision Tree, Logistic Regression, Conditional Random Field (CRF), Hidden Markov Model (HMM).</td>
</tr>
<tr>
<td align="left">Content&#x2013;related attributes</td>
<td align="left">Deep learning</td>
<td align="left">Recurrent Neural Networks (RNN), Convolutional Neural Networks (CNN)</td>
</tr>
<tr>
<td align="left" colspan="2">Other methods</td>
<td align="left">Retweet Behavior, Diffusion Patterns, Anolamy Detection, Hawkes process, Crowdsourcing, Computational Fact-Checking</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s2_2"><label>2.2</label><title>Machine Learning-Based Fake News Detection</title>
<p>Most of the fake news detection works to formulate the problem as a binary classification problem. The literature research may fall under the umbrella of three main classes [<xref ref-type="bibr" rid="ref-15">15</xref>]; feature-based machine learning approaches, networking approaches, and deep learning approaches. Feature-based machine learning approaches aim at learning features from data using feature-engineering techniques before classification. Textual, visual, and user features are extracted and fed into a classifier, then evaluated to identify the best performance given those sets of features. The most widely used supervised learning techniques include logistic regression (LR) [<xref ref-type="bibr" rid="ref-16">16</xref>], ensemble learning techniques such as random forest (RF) and adaptive boosting (Adaboost) [<xref ref-type="bibr" rid="ref-16">16</xref>&#x2013;<xref ref-type="bibr" rid="ref-18">18</xref>], decision trees [<xref ref-type="bibr" rid="ref-18">18</xref>], artificial neural networks [<xref ref-type="bibr" rid="ref-18">18</xref>], support vector machines(SVM) [<xref ref-type="bibr" rid="ref-16">16</xref>,<xref ref-type="bibr" rid="ref-18">18</xref>], na&#x00EF;ve Bayesian (NB) [<xref ref-type="bibr" rid="ref-16">16</xref>,<xref ref-type="bibr" rid="ref-18">18</xref>,<xref ref-type="bibr" rid="ref-19">19</xref>] and k-nearest neighbor (KNN) [<xref ref-type="bibr" rid="ref-16">16</xref>&#x2013;<xref ref-type="bibr" rid="ref-18">18</xref>] and linear discriminant analysis (LDA) [<xref ref-type="bibr" rid="ref-16">16</xref>,<xref ref-type="bibr" rid="ref-20">20</xref>,<xref ref-type="bibr" rid="ref-21">21</xref>]. However, feature-based machine learning models suffer the issue of requiring feature engineering ahead of classification, cumbersome. Networking approaches evaluate user/author credibility by extracting features such as the number of followers, comments/replies content, timestamp, and using graph-based algorithms to analyze structural social networks [<xref ref-type="bibr" rid="ref-22">22</xref>,<xref ref-type="bibr" rid="ref-23">23</xref>].</p>
</sec>
<sec id="s2_3"><label>2.3</label><title>Deep Learning for Fake News Detection</title>
<p>Deep learning (DL) [<xref ref-type="bibr" rid="ref-9">9</xref>,<xref ref-type="bibr" rid="ref-15">15</xref>] approaches use deep learning algorithms to learn features from data without feature engineering during training automatically. Deep learning models have proven a substantial performance improvement and eliminated the need for the feature extraction process. As stated earlier, deep learning models can automatically overcome the burden of feature engineering steps and automatically use training data to identify discriminatory features for fake news detection. Deep learning models showed remarkable performance in text classification general tasks [<xref ref-type="bibr" rid="ref-24">24</xref>&#x2013;<xref ref-type="bibr" rid="ref-27">27</xref>] and have been widely used for fake news detection [<xref ref-type="bibr" rid="ref-28">28</xref>]. Due to their efficiency for text classification tasks, deep learning models have been applied for fake news detection from the NLP perspective using only text. For instance, in [<xref ref-type="bibr" rid="ref-29">29</xref>], deep neural networks were used to predict real/fake news using only the news&#x0027;s textual content. Different machines and deep learning algorithms were applied in this work, and their results showed the superiority of Gated Recurrent Units (GRUs) over other tested methods.</p>
<p>In [<xref ref-type="bibr" rid="ref-30">30</xref>], text content was preprocessed and input to recurrent neural networks (RNN), GRU, Vanilla, and Long Short Term Memories (LSTMs) for classifying fake news. Their results showed that LSTM follows GRU&#x0027;s best performance and finally comes vanilla. Text-based only classification in [<xref ref-type="bibr" rid="ref-31">31</xref>] using tanh-RNNs and LSTM with a single hidden layer, GRU with one hidden layer, and enhanced with an extra GRU hidden layer. In [<xref ref-type="bibr" rid="ref-32">32</xref>], bidirectional LSTMs together with Concurrent neural networks (CNN) were used for classification. The bi-directional LSTMs considered contextual information in both directions: forward and backward in the text. In [<xref ref-type="bibr" rid="ref-33">33</xref>], the feasibility of applying deep learning architecture of CNN with LSTM cells for text-only &#x2013; based classification of RNN and LSTMs and GRUs were used in [<xref ref-type="bibr" rid="ref-34">34</xref>] for text-based classification. Deep CNN multiple hidden layers were used in [<xref ref-type="bibr" rid="ref-35">35</xref>].</p>
<p>Other attempts were made using text and one or more of other features in [<xref ref-type="bibr" rid="ref-36">36</xref>] Arjun et al. utilized ensemble-based CNN and (BiLSTM) for multi-label classification of fake news based on textual content and features related to the source&#x0027;s behavior speaker. In their model, fake news is assigned to one of six classes of fake news. (Pants-fire, False, Barely-true, Half-true, Mostly-true, and True). In [<xref ref-type="bibr" rid="ref-37">37</xref>], neural ensemble architecture (RNN, GRN, and CNN) and used content-based and author-based features to detect rumors on Twitter. Textual and propagation-based features were used in [<xref ref-type="bibr" rid="ref-38">38</xref>,<xref ref-type="bibr" rid="ref-39">39</xref>] for classification: the former used RNN, CNN, and recursive neural networks (RvNN). CNN and RNN used only text and sentiment polarity of tweet's response, and the RvNN model was used on the text and the propagation. At the same time, the latter study constructed RvRNN based on top-down and bottom-up tree structures. These models were tested compared to the traditional state-of-the-art models such as decision tree (Tree-based Ranki DTR, Decision-Tree Classifier DTC), RFC, different variations of SVM, GRU-RNN. Their results showed that TD-RvRNN gave the best performance.</p>
<p>Post-based, together with user-based features, were used for fake news predictions [<xref ref-type="bibr" rid="ref-40">40</xref>]. They applied RNN, LSTMs, and GRUs and found out that LSTMs outperformed the other two models. In [<xref ref-type="bibr" rid="ref-41">41</xref>], text, user-based, content-based features, signal features were used for prediction tasks using a hierarchical recurrent convolutional neural network. Their experiments included (Tree-based Ranki DTR, Decision-Tree Classifier DTC), SVM, GRUs, BLSTMs. <?A3B2 "tbl2",5,"anchor"?><xref ref-type="table" rid="table-2">Tab. 2</xref> summarizes the surveyed works in the literature.</p>
<table-wrap id="table-2"><label>Table 2</label><caption><title>The summarization of the related works</title></caption>
<table frame="hsides" >
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Study</th>
<th align="left">Used features</th>
<th align="left">Experimented models</th>
<th align="left">Dataset</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">Umer<break/> et al. [<xref ref-type="bibr" rid="ref-29">29</xref>]</td>
<td align="left">Only text</td>
<td align="left">Logistic regression<break/>Two-layer FNN<break/>GRUs<break/>Bidirectional RNN with LSTMs<break/>CNN with MaxPooling<break/>Attention-Augmented CNN</td>
<td align="left">size: 63,000 articles.</td>
</tr>
<tr>
<td align="left">Girgis<break/> et al. [<xref ref-type="bibr" rid="ref-30">30</xref>]</td>
<td align="left">Only text</td>
<td align="left">RNN<break/>(vanilla, GRU)<break/>LSTMs.</td>
<td align="left">LIAR Dataset</td>
</tr>
<tr>
<td align="left">Jing et al. [<xref ref-type="bibr" rid="ref-31">31</xref>]</td>
<td align="left">Only text</td>
<td align="left">SVM-TS<break/>RFC<break/>tanh-RNNs (single hidden layer)<break/>LSTM-1 (one hidden layer)<break/>GRU-1 (one-layer)<break/>GRU-2 (extra GRU hidden layer)</td>
<td align="left">Twitter 498 rumors and 494 non-rumors,<break/>Sina Weibo 2,313 rumors and 2,351 non-rumors.</td>
</tr>
<tr>
<td align="left">Muhammad<break/> et al. [<xref ref-type="bibr" rid="ref-32">32</xref>]</td>
<td align="left">Only text</td>
<td align="left">K-nearest neighbors (KNN),<break/>Decision tree (DT),<break/>Random forest (RF),<break/>Logistic regression (LR),<break/>Na&#x00EF;ve Bayes (NB)<break/>LSTM,<break/>CNN,<break/>RNN,<break/>LSTM-CNN.<break/>BiLSTM-CNN</td>
<td align="left">Pheme rumor dataset (5800 tweets)</td>
</tr>
<tr>
<td align="left">Sansiri<break/> et al. [<xref ref-type="bibr" rid="ref-37">37</xref>]</td>
<td align="left">Only text</td>
<td align="left">LSTM<break/>CNN</td>
<td align="left">20,015 news, (11,941 fake 8,074 real news)</td>
</tr>
<tr>
<td align="left">Verma<break/> et al. [<xref ref-type="bibr" rid="ref-34">34</xref>]</td>
<td align="left">Only text</td>
<td align="left">GRU<break/>RRN &#x002B; LSTM</td>
<td align="left">50,000 real news dataset</td>
</tr>
<tr>
<td align="left">Kaliyar<break/> et al. [<xref ref-type="bibr" rid="ref-35">35</xref>]</td>
<td align="left">Only text</td>
<td align="left">CNN</td>
<td align="left">Kaggle dataset<break/>FakeNews Dataset</td>
</tr>
<tr>
<td align="left">Kaliyar<break/> et al. [<xref ref-type="bibr" rid="ref-36">36</xref>]</td>
<td align="left">Text<break/>the behavior of the source speaker.</td>
<td align="left">CNN<break/>BSLTM<break/>MLP</td>
<td align="left">LIAR Dataset</td>
</tr>
<tr>
<td align="left">Sansiri<break/> et al. [<xref ref-type="bibr" rid="ref-37">37</xref>]</td>
<td align="left">Content-based<break/>author-based features</td>
<td align="left">DTC<break/>RFC<break/>SVM<break/>Ensemble RNN<break/>GRU-2<break/>CNN</td>
<td align="left">Real-world dataset (nearly 615000 tweets from 284000 users)</td>
</tr>
<tr>
<td align="left">Pavithra<break/> et al. [<xref ref-type="bibr" rid="ref-38">38</xref>]</td>
<td align="left">Textural<break/>propagation features</td>
<td align="left">CNN<break/>RNN<break/>RvNN</td>
<td align="left">Pheme-RNR (1971 rumor &#x002B; 3819 non-rumor)</td>
</tr>
<tr>
<td align="left">Jing et al. [<xref ref-type="bibr" rid="ref-39">39</xref>]</td>
<td align="left">Text<break/>propagation-based features</td>
<td align="left">DTR,<break/>DTC<break/>RFC,<break/>Different variations of SVM,<break/>GRU-RNN.<break/>BU-RvRNN<break/>TD-RvRNN</td>
<td align="left">Twitter15 (1,381 propagation trees)<break/>Twitter16 (1,181 propagation trees)</td>
</tr>
<tr>
<td align="left">Yichun<break/> et al. [<xref ref-type="bibr" rid="ref-40">40</xref>]</td>
<td align="left">Post-based,<break/>User-based</td>
<td align="left">DLSTM<break/>DGRU<break/>DSRNN</td>
<td align="left">&#x00A0;Sina Weibo, (2313 rumors, 2351 non-rumors)<break/>Twitter dataset 111 events (60 rumors and 51 non-rumors) and 192,350 posts.</td>
</tr>
<tr>
<td align="left">Lin et al. [<xref ref-type="bibr" rid="ref-41">41</xref>]</td>
<td align="left">Text<break/>user-based<break/>content-based<break/>signal features</td>
<td align="left">DTR<break/>DTC<break/>SVM<break/>GRUs<break/>hierarchical RNN<break/>BSLTMs</td>
<td align="left">Twitter (992 events)<break/>Sina Weibo (4664 events)</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s2_4"><label>2.4</label><title>Arabic Rumor Detection</title>
<p>The Arabic language has a complex structure that imposes challenges in addition to the lack of datasets. Thus, the researches on rumor detection in Arabic social media are few and require more attention and effort to achieve optimal results. The studies that are focused on Arabic rumor detection are summarized in <?A3B2 "tbl3",5,"anchor"?><xref ref-type="table" rid="table-3">Tab. 3</xref>.</p>
<table-wrap id="table-3"><label>Table 3</label><caption><title>The researches that are concerning Arabic rumors detection</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Reference</th>
<th align="left">Year</th>
<th align="left">Objective</th>
<th align="left">Method</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">[<xref ref-type="bibr" rid="ref-42">42</xref>]</td>
<td align="left">2016</td>
<td align="left">- Determining the attributes of Arabic rumor (fake information) patterns.<break/>Showing Arabic rumor identification in Twitter social.</td>
<td align="left">Using natural language processing and machine learning<break/>Using TF-IDF (Term frequency-inverse document frequency) for determining the weights for the terms.</td>
</tr>
<tr>
<td align="left">[<xref ref-type="bibr" rid="ref-43">43</xref>]</td>
<td align="left">2018</td>
<td align="left">- Detecting truth in the Arabic tweets<break/>- Preview the related work of credibility check for Arabic.</td>
<td align="left">- Introducing main procedures of the public model of Truth Detection for Arabic Tweets</td>
</tr>
<tr>
<td align="left">[<xref ref-type="bibr" rid="ref-44">44</xref>]</td>
<td align="left">2018</td>
<td align="left">- Detecting the trustiness of the information that is prevalent through social media.</td>
<td align="left">- Using classical machine learning algorithms<break/>- Using Cluster features and Cluster-Post features that were collected from social media</td>
</tr>
<tr>
<td align="left">[<xref ref-type="bibr" rid="ref-45">45</xref>]</td>
<td align="left">2018</td>
<td align="left">- Credibility checking model for Arabic news</td>
<td align="left">- Content parsing, features extraction, content verification, users&#x2019; comments, polarity evaluation, and credibility classification.</td>
</tr>
<tr>
<td align="left">[<xref ref-type="bibr" rid="ref-46">46</xref>]</td>
<td align="left">2019</td>
<td align="left">- Detecting rumors in Arabic social media</td>
<td align="left">- Using extracted features from the user and the content and analyzing to identify their importance.<break/>- Utilizing two different learning models, semi-supervised learning and unsupervised learning using expectation&#x2013;maximization</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s3"><label>3</label><title>Proposed Architecture</title>
<p>The proposed methodology investigates the most famous state of the arts deep learning algorithms for Arabic text classification; like deep learning techniques, they have the advantage of their ability to capture semantic features from textual data [<xref ref-type="bibr" rid="ref-47">47</xref>] automatically. The four different combinations of deep neural networks, namely, CNN, LSTM, BiLSTM, CNN&#x002B;LSTM, and CNN &#x002B; BiLSTM classification, have been performed. The proposed methodology is shown in <?A3B2 "fig1",5,"anchor"?><xref ref-type="fig" rid="fig-1">Fig. 1</xref>.</p>
<fig id="fig-1"><label>Figure 1</label><caption><title>System architecture</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_21449-fig-1.png"/></fig>
<sec id="s3_1"><label>3.1</label><title>Datasets</title>
<p>The first dataset consists of news and tweets that were manually collected and annotated by rumor/non-rumor. The actual dataset was collected from the Arabic news portals such as; Youm7, Akhbarelyom, and Ahram. This fake news was announced to make people aware that news is not absolute and fake. This effort is the responsibility of the information and decision support center, the Egyptian cabinet.</p>
<p>The second dataset is a benchmark dataset published in [<xref ref-type="bibr" rid="ref-48">48</xref>]. Then, the two datasets are merged into one large combined dataset to test deep learning performance using a larger dataset. The details of each dataset are shown in the <xref ref-type="table" rid="table-4">Tab. 4</xref>.</p>
<table-wrap id="table-4"><label>Table 4</label><caption><title>The description of the datasets</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Dataset</th>
<th align="left">Number of tweets</th>
<th align="left">Class label 0 (non-rumor)</th>
<th align="left">Class label 1 (rumor)</th>
<th align="left">No unique tokens</th>
<th align="left">The average length of tokens</th>
<th align="left">Max text length</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">Real dataset<break/>(Dataset 1)</td>
<td align="left">1980</td>
<td align="left">801</td>
<td align="left">1178</td>
<td align="left">8881</td>
<td align="left">14.1</td>
<td align="left">46</td>
</tr>
<tr>
<td align="left">Benchmark dataset<break/>(Dataset 2)</td>
<td align="left">2578</td>
<td align="left">1218</td>
<td align="left">1364</td>
<td align="left">12919</td>
<td align="left">18.4</td>
<td align="left">88</td>
</tr>
<tr>
<td align="left">Merged dataset<break/>(Dataset 3)</td>
<td align="left">4561</td>
<td align="left">2019</td>
<td align="left">2542</td>
<td align="left">18483</td>
<td align="left">16.3</td>
<td align="left">88</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><?A3B2 "tbl5",5,"anchor"?><xref ref-type="fig" rid="fig-5">Tab. 5</xref> shows samples of that real dataset collected from Arabic news websites.</p>
<fig id="fig-5"><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_21449-fig-5.png"/></fig>
</sec>
<sec id="s3_2"><label>3.2</label><title>Preprocessing</title>
<p>Preprocessing the text before it is fed into the classifier is very important and impacts the overall performance of the classification model. In this step, the text is cleaned using filters to remove punctuations and all non-Unicode characters. Afterward, stop words are removed, then sentences are tokenized, and tokens are stemmed. The resulting sentences are then encoded as numerical sequences, the number of unique tokens and the maximum sentences&#x2019; length is calculated. This maximum length is used to pad all sentences to be of the same size, equal to the maximum length. Labels are then encoded using one-hot-encoding.</p>
</sec>
<sec id="s3_3"><label>3.3</label><title>Word Embedding</title>
<p>Recently, word embeddings proved to outperform traditional text representation techniques. It represents each word as a real-valued vector in a dimensional space while preserving the semantic relationship between words. As vectors of words with similar meanings are placed close to each other. Word embeddings can be learned from the text to fit a deep neural model on text data. For our work, the Tensorflow Keras embedding layer was used. It takes, as an input, the numerically encoded text. It is implemented as the first hidden layer of the deep neural network where the word embeddings will be learned while training the network. The embedding layer stores a lookup table to map the words represented by numeric indexes to their dense vector representations.</p>
</sec>
<sec id="s3_4"><label>3.4</label><title>Proposed Models</title>
<p>Our system explores the usage of three deep neural networks, namely, CNN, LSTM, and BiLSTM, and two combinations between CNN&#x002B;LSTM and CNN&#x002B; BiLSTM as illustrated in <?A3B2 "fig2",5,"anchor"?><xref ref-type="fig" rid="fig-2">Fig. 2</xref>.</p>
<fig id="fig-2"><label>Figure 2</label><caption><title>Deep learning model</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_21449-fig-2.png"/></fig>
<p>CNN model consists of one conventional layer which learns to extract features from sequences represented using a word embedding and derive meaningful and useful sub-structures for the overall prediction task. It is implemented with 64 filters (parallel fields for processing words) with a linear (&#x2018;relu&#x2019;) activation function. The second layer is a pooling layer to reduce the output of the convolutional layer by half. The 2D output from the CNN part of the model is flattened to one long 2D vector to represent the &#x2018;features&#x2019; extracted by the CNN. Finally, two dense layers are used to scale, rotate and transform the vector by multiplying Matrix and vector.</p>
<p>The output of the word embedding layer is fed into one LSTM layer with 128 memory units. The output of the LSTM is fed into a dense layer of size&#x2009;&#x003D;&#x2009;64. This is used to enhance the complexity of the LSTM&#x0027;s output threshold. The activation function is natural for binary classification. It is a binary-class classification problem; binary cross-entropy is used as the loss function.</p>
<p>The third model combines both CNN with LSTM, where two conventional layers are added with max-pooling and dropout layers. The conventional layers act as feature extractors for the LSTMs on input data. The CNN layer uses the output of the word embeddings layer. Afterward, the pooling layer reduces the features extracted by the CNN layer. A dropout layer is added to help to prevent neural from overfitting. The LSTM layer with a hidden size of 128 is added. We use one LSTM layer with a state output of size&#x2009;&#x003D;&#x2009;128. Note, as per default return sequence is False, we only get one output, i.e., of the last state of the LSTM. The output of LSTM is connected with a dense layer of size&#x2009;&#x003D;&#x2009;64 to produce the final class label by calculating the probability of the LSTM output. The softmax activation function is used to generate the final classification.</p>
<p>The BiLSTM model uses a combination of recurrent and convolutional cells for learning. The output from the word embeddings layers is fed into a bi-directional LSTM. Afterward, dense layers are used to find the most suitable class based on probability.</p>
<p>BiLSTM-CNN model architecture uses a combination of convolutional and recurrent neurons for learning. As input, the output of the embeddings layers is fed into two-level conventional layers for feature learning for the BiLSTM layer. The features extracted by the CNN layers are max-pooled and concatenated. The fully connected dense layer predicts the probability of each class label.</p>
<p>For Training the deep learning models, Adam optimizer with 0.01 learning rate, weight decay of 0.0005, and 128 batch size. A dropout value of 0.5 is used to avoid overfitting and speed up the learning. The output layer uses a softmax activation function.</p>
<p>The experiments used python programming Tensorflow and Keras libraries for machine learning and deep learning models. A windows 10&#x2013;based machine with core i7 and 16 GB RAM was used.</p>
</sec>
</sec>
<sec id="s4"><label>4</label><title>Results and Discussion</title>
<p>Two experiments have been performed on three different datasets. The first experiment utilizes the proposed deep learning algorithms. The second experiment utilizes machine-learning algorithms using n-gram feature extraction and compares their results with deep learning algorithms.</p>
<p>Experiments included two phases; first, the most famous machine learning algorithms have been applied for classification with different n-grams. Machine learning techniques were evaluated using accuracy, f1-measure, and AUC (Area Under Curve) measures. The second phase of experiments included applying deep learning models for classification. Deep learning algorithms were first trained using simple data spilled with 80&#x0025; training and 20&#x0025; testing. Then same algorithms were trained using 5-fold cross-validation [<xref ref-type="bibr" rid="ref-49">49</xref>].</p>
<sec id="s4_1"><label>4.1</label><title>Machine Learning Experiments</title>
<p>The experiments are conducted using many machine learning algorithms, including Linear SVC, SVC, multinomialNB, bernoulliNB, stochastic gradient descent (SGD), decision tree, random forest, and k-neighbors. Each algorithm is evaluated using accuracy, F-score, and area under the curve (AUC). The results of the first dataset experiment are shown in <?A3B2 "tbl6",5,"anchor"?><xref ref-type="table" rid="table-6">Tab. 6</xref>. The table shows that the SGD classifier gives the best results. The figure shows that SVC, decision tree, and random forest give lower performance than other algorithms.</p>
<table-wrap id="table-6"><label>Table 6</label><caption><title>The results of the first dataset experiment</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Algorithm</th>
<th align="left">Accuracy</th>
<th align="left">F1_score</th>
<th align="left">AUC</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">Linear SVC</td>
<td align="left">0.859</td>
<td align="left">0.857</td>
<td align="left">0.845</td>
</tr>
<tr>
<td align="left">SVC</td>
<td align="left">0.596</td>
<td align="left">0.445</td>
<td align="left">0.5</td>
</tr>
<tr>
<td align="left">MultinomialNB</td>
<td align="left">0.823</td>
<td align="left">0.814</td>
<td align="left">0.788</td>
</tr>
<tr>
<td align="left">BernoulliNB</td>
<td align="left">0.821</td>
<td align="left">0.811</td>
<td align="left">0.786</td>
</tr>
<tr>
<td align="left">SGD classifier</td>
<td align="left">0.861</td>
<td align="left">0.861</td>
<td align="left">0.853</td>
</tr>
<tr>
<td align="left">Decision tree classifier</td>
<td align="left">0.641</td>
<td align="left">0.56</td>
<td align="left">0.563</td>
</tr>
<tr>
<td align="left">Random forest classifier</td>
<td align="left">0.596</td>
<td align="left">0.445</td>
<td align="left">0.5</td>
</tr>
<tr>
<td align="left">KNeighbors classifier</td>
<td align="left">0.806</td>
<td align="left">0.796</td>
<td align="left">0.77</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The results of the second dataset experiment are shown in <?A3B2 "tbl7",5,"anchor"?><xref ref-type="table" rid="table-7">Tab. 7</xref>. Each algorithm is evaluated using accuracy, F-score, and area under the curve (AUC). <xref ref-type="table" rid="table-7">Tab. 7</xref> shows that the LinerarSVC classifier gives the best results. The table shows that SVC, decision tree, and random forest give lower performance than other algorithms.</p>
<table-wrap id="table-7"><label>Table 7</label><caption><title>The results of the second dataset experiment</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Algorithm</th>
<th align="left">Accuracy</th>
<th align="left">F1_score</th>
<th align="left">AUC</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">Linear SVC</td>
<td align="left">0.762</td>
<td align="left">0.762</td>
<td align="left">0.762</td>
</tr>
<tr>
<td align="left">SVC</td>
<td align="left">0.511</td>
<td align="left">0.345</td>
<td align="left">0.5</td>
</tr>
<tr>
<td align="left">MultinomialNB</td>
<td align="left">0.743</td>
<td align="left">0.738</td>
<td align="left">0.74</td>
</tr>
<tr>
<td align="left">BernoulliNB</td>
<td align="left">0.735</td>
<td align="left">0.725</td>
<td align="left">0.731</td>
</tr>
<tr>
<td align="left">SGD classifier</td>
<td align="left">0.737</td>
<td align="left">0.737</td>
<td align="left">0.736</td>
</tr>
<tr>
<td align="left">Decision tree classifier</td>
<td align="left">0.584</td>
<td align="left">0.573</td>
<td align="left">0.588</td>
</tr>
<tr>
<td align="left">Random forest classifier</td>
<td align="left">0.515</td>
<td align="left">0.373</td>
<td align="left">0.504</td>
</tr>
<tr>
<td align="left">KNeighbors classifier</td>
<td align="left">0.725</td>
<td align="left">0.725</td>
<td align="left">0.724</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The results of the third dataset experiment are shown in <?A3B2 "tbl8",5,"anchor"?><xref ref-type="table" rid="table-8">Tab. 8</xref>. Each algorithm is evaluated using accuracy, F-score, and area under the curve (AUC). The table shows that the LinerarSVC classifier gives the best F-score and AUC while MultinomialNB gives the best accuracy. The figure shows that SVC, decision tree, and random forest give lower performance than other algorithms.</p>
<table-wrap id="table-8"><label>Table 8</label><caption><title>The results of the third dataset experiment</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Algorithm</th>
<th align="left">Accuracy</th>
<th align="left">F1_score</th>
<th align="left">AUC</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">Linear SVC</td>
<td align="left">0.779</td>
<td align="left">0.779</td>
<td align="left">0.775</td>
</tr>
<tr>
<td align="left">SVC</td>
<td align="left">0.561</td>
<td align="left">0.403</td>
<td align="left">0.5</td>
</tr>
<tr>
<td align="left">MultinomialNB</td>
<td align="left">0.783</td>
<td align="left">0.778</td>
<td align="left">0.767</td>
</tr>
<tr>
<td align="left">BernoulliNB</td>
<td align="left">0.779</td>
<td align="left">0.772</td>
<td align="left">0.761</td>
</tr>
<tr>
<td align="left">SGD classifier</td>
<td align="left">0.769</td>
<td align="left">0.769</td>
<td align="left">0.767</td>
</tr>
<tr>
<td align="left">Decision tree classifier</td>
<td align="left">0.593</td>
<td align="left">0.489</td>
<td align="left">0.54</td>
</tr>
<tr>
<td align="left">Random forest classifier</td>
<td align="left">0.56</td>
<td align="left">0.419</td>
<td align="left">0.501</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The methods SVC, decision tree, and random forest can be concluded that is not suitable for this problem. The following graphs depicted in <?A3B2 "fig3",5,"anchor"?><xref ref-type="fig" rid="fig-3">Figs. 3a</xref>&#x2013;<xref ref-type="fig" rid="fig-3">3h</xref> show the performance of each ML algorithm applied to each dataset. The BernoulliNB method shows its performance with the first dataset, found in <xref ref-type="fig" rid="fig-3">Fig. 3a</xref>. <xref ref-type="fig" rid="fig-3">Fig. 3b</xref> shows that MultinomialNB gives its lower performance for the third dataset. <xref ref-type="fig" rid="fig-3">Fig. 3c</xref> shows that k neighbors give their best performance for the first dataset. <xref ref-type="fig" rid="fig-3">Fig. 3d</xref> shows that random forest gives its best performance for the first dataset. <xref ref-type="fig" rid="fig-3">Fig. 3e</xref> shows that the decision tree performs best for the first dataset. <xref ref-type="fig" rid="fig-3">Fig. 3f</xref> shows that the SGD classifier gives an almost equivalent performance for all datasets. <xref ref-type="fig" rid="fig-3">Fig. 3g</xref> shows that SVC gives its best performance for the first dataset. <xref ref-type="fig" rid="fig-3">Fig. 3h</xref> shows that Linear SVC gives almost equivalent performance for all datasets. The first dataset is the manual collected and annotated data, which is the real-life data. Therefore, the machine learning algorithms give excellent performance for real-life data.</p>
<fig id="fig-3"><label>Figure 3</label><caption><title>The performance of each ML algorithms with each dataset</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_21449-fig-3.png"/></fig>
</sec>
<sec id="s4_2"><label>4.2</label><title>Deep Learning Algorithms</title>
<p>Many deep learning algorithms are conducted; CNN, LSTM, CNN &#x002B; LSTM, BiLSTM, and CNN &#x002B; BiLSTM. The evaluation metrics of accuracy, loss, and AUC are used.</p>
<p><?A3B2 "tbl9",5,"anchor"?><xref ref-type="table" rid="table-9">Tab. 9</xref> shows the performance of each algorithm applied to the first dataset. <xref ref-type="table" rid="table-9">Tab. 9</xref> shows that BiLSTM gives the slightest loss, best accuracy, and best AUC. Thus, BiLSTM gives good performance with reasonable loss compared to other algorithms, which are close to each other in performance.</p>
<table-wrap id="table-9"><label>Table 9</label><caption><title>The performance of each algorithm applied on the first dataset</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Deep learning model</th>
<th align="left">Accuracy</th>
<th align="left">Loss</th>
<th align="left">AUC</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">CNN</td>
<td align="left">0.780303</td>
<td align="left">0.582185</td>
<td align="left">0.839837</td>
</tr>
<tr>
<td align="left">LSTM</td>
<td align="left">0.838384</td>
<td align="left">1.387679</td>
<td align="left">0.887626</td>
</tr>
<tr>
<td align="left">CNN &#x002B; LSTM</td>
<td align="left">0.825758</td>
<td align="left">0.837274</td>
<td align="left">0.886626</td>
</tr>
<tr>
<td align="left">BiLSTM</td>
<td align="left">0.848283</td>
<td align="left">0.583095</td>
<td align="left">0.902887</td>
</tr>
<tr>
<td align="left">CNN &#x002B; BiLSTM</td>
<td align="left">0.8283</td>
<td align="left">0.909429</td>
<td align="left">0.886772</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><?A3B2 "tbl10",5,"anchor"?><xref ref-type="table" rid="table-10">Tab. 10</xref> shows the performance of each algorithm on the second dataset. Results on the second dataset also show that BiLSTM gives the least amount of loss, the best accuracy, and the best. Additionally, CNN gives a bad performance with a significant loss and the lowest accuracy. Other algorithms give almost similar performances.</p>
<table-wrap id="table-10"><label>Table 10</label><caption><title>The performance of each algorithm applied on the second dataset</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Deep learning model</th>
<th align="left">Accuracy</th>
<th align="left">Loss</th>
<th align="left">AUC</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">CNN</td>
<td align="left">0.597679</td>
<td align="left">4.226126</td>
<td align="left">0.61282</td>
</tr>
<tr>
<td align="left">LSTM</td>
<td align="left">0.709865</td>
<td align="left">1.902138</td>
<td align="left">0.792778</td>
</tr>
<tr>
<td align="left">CNN &#x002B; LSTM</td>
<td align="left">0.727273</td>
<td align="left">1.401511</td>
<td align="left">0.792778</td>
</tr>
<tr>
<td align="left">BiLSTM</td>
<td align="left">0.742747</td>
<td align="left">1.236142</td>
<td align="left">0.803262</td>
</tr>
<tr>
<td align="left">CNN &#x002B; BiLSTM</td>
<td align="left">0.7273</td>
<td align="left">1.768434</td>
<td align="left">0.791577</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><?A3B2 "tbl11",5,"anchor"?><xref ref-type="table" rid="table-11">Tab. 11</xref> shows the performance of each algorithm on the third dataset. <xref ref-type="table" rid="table-11">Tab. 11</xref> shows that CNN gives the least amount of loss. BiLSTM gives the best accuracy and the best AUC. <xref ref-type="table" rid="table-11">Tab. 11</xref> shows that LSTM and CNN &#x002B; BiLSTM give a significant loss while accuracies and AUC are almost similar to other algorithms.</p>
<table-wrap id="table-11"><label>Table 11</label><caption><title>The performance of each algorithm applied on the third dataset</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Deep learning model</th>
<th align="left">Accuracy</th>
<th align="left">Loss</th>
<th align="left">AUC</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">CNN</td>
<td align="left">0.710843</td>
<td align="left">1.118824</td>
<td align="left">0.765603</td>
</tr>
<tr>
<td align="left">LSTM</td>
<td align="left">0.765608</td>
<td align="left">2.172314</td>
<td align="left">0.829969</td>
</tr>
<tr>
<td align="left">CNN &#x002B; LSTM</td>
<td align="left">0.760131</td>
<td align="left">1.359961</td>
<td align="left">0.831611</td>
</tr>
<tr>
<td align="left">BiLSTM</td>
<td align="left">0.773275</td>
<td align="left">1.151952</td>
<td align="left">0.837928</td>
</tr>
<tr>
<td align="left">CNN &#x002B; BiLSTM</td>
<td align="left">0.7558</td>
<td align="left">1.741797</td>
<td align="left">0.829785</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The following graphs depicted in <?A3B2 "fig4",5,"anchor"?><xref ref-type="fig" rid="fig-4">Figs. 4a</xref>&#x2013;<xref ref-type="fig" rid="fig-4">4h</xref> show the performance of each deep learning algorithm with each dataset. The BiLSTM method is more suitable for the first and second datasets because it significantly loses the third dataset, as shown in <xref ref-type="fig" rid="fig-4">Fig. 4a</xref>. <xref ref-type="fig" rid="fig-4">Fig. 4b</xref> shows that CNN is not suitable for this problem as it gives low performance for all datasets and a considerable amount of loss in the third dataset. <xref ref-type="fig" rid="fig-4">Fig. 4c</xref> shows that CNN &#x002B; BiLSTM is more suitable for datasets one and second as it significantly loses dataset three. <xref ref-type="fig" rid="fig-4">Fig. 4d</xref> shows that CNN&#x002B;LSTM is more suitable for the first and second datasets as it gives a significant amount of loss in the third dataset. <xref ref-type="fig" rid="fig-4">Fig. 4e</xref> shows that LSTM provides accepted performance for the first and second datasets but encounters a significant loss for the third dataset. Therefore, deep learning algorithms give better performance when combining with real-life data.</p>
<fig id="fig-4"><label>Figure 4</label><caption><title>The performance of each deep learning algorithm with each dataset</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_21449-fig-4.png"/></fig>
</sec>
<sec id="s4_3"><label>4.3</label><title>Cross-Validation</title>
<p>To verify the experiments done with deep learning algorithms, five-fold cross-validation on the three datasets have experimented. The results of each dataset are shown in <?A3B2 "tbl12",5,"anchor"?><?A3B2 "tbl13",5,"anchor"?><?A3B2 "tbl14",5,"anchor"?><xref ref-type="table" rid="table-12 table-13 table-14">Tabs. 12&#x2013;14</xref>. Results show that BiLSTM and BiLSTM &#x002B; CNN give the highest accuracy and most negligible loss for all three datasets. On the other hand, CNN achieved the worest performance among all experimented models.</p>
<table-wrap id="table-12"><label>Table 12</label><caption><title>The five-fold cross-validation of the first dataset</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Deep learning model</th>
<th align="left">Accuracy</th>
<th align="left">Loss</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">CNN</td>
<td align="left">58.27</td>
<td align="left">4.54</td>
</tr>
<tr>
<td align="left">LSTM</td>
<td align="left">82.06</td>
<td align="left">1.13</td>
</tr>
<tr>
<td align="left">CNN &#x002B; LSTM</td>
<td align="left">81.86</td>
<td align="left">1.01</td>
</tr>
<tr>
<td align="left">BiLSTM</td>
<td align="left">83.92</td>
<td align="left">0.86</td>
</tr>
<tr>
<td align="left">CNN &#x002B; BiLSTM</td>
<td align="left">83.88</td>
<td align="left">1.08</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="table-13"><label>Table 13</label><caption><title>The five-fold cross-validation of the second dataset</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Deep learning model</th>
<th align="left">Accuracy</th>
<th align="left">Loss</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">CNN</td>
<td align="left">58.43</td>
<td align="left">4.25</td>
</tr>
<tr>
<td align="left">LSTM</td>
<td align="left">70.02</td>
<td align="left">2.51</td>
</tr>
<tr>
<td align="left">CNN &#x002B; LSTM</td>
<td align="left">70.19</td>
<td align="left">2.09</td>
</tr>
<tr>
<td align="left">BiLSTM</td>
<td align="left">70.83</td>
<td align="left">2.06</td>
</tr>
<tr>
<td align="left">CNN &#x002B; BiLSTM</td>
<td align="left">71.72</td>
<td align="left">2.01</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="table-14"><label>Table 14</label><caption><title>The five-fold cross-validation of the third dataset</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Deep learning model</th>
<th align="left">Accuracy</th>
<th align="left">Loss</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">CNN</td>
<td align="left">57.42</td>
<td align="left">5.94</td>
</tr>
<tr>
<td align="left">LSTM</td>
<td align="left">73.34</td>
<td align="left">2.59</td>
</tr>
<tr>
<td align="left">CNN &#x002B; LSTM</td>
<td align="left">73.03</td>
<td align="left">2.13</td>
</tr>
<tr>
<td align="left">BiLSTM</td>
<td align="left">74.98</td>
<td align="left">1.97</td>
</tr>
<tr>
<td align="left">CNN &#x002B; BiLSTM</td>
<td align="left">73. 87</td>
<td align="left">2.02</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s5"><label>5</label><title>Conclusions and Future Works</title>
<p>This paper aims at investigating machine learning and deep learning models for content-based Arabic fake news classification. A series of experiments were conducted to evaluate the task-specific deep learning models. Three datasets were used in the experiments to assess the most well-known models in the literature. Our findings indicate that machine learning and deep learning approaches can identify fake news using text-based linguistic features. There was no single model that performed optimally across all datasets in terms of machine learning algorithms. On the other hand, our results show that the BiLSTM model achieves the highest accuracy among all models assessed across all datasets.</p>
<p>We intend to thoroughly examine the existing architectures combining various layers as part of our future work. Furthermore, examine the effect of various pre-trained word embeddings on the performance of deep learning models.</p>
</sec>
</body>
<back>
<fn-group>
<fn fn-type="other"><p><bold>Funding Statement:</bold> The authors received no specific funding for this study.</p></fn>
<fn fn-type="conflict"><p><bold>Conflicts of Interest:</bold> Authors declare that they have no conflicts of interest to report regarding the present study.</p></fn>
</fn-group>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Vohra</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Kakkar</surname></string-name></person-group>, &#x201C;<article-title>Detection of rumor in social media</article-title>,&#x201D; in <conf-name>Proc. 8th Int. Conf. on Cloud Computing, Data Science &#x0026; Engineering</conf-name>, <conf-loc>Noida, India</conf-loc>, pp. <fpage>485</fpage>&#x2013;<lpage>490</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>F.</given-names> <surname>Pierri</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Ceri</surname></string-name></person-group>, &#x201C;<article-title>False news on social media: A data-driven survey</article-title>,&#x201D; <source>ACM SIGMOD Record</source>, vol. <volume>48</volume>, no. <issue>2</issue>, pp. <fpage>18</fpage>&#x2013;<lpage>27</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Kumar</surname></string-name> and <string-name><given-names>N.</given-names> <surname>Shah</surname></string-name></person-group>, &#x201C;<article-title>False information on web and social media: A survey</article-title>,&#x201D; <source>ArXiv</source>, vol. <volume>abs/1804.08559</volume>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Zannettou</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Sirivianos</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Blackburn</surname></string-name> and <string-name><given-names>N.</given-names> <surname>ourtellis</surname></string-name></person-group>, &#x201C;<article-title>The web of false information: Rumors, fake news, hoaxes, clickbait, and various other shenanigans</article-title>,&#x201D; <source>Journal of Data and Information Quality (JDIQ)</source>, vol. <volume>11</volume>, no. <issue>10</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>37</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Tandoc</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Lim</surname></string-name> and <string-name><given-names>R.</given-names> <surname>Ling</surname></string-name></person-group>, &#x201C;<article-title>Defining &#x201C;fake news&#x201D; a typology of scholarly definitions</article-title>,&#x201D; <source>Digital Journalism</source>, vol. <volume>6</volume>, no. <issue>2</issue>, pp. <fpage>137</fpage>&#x2013;<lpage>153</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Q.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Yang</surname></string-name> and <string-name><given-names>W.</given-names> <surname>Xi</surname></string-name></person-group>, &#x201C;<article-title>Effects of group arguments on rumor belief and t]transmission in online communities: An information cascade and group polarization perspective</article-title>,&#x201D; <source>Information &#x0026; Management</source>, vol. <volume>55</volume>, no. <issue>4</issue>, pp. <fpage>441</fpage>&#x2013;<lpage>449</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>V.</given-names> <surname>Agarwal</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Sultana</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Malhotra</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Sarkar</surname></string-name></person-group>, &#x201C;<article-title>Analysis of classifiers for fake news detection</article-title>,&#x201D; <source>Procedia Computer Science</source>, vol. <volume>165</volume>, no. <issue>1</issue>, pp. <fpage>377</fpage>&#x2013;<lpage>383</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Peng</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Yang</surname></string-name> and <string-name><given-names>L.</given-names> <surname>Lu</surname></string-name></person-group>, &#x201C;<chapter-title>Chapter 5-text mining and deep learning for disease classification</chapter-title>,&#x201D; in <source>The Elsevier and MICCAI Society Book Series, Handbook of Medical Image Computing and Computer Assisted Intervention</source>, <publisher-name>Academic Press</publisher-name>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>R.</given-names> <surname>Kaliyar</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Goswami</surname></string-name>, <string-name><given-names>P.</given-names> <surname>Narang</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Sinha</surname></string-name></person-group>, &#x201C;<article-title>FNDNet a deep convolutional neural network for fake news detection</article-title>,&#x201D; <source>Cognitive Systems Research</source>, vol. <volume>61</volume>, no. <issue>1</issue>, pp. <fpage>32</fpage>&#x2013;<lpage>44</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Elnagar</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Al-Debsi</surname></string-name> and <string-name><given-names>O.</given-names> <surname>Einea</surname></string-name></person-group>, &#x201C;<article-title>Arabic text classification using deep learning models</article-title>,&#x201D; <source>Information Processing and Management</source>, vol. <volume>57</volume>, no. <issue>1</issue>, pp. <fpage>102</fpage>&#x2013;<lpage>121</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>V.</given-names> <surname>P&#x00E9;rez-Rosas</surname></string-name>, <string-name><given-names>B.</given-names> <surname>Kleinberg</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Lefevre</surname></string-name> and <string-name><given-names>R.</given-names> <surname>Mihalcea</surname></string-name></person-group>, &#x201C;<article-title>Automatic detection of fake news</article-title>,&#x201D; in <source>proc. 27th Int. Conf. on Computational Linguistics</source>, <conf-loc>Santa Fe, New Mexico, USA</conf-loc>, pp. <fpage>3391</fpage>&#x2013;<lpage>3401</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Jin</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Cao</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Zhou</surname></string-name> and <string-name><given-names>Q.</given-names> <surname>Tian</surname></string-name></person-group>, &#x201C;<article-title>Novel visual and statistical image features for microblogs news verification</article-title>,&#x201D; <source>IEEE Transactions on Multimedia</source>, vol. <volume>19</volume>, no. <issue>3</issue>, pp. <fpage>598</fpage>&#x2013;<lpage>608</lpage>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Ahsan</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Kumari</surname></string-name> and <string-name><given-names>T.</given-names> <surname>Sharma</surname></string-name></person-group>, &#x201C;<article-title>Rumors detection, verification and controlling mechanisms in online social networks: A survey</article-title>,&#x201D; <source>Online Social Networks and Media</source>, vol. <volume>14</volume>, no. <issue>1</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>12</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Bondielli</surname></string-name> and <string-name><given-names>F.</given-names> <surname>Marcelloni</surname></string-name></person-group>, &#x201C;<article-title>A survey on fake news and rumour detection techniques</article-title>,&#x201D; <source>Information Sciences</source>, vol. <volume>497</volume>, no. <issue>1</issue>, pp. <fpage>38</fpage>&#x2013;<lpage>55</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Al-Sarem</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Boulila</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Al-Harby</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Qadir</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Alsaeedi</surname></string-name></person-group>, &#x201C;<article-title>Deep learning-based rumor detection on microblogging platforms: A systematic review</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>7</volume>, pp. <fpage>152788</fpage>&#x2013;<lpage>152812</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Sabbeh</surname></string-name></person-group>,&#x0022;<article-title>Performance evaluation of different data mining techniques for social media news credibility assessment,&#x201D;</article-title> <source>International Journal of Advanced Computer Science and Applications</source>, vol. <volume>10</volume>, no. <issue>9</issue>, pp. <fpage>245</fpage>&#x2013;<lpage>256</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Abbasi</surname></string-name>, <string-name><given-names>A. R.</given-names> <surname>Javed</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Chakraborty</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Nebhen</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Zehra</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Elstream: An ensemble learning approach for concept drift detection in dynamic social big data stream learning</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>9</volume>, pp. <fpage>66408</fpage>&#x2013;<lpage>66419</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Reis</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Correia</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Murai</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Veloso</surname></string-name> and <string-name><given-names>F.</given-names> <surname>Benevenuto</surname></string-name></person-group>, &#x201C;<article-title>Supervised learning for fake news detection</article-title>,&#x201D; <source>IEEE Intelligent Systems</source>, vol. <volume>34</volume>, no. <issue>2</issue>, pp. <fpage>76</fpage>&#x2013;<lpage>81</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Granik</surname></string-name> and <string-name><given-names>V.</given-names> <surname>Mesyura</surname></string-name></person-group>, &#x201C;<article-title>Fake news detection using naive Bayes classifier</article-title>,&#x201D; in <conf-name>IEEE First Ukraine Conf. on Electrical and Computer Engineering (UKRCON)</conf-name>, <conf-loc>Kyiv, Ukraine</conf-loc>, pp. <fpage>900</fpage>&#x2013;<lpage>903</lpage>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Benjamin</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Ciro</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Filippo</surname></string-name></person-group>, &#x201C;<article-title>Social spam detection</article-title>,&#x201D; in <conf-name>5th Int. Workshop on Adversarial Information Retrieval on the Web</conf-name>, <conf-loc>ACM, Madrid, Spain</conf-loc>, pp. <fpage>41</fpage>&#x2013;<lpage>48</lpage>, <year>2009</year>.</mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>R.</given-names> <surname>Deepa</surname></string-name> and <string-name><given-names>N.</given-names> <surname>Radha</surname></string-name></person-group>, &#x201C;<article-title>Supervised learning approach for spam classification analysis using data mining tools</article-title>,&#x201D; <source>(IJCSE) International Journal on Computer Science and Engineering</source>, vol. <volume>2</volume>, no. <issue>9</issue>, pp. <fpage>2783</fpage>&#x2013;<lpage>2789</lpage>, <year>2010</year>.</mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>K.</given-names> <surname>Fouad</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Elsheshtawy</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Dawood</surname></string-name></person-group>, &#x201C;<article-title>Intelligent approach for large-scale data mining</article-title>,&#x201D; <source>International Journal of Sociotechnology and Knowledge Development</source>, vol. <volume>13</volume>, no. <issue>2</issue>, pp. <fpage>119</fpage>&#x2013;<lpage>152</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Jin</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Cao</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Zhang</surname></string-name> and <string-name><given-names>J.</given-names> <surname>Luo</surname></string-name></person-group>, &#x201C;<article-title>News verification by exploiting conflicting social viewpoints in microblogs</article-title>,&#x201D; in <conf-name>Thirtieth AAAI Conf. on Artificial Intelligence</conf-name>, <conf-loc>Phoenix, Arizona, USA</conf-loc>, pp. <fpage>2972</fpage>&#x2013;<lpage>2978</lpage>, <year>2016</year>.</mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>L.</given-names> <surname>Siwei</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Liheng</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Kang</surname></string-name> and <string-name><given-names>Z.</given-names> <surname>Jun</surname></string-name></person-group>, &#x201C;<article-title>Recurrent convolutional neural networks for text classification</article-title>,&#x201D; in <conf-name>Twenty-Ninth AAAI Conf. on Artificial Intelligence</conf-name>, <conf-loc>Austin, Texas, USA</conf-loc>, pp. <fpage>2267</fpage>&#x2013;<lpage>2273</lpage>, <year>2015</year>.</mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Oscar</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Ignacio</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Fernando</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Carlos</surname></string-name></person-group>, &#x201C;<article-title>Enhancing deep learning sentiment analysis with ensemble techniques in social applications</article-title>,&#x201D; <source>Expert Systems with Applications</source>, vol. <volume>77</volume>, pp. <fpage>246</fpage>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-26"><label>[26]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Elnagar</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Al-Debsi</surname></string-name> and <string-name><given-names>O.</given-names> <surname>Einea</surname></string-name></person-group>, &#x201C;<article-title>Arabic text classification using deep learning models</article-title>,&#x201D; <source>Information Processing and Management</source>, vol. <volume>57</volume>, no. <issue>1</issue>, pp. <fpage>102121</fpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-27"><label>[27]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Heikal</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Torki</surname></string-name> and <string-name><given-names>N.</given-names> <surname>El-Makky</surname></string-name></person-group>, &#x201C;<article-title>Sentiment analysis of arabic tweets using deep learning</article-title>,&#x201D; in <conf-name>4th Int. Conf. on Arabic Computational Linguistics (ACLing 2018)</conf-name>, <conf-loc>Dubai, UAE</conf-loc>, pp. <fpage>114</fpage>&#x2013;<lpage>122</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-28"><label>[28]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Kumar</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Asthana</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Upadhyay</surname></string-name>, <string-name><given-names>N.</given-names> <surname>Upreti</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Akbar</surname></string-name></person-group>, &#x201C;<article-title>Fake news detection using deep learning models: A novel approach</article-title>,&#x201D; <source>Transactions on Emerging Telecommunication Technologies</source>, vol. <volume>31</volume>, no. <issue>2</issue>, pp. <fpage>e3767</fpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-29"><label>[29]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Umer</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Imtiaz</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Ullah</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Mehmood</surname></string-name>, <string-name><given-names>G. S.</given-names> <surname>Choi</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Fake news stance detection using deep learning srchitecture (CNN-lSTM)</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>8</volume>, pp. <fpage>156695</fpage>&#x2013;<lpage>156706</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-30"><label>[30]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Girgis</surname></string-name>, <string-name><given-names>E.</given-names> <surname>Amer</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Gadallah</surname></string-name></person-group>, &#x201C;<article-title>Deep learning algorithms for detecting fake news in online text</article-title>,&#x201D; in <conf-name>13th Int. Conf. on Computer Engineering and Systems (ICCES)</conf-name>, <conf-loc>Cairo, Egypt</conf-loc>, pp. <fpage>93</fpage>&#x2013;<lpage>97</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-31"><label>[31]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Jing</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Wei</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Prasenjit</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Sejeong</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Jim</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Detecting rumors from microblogs with recurrent neural networks</article-title>,&#x201D; in <conf-name>Twenty-Fifth Int. Joint Conf. on Artificial Intelligence IJCAI-16</conf-name>, <conf-loc>New York, USA</conf-loc>, pp. <fpage>3818</fpage>&#x2013;<lpage>3824</lpage>, <year>2016</year>.</mixed-citation></ref>
<ref id="ref-32"><label>[32]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Muhammad</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Ammara</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Anam</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Adil</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Rehman</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Exploring deep neural networks for rumor detection</article-title>,&#x201D; <source>Journal of Ambient Intelligence and Humanized Computing</source>, vol. <volume>12</volume>, pp. <fpage>4315</fpage>&#x2013;<lpage>4333</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-33"><label>[33]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>R.</given-names> <surname>AlvaroIbrain</surname></string-name> and <string-name><given-names>L.</given-names> <surname>Lara</surname></string-name></person-group>, &#x201C;<article-title>Fake news detection using deep learning</article-title>,&#x201D; <source>Journal of Information Process Systems</source>, vol. <volume>15</volume>, no. <issue>5</issue>, pp. <fpage>1119</fpage>&#x2013;<lpage>1130</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-34"><label>[34]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Verma</surname></string-name>, <string-name><given-names>V.</given-names> <surname>Mittal</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Dawn</surname></string-name></person-group>, &#x201C;<article-title>FIND: Fake information and news detections using deep learning</article-title>,&#x201D; in <conf-name>Twelfth Int. Conf. on Contemporary Computing (IC3)</conf-name>, <conf-loc>Noida, India</conf-loc>, pp. <fpage>1</fpage>&#x2013;<lpage>7</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-35"><label>[35]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>K.</given-names> <surname>Kaliyar</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Goswami</surname></string-name>, <string-name><given-names>P.</given-names> <surname>Narang</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Sinha</surname></string-name></person-group>, &#x201C;<article-title>FNDNet- a deep convolutional neural network for fake news detection</article-title>,&#x201D; <source>Cognitive Systems Research</source>, vol. <volume>61</volume>, pp. <fpage>32</fpage>&#x2013;<lpage>44</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-36"><label>[36]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>R.</given-names> <surname>Kaliyar</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Goswami</surname></string-name> and <string-name><given-names>P.</given-names> <surname>Narang</surname></string-name></person-group>, &#x201C;<article-title>FakeBERT: Fake news detection in social media with a BERT-based deep learning approach</article-title>,&#x201D; <source>Multimedia Tools and Application</source>, vol. <volume>80</volume>, no. <issue>8</issue>, pp. <fpage>11765</fpage>&#x2013;<lpage>11788</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-37"><label>[37]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>T.</given-names> <surname>Sansiri</surname></string-name> and <string-name><given-names>H.</given-names> <surname>Kien</surname></string-name></person-group> &#x201C;<article-title>Attention based neural architecture for rumor detection with author context awareness</article-title>,&#x201D; in <conf-name>Thirteenth Int. Conf. on Digital Information Management (ICDIM)</conf-name>, <conf-loc>Luton, UK</conf-loc>, pp. <fpage>82</fpage>&#x2013;<lpage>87</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-38"><label>[38]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Pavithra</surname></string-name> and <string-name><given-names>J.</given-names> <surname>Shibily</surname></string-name></person-group>, &#x201C;<article-title>Deep learning approach for rumour detection in twitter: A comparative analysis</article-title>,&#x201D; in <conf-name>Int. Conf. on Systems, Energy &#x0026; Environment (ICSEE)</conf-name>, <conf-loc>GCE Kannur, Kerala</conf-loc>, <month>July</month> <year>2019</year>.</mixed-citation></ref>
<ref id="ref-39"><label>[39]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Jing</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Wei</surname></string-name> and <string-name><given-names>W.</given-names> <surname>Kam-Fai</surname></string-name></person-group>, &#x201C;<article-title>Rumor detection on twitter with tree-structured recursive neural networks</article-title>,&#x201D; in <conf-name>56th Annual Meeting of the Association for Computational Linguistics (Long Papers)</conf-name>, <conf-loc>Melbourne, Australia</conf-loc>, pp. <fpage>1980</fpage>&#x2013;<lpage>1989</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-40"><label>[40]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>X.</given-names> <surname>Yichun</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Chen</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Zhiping</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Shuifa</surname></string-name> and <string-name><given-names>D.</given-names> <surname>Fangmin</surname></string-name></person-group>, &#x201C;<article-title>Deep recurrent neural network and data filtering for rumor detection on sina weibo</article-title>,&#x201D; <source>Symmetry</source>, vol. <volume>11</volume>, no. <issue>11</issue>, pp. <fpage>1408</fpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-41"><label>[41]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>X.</given-names> <surname>Lin</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Liao</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Xu</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Pianand</surname></string-name> and <string-name><given-names>K.</given-names> <surname>Wong</surname></string-name></person-group>, &#x201C;<article-title>Rumor detection with hierarchical recurrent convolutional neural network</article-title>,&#x201D; in <conf-name>8th CCF Int. Conf., NLPCC</conf-name>, <conf-loc>Dunhuang, China</conf-loc>, pp. <fpage>338</fpage>&#x2013;<lpage>348</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-42"><label>[42]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Yahya</surname></string-name></person-group>, &#x201C;<article-title>Arabic rumours identification by measuring the credibility of arabic tweet content</article-title>,&#x201D; <source>International Journal of Knowledge Society Research (IJKSR)</source>, vol. <volume>7</volume>, no. <issue>2</issue>, pp. <fpage>72</fpage>&#x2013;<lpage>83</lpage>, <year>2016</year>.</mixed-citation></ref>
<ref id="ref-43"><label>[43]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>R.</given-names> <surname>Mouty</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Gazdar</surname></string-name></person-group>, &#x201C;<article-title>Survey on steps of truth detection on arabic tweets</article-title>,&#x201D; in <conf-name>21st Saudi Computer Society National Computer Conf. (NCC)</conf-name>, <conf-loc>Riyadh, Saudi Arabia</conf-loc>, pp. <fpage>1</fpage>&#x2013;<lpage>6</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-44"><label>[44]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Alzanin</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Azmi</surname></string-name></person-group>, &#x201C;<article-title>Detecting rumors in social media: A survey</article-title>,&#x201D; <source>Procedia Computer Science</source>, vol. <volume>142</volume>, pp. <fpage>294</fpage>&#x2013;<lpage>300</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-45"><label>[45]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Sabbeh</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Baatwah</surname></string-name></person-group>, &#x201C;<article-title>Arabic news credibility on twitter: An enhanced model using hybrid features</article-title>,&#x201D; <source>Journal of Theoretical and Applied Information Technology</source>, vol. <volume>96</volume>, no. <issue>8</issue>, pp. <fpage>2327</fpage>&#x2013;<lpage>2338</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-46"><label>[46]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Alzanin</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Azmi</surname></string-name></person-group>, &#x201C;<article-title>Rumor detection in arabic tweets using semi-supervised and unsupervised expectation&#x2013;maximization</article-title>,&#x201D; <source>Knowledge-Based Systems</source>, vol. <volume>185</volume>, pp. <fpage>104945</fpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-47"><label>[47]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Q.</given-names> <surname>Liu</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Huang</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Gao</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Wei</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Tian</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>Task-oriented word embedding for text classification</article-title>,&#x201D; in <conf-name>27th Int. Conf. on Computational Linguistics</conf-name>, <conf-loc>Santa Fe, New-Mexico, USA</conf-loc>, pp. <fpage>2023</fpage>&#x2013;<lpage>2032</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-48"><label>[48]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><given-names>R.</given-names> <surname>Francisco</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Paolo</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Anis</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Wajdi</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Bilal</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<chapter-title>Overview of the track on author profiling and deception detection in arabic</chapter-title>,&#x201D; in <source>Working Notes of the Forum for Information Retrieval Evaluation (FIRE 2019)</source>, <publisher-loc>Kolkata</publisher-loc>, <publisher-loc>India</publisher-loc>, pp. <fpage>70</fpage>&#x2013;<lpage>83</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-49"><label>[49]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>K.</given-names> <surname>Fouad</surname></string-name> and <string-name><given-names>D.</given-names> <surname>El-Bably</surname></string-name></person-group>, &#x201C;<article-title>Intelligent approach for large-scale data mining</article-title>,&#x201D; <source>International Journal of Computer Applications in Technology</source>, vol. <volume>63</volume>, no. <issue>1&#x2013;2</issue>, pp. <fpage>93</fpage>&#x2013;<lpage>113</lpage>, <year>2020</year>.</mixed-citation></ref>
</ref-list>
</back>
</article>