The Detection of Fake News in Arabic Tweets Using Deep Learning
Abstract
:1. Introduction
- Construction of a comprehensive, manually labelled dataset that contains Arabic fake and real news Tweets.
- Proposed a deep learning model for eliminating and detecting fake news on different topics.
- Exploited the news content and social context for more accurate detection of news credibility.
- Investigated the influence of contextual embeddings and compare them to classic word embeddings.
- Ascertained the generalization capability of the proposed model using a publicly available dataset.
- Verified the efficiency of the proposed model architecture by comparing it with current models.
2. Related Work
Ref | Year | Dataset | Topic | Classification Approach | Feature Type | Textual Feature Representations | Result | |
---|---|---|---|---|---|---|---|---|
News Content | Social Context | |||||||
[19] | 2018 | 800 Tweets | General | NB, SVM, DT | ✔ | ✔ | - | Accuracy 0.899 |
[20] | 2019 | 177 Tweets | General | EM | ✔ | ✔ | - | F1-score 0.80 |
[10] | 2019 | 268 labeled blog posts, 20,392 unlabeled blog posts | General | CNN | ✔ | Word2vec(CBOW), char-level embeddings | F1-score 0.63 | |
[21] | 2019 | 9000 Tweets | General | RF, SVM, DT, NB | ✔ | ✔ | - | F1-score 0.776 |
[22] | 2019 | 1862 Tweets | Syrian crisis | LR, RF, DT, AdaBoost | ✔ | ✔ | - | Accuracy 0.76 |
[11] | 2020 | 4547 news | General | LSTM, mBERT | ✔ | Word-level embeddings, char-level embeddings, mBERT | F1-score 0.643 | |
[12] | 2020 | AraNews (97,310 news), ATB (48,655 news), ANS (4547 news) | General | mBERT, AraBERT, XLM-RBase, XLM-RLarg | ✔ | mBERT, AraBERT, XLM-RBase, XLM-RLarg | F1-score 0.70 | |
[15] | 2020 | 6895 news articles | Political | NB, XGBoost, CNN | ✔ | BOW, TF-IDF, fastText | F1-score 0.984 | |
[1] | 2021 | 1862 Tweets | Syrian crisis | KNN, DT, NB, LR, LDA, SVM, RF, XGboost | ✔ | ✔ | TF, TF-IDF, BoW | Accuracy 0.82 |
[16] | 2021 | 37,000 Tweets | COVID-19 | NB, LR, SVM, MLP, RF, XGB | ✔ | BOW, TF-IDF | F1-score 0.933 | |
[6] | 2021 | 10,828 Tweets | COVID-19 | AraBERT, mBERT, distilBERT-multi, mBERT COV19, AraBERT COV19 | ✔ | AraBERT, mBERT, distilBERT-multi, mBERT COV19, AraBERT COV19 | F1-score 0.9578 | |
[14] | 2021 | COVID-19-Fakes (70,959 Tweets), ArCOV19-Rumors (3032 Tweets), ANS (4091 news), AraNews (108,194 news) | COVID-19, general | CNN, RNN, GRU, AraBERT v1, AraBERT v2, AraBERT v02, QARiB, Ar-Electra, Marbert, Arbert | ✔ | Word2vec, fastText, doc2vec, glove, AraBERT v1, AraBERT v2, AraBERT v02, QARiB, Ar-Electra, MARBERT, Arbert | F1-score 0.95 | |
[3] | 2021 | 8786 Tweets | COVID-19 | XGB, RF, NB, SVM, SGD, CNN, RNN, CRNN | ✔ | TF-IDF, word2vec, fastText | F1-score 0.54 | |
[17] | 2022 | 3157 Tweets | COVID-19 | LR, KNN, CART, SVM, NB, RF, AdaBoost, Bagging, ExtraTree | ✔ | ✔ | TF-IDF, glove | F1-score 0.935 |
[18] | 2022 | 4299 Tweets | COVID-19 | RF, DT, XGBoost, SVM, KNN, NB, SGD, LR, RNN, BiRNN, GRU, BiGRU, LSTM, BiLSTM | ✔ | N-Gram, TF-IDF, word2vec | Accuracy 0.81 | |
[13] | 2022 | 1098 news articles | Hajj | SVM, RF, NB | ✔ | - | F1-score 0.79 |
3. Methodology
3.1. Tweet Collection
3.2. Tweets Preprocessing
- The cleaning technique filters out the noisy data and irrelevant text. Cleaning the Tweets included the following steps:
- -
- Removing user mentions (@), hashtag signs (#), URLs, punctuation marks, white spaces, and emojis.
- -
- Removing non-Arabic words and characters.
- -
- Removing repeated letters of more than two letters.
- The normalization technique unifies the disparate shapes of characters into a single canonical form. This was performed by carrying out the following steps:
- -
- Replacing non-Arabic numerals with Arabic numbers.
- -
- Unifying variants of the Arabic letters (alif, waw, ya, and ha) into a single form.
3.3. Features Extraction
3.3.1. Textual Features Extraction
- Keras Embedding LayerEach word in the Keras embedding layer is represented by an embedding vector of a fixed length that is defined as part of the model, often 100, 200, or 300 dimensions. An embedding vector is learned during training the neural network on a specific NLP task. It requires each word to be represented using one-hot representation and padded with zeros for texts shorter than the maximum length. The embedding layer is initialized with small random numbers and learns the optimal values of the embeddings for all words using the backpropagation algorithm [29].
- Word2vecWord2vec is one of the most famous pre-trained word embeddings that is based on continuous bag-of-words (CBOW) or skip-gram (SG) learning models for obtaining vector representations for different words. These models use neural networks to learn word representation but differ in their input and output variables. CBOW learns word vectors by training the model to predict a target word given its context, while in the SG model, the input word is used to predict the surrounding context words [30]. In [31], the word2vec model is used to release pre-trained word vectors for the Arabic language called AraVec, which has been trained on two different Arabic content sources, Twitter and Wikipedia. In our experiment, we chose the version of AraVec built using Twitter data by applying CBOW for 100 and 300 dimensions with a window size of 3.
- FastTextThis is essentially an extension of the word2vec model. FastText treats each word as a bag of character n-grams instead of a word whole [29]. A vector representation is generated for each character n-gram, and the sum of these representations gives the final word vector. This feature enables it to find the vector representation for rare, misspelled, and out-of-vocabulary words [32]. FastText provides pre-trained word vectors for 157 different languages, including Arabic, which have been trained on datasets composed of a mixture of Common Crawl and Wikipedia [8]. These models were learned dense vectors using CBOW with a window size of 5 and character n-grams of length 5. In our experiment, we selected the 100–300-dimension Arabic version of pre-trained FastText embedding.
- ARBERT and MARBERTBERT (Bidirectional Encoder Representations from Transformers) is a bidirectional transformer model that was pre-trained on a massive unlabelled English dataset. The bidirectional capability of BERT means that the model can read each word from both the right and left side of a text statement simultaneously; this helps the model to truly and deeply perceive the meaning of the word by all of its surrounding words [33]. BERT has many pre-trained model variations that are trained on the source dataset with a different domain or language. These models can be used either to extract numerical representations of text data or to fine-tune these models on a diverse range of NLP tasks that may not be relevant to the task the model was trained on to provide state-of-the-art predictions [32]. A variety of pre-trained BERT-based language models for Arabic are available, namely AraBERT [34], QARiB [35], ARBERT, and MARBERT [9]. This work focuses on ARBERT and MARBERT to extract contextualized embeddings. ARBERT is trained on Modern Standard Arabic (MSA) text extracted from six resources covering different Arabic countries. MARBERT is a large-scale model trained on 6 billion Arabic Tweets. Therefore, it has shown promising results for social media tasks. The following paragraph will explain the steps to extracting contextualized embeddings using BERT-based models.First, the input text is passed to the tokenizer to truncate it into tokens. The tokenizer uses the word-piece vocabulary size of around 100K. The word-piece tokenizer starts by searching for the whole word in the vocabulary, and if it does not exist, it iteratively splits this word into sub-words until the predefined words are reached. The advantage of this is that it forces the model to represent any word, and hence overcome the OOV issue [33]. After decomposing the text into tokens, it transforms the list of tokens into numerical IDs corresponding to vocabulary indices, indicated as input IDs. To make all input sequences have a fixed length, padding is performed, and to distinguish between tokens and padding elements in the input sequence, the attention mask was created. Input IDs and attention masks are fed into the model, giving an embedding vector of 768 for each token as the output. Two different approaches were considered here to use their embeddings: word-level embedding, which encodes each token as a 768-dimensional vector of the numerical values, and sentence-level embedding, which averages all the tokens’ vectors to create a single 768-dimensional vector for the entire sentence.
3.3.2. User Features Extraction
3.4. Proposed Model Architecture
3.4.1. CNN Architecture
3.4.2. BiLSTM Architecture
3.4.3. Fully Connected Layers
4. Experiments and Results
4.1. Experimental Setup
4.2. Evaluation Metrics
- Accuracy (A) is the average between the number of samples classified accurately and the overall number of samples.
- Precision (P) is the average between the number of positive samples classified accurately and the overall number of samples classified as positive.
- Recall (R) is the average between the number of positive samples classified accurately and the overall number of samples that should be classified as positive.
- F1-score (F1) is the harmonic mean between the precision and recall scores.
4.3. Experimental Results
4.4. Impact of Oversampling and Undersampling
4.5. Discussion
4.6. Comparison with Models from the Literature
5. Conclusions and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Thaher, T.; Saheb, M.; Turabieh, H.; Chantar, H. Intelligent detection of false information in arabic tweets utilizing hybrid harris hawks based feature selection and machine learning models. Symmetry 2021, 13, 556. [Google Scholar] [CrossRef]
- Liu, Y.; Wu, Y.F.B. Fned: A deep network for fake news early detection on social media. ACM Trans. Inf. Syst. (TOIS) 2020, 38, 1–33. [Google Scholar] [CrossRef]
- Alqurashi, S.; Hamoui, B.; Alashaikh, A.; Alhindi, A.; Alanazi, E. Eating garlic prevents COVID-19 infection: Detecting misinformation on the Arabic content of Twitter. arXiv 2021, arXiv:2101.05626. [Google Scholar]
- Kaliyar, R.K.; Goswami, A.; Narang, P. DeepFakE: Improving fake news detection using tensor decomposition-based deep neural network. J. Supercomput. 2021, 77, 1015–1037. [Google Scholar] [CrossRef]
- Al-Sarem, M.; Alsaeedi, A.; Saeed, F.; Boulila, W.; AmeerBakhsh, O. A novel hybrid deep learning model for detecting COVID-19-related rumors on social media based on LSTM and concatenated parallel CNNs. Appl. Sci. 2021, 11, 7940. [Google Scholar] [CrossRef]
- Ameur, M.S.H.; Aliane, H. Aracovid19-mfh: Arabic COVID-19 multi-label fake news & hate speech detection dataset. Procedia Comput. Sci. 2021, 189, 232–241. [Google Scholar]
- Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. Adv. Neural Inf. Process. Syst. 2013, 26. [Google Scholar]
- Grave, E.; Bojanowski, P.; Gupta, P.; Joulin, A.; Mikolov, T. Learning word vectors for 157 languages. arXiv 2018, arXiv:1802.06893. [Google Scholar]
- Abdul-Mageed, M.; Elmadany, A.; Nagoudi, E.M.B. ARBERT & MARBERT: Deep bidirectional transformers for Arabic. arXiv 2020, arXiv:2101.01785. [Google Scholar]
- Helwe, C.; Elbassuoni, S.; Al Zaatari, A.; El-Hajj, W. Assessing arabic weblog credibility via deep co-learning. In Proceedings of the Fourth Arabic Natural Language Processing Workshop, Florence, Italy, August 2019; pp. 130–136. [Google Scholar]
- Khouja, J. Stance prediction and claim verification: An Arabic perspective. arXiv 2020, arXiv:2005.10410. [Google Scholar]
- Nagoudi, E.M.B.; Elmadany, A.; Abdul-Mageed, M.; Alhindi, T.; Cavusoglu, H. Machine generation and detection of Arabic manipulated and fake news. arXiv 2020, arXiv:2011.03092. [Google Scholar]
- Himdi, H.; Weir, G.; Assiri, F.; Al-Barhamtoshy, H. Arabic fake news detection based on textual analysis. Arab. J. Sci. Eng. 2022, 47, 10453–10469. [Google Scholar] [CrossRef] [PubMed]
- Al-Yahya, M.; Al-Khalifa, H.; Al-Baity, H.; AlSaeed, D.; Essam, A. Arabic fake news detection: Comparative study of neural networks and transformer-based approaches. Complexity 2021, 2021, 5516945. [Google Scholar] [CrossRef]
- Saadany, H.; Mohamed, E.; Orasan, C. Fake or real? A study of Arabic satirical fake news. arXiv 2020, arXiv:2011.00452. [Google Scholar]
- Mahlous, A.R.; Al-Laith, A. Fake news detection in Arabic tweets during the COVID-19 pandemic. Int. J. Adv. Comput. Sci. Appl. 2021, 12, 778–788. [Google Scholar] [CrossRef]
- Qasem, S.N.; Al-Sarem, M.; Saeed, F. An ensemble learning based approach for detecting and tracking COVID19 rumors. Comput. Mater. Contin. 2021, 70, 1721–1747. [Google Scholar]
- Amoudi, G.; Albalawi, R.; Baothman, F.; Jamal, A.; Alghamdi, H.; Alhothali, A. Arabic rumor detection: A comparative study. Alex. Eng. J. 2022, 61, 12511–12523. [Google Scholar] [CrossRef]
- Sabbeh, S.F.; Baatwah, S.Y. Arabic News Credibility on Twitter: An Enhanced Model Using Hybrid Features. J. Theor. Appl. Inf. Technol. 2018, 96, 2327–2338. [Google Scholar]
- Alzanin, S.M.; Azmi, A.M. Rumor detection in Arabic tweets using semi-supervised and unsupervised expectation–maximization. Knowl.-Based Syst. 2019, 185, 104945. [Google Scholar] [CrossRef]
- Mouty, R.; Gazdar, A. The effect of the similarity between the two names of twitter users on the credibility of their publications. In Proceedings of the 2019 Joint 8th International Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd International Conference on Imaging, Vision & Pattern Recognition (icIVPR), Spokane, WA, USA, 30 May–2 June 2019; pp. 196–201. [Google Scholar]
- Jardaneh, G.; Abdelhaq, H.; Buzz, M.; Johnson, D. Classifying Arabic tweets based on credibility using content and user features. In Proceedings of the 2019 IEEE Jordan International Joint Conference on Electrical Engineering and Information Technology (JEEIT), Amman, Jordan, 9–11 April 2019; pp. 596–601. [Google Scholar]
- Twitter API Documentation. Available online: https://developer.twitter.com/en/docs (accessed on 18 January 2022).
- Alharthi, R.; Alhothali, A.; Moria, K. A real-time deep-learning approach for filtering Arabic low-quality content and accounts on Twitter. Inf. Syst. 2021, 99, 101740. [Google Scholar] [CrossRef]
- Hegazi, M.O.; Al-Dossari, Y.; Al-Yahy, A.; Al-Sumari, A.; Hilal, A. Preprocessing Arabic text on social media. Heliyon 2021, 7, e06191. [Google Scholar] [CrossRef] [PubMed]
- Alwehaibi, A.; Roy, K. Comparison of pre-trained word vectors for arabic text classification using deep learning approach. In Proceedings of the 2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, 17–20 December 2018; pp. 1471–1474. [Google Scholar]
- El-Alami, F.z.; El Alaoui, S.O.; Nahnahi, N.E. Contextual semantic embeddings based on fine-tuned AraBERT model for Arabic text multi-class categorization. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 8422–8428. [Google Scholar] [CrossRef]
- Ashi, M.M.; Siddiqui, M.A.; Nadeem, F. Pre-trained word embeddings for Arabic aspect-based sentiment analysis of airline tweets. In Proceedings of the International Conference on Advanced Intelligent Systems and Informatics 2018, Cairo, Egypt, 1–3 September 2018; Springer: Cham, Switzerland, 2019; pp. 241–251. [Google Scholar]
- Bogale Gereme, F.; Zhu, W. Fighting fake news using deep learning: Pre-trained word embeddings and the embedding layer investigated. In Proceedings of the 2020 3rd International Conference on Computational Intelligence and Intelligent Systems, Tokyo, Japan, 13–15 November 2020; pp. 24–29. [Google Scholar]
- Saleh, H.; Alhothali, A.; Moria, K. Detection of hate speech using bert and hate speech word embedding with deep model. Appl. Artif. Intell. 2023, 37, 2166719. [Google Scholar] [CrossRef]
- Soliman, A.B.; Eissa, K.; El-Beltagy, S.R. Aravec: A set of arabic word embedding models for use in arabic nlp. Procedia Comput. Sci. 2017, 117, 256–265. [Google Scholar] [CrossRef]
- d’Sa, A.G.; Illina, I.; Fohr, D. Bert and fasttext embeddings for automatic detection of toxic speech. In Proceedings of the 2020 International Multi-Conference on: “Organization of Knowledge and Advanced Technologies” (OCTA), Tunis, Tunisia, 6–8 February 2020; pp. 1–5. [Google Scholar]
- Alammary, A.S. BERT models for Arabic text classification: A systematic review. Appl. Sci. 2022, 12, 5720. [Google Scholar] [CrossRef]
- Antoun, W.; Baly, F.; Hajj, H. Arabert: Transformer-based model for arabic language understanding. arXiv 2020, arXiv:2003.00104. [Google Scholar]
- Abdelali, A.; Hassan, S.; Mubarak, H.; Darwish, K.; Samih, Y. Pre-training bert on arabic tweets: Practical considerations. arXiv 2021, arXiv:2102.10684. [Google Scholar]
- Thompson, R.C.; Joseph, S.; Adeliyi, T.T. A Systematic Literature Review and Meta-Analysis of Studies on Online Fake News Detection. Information 2022, 13, 527. [Google Scholar] [CrossRef]
- Kaliyar, R.K.; Goswami, A.; Narang, P. FakeBERT: Fake news detection in social media with a BERT-based deep learning approach. Multimed. Tools Appl. 2021, 80, 11765–11788. [Google Scholar] [CrossRef]
- Rai, N.; Kumar, D.; Kaushik, N.; Raj, C.; Ali, A. Fake News Classification using transformer based enhanced LSTM and BERT. Int. J. Cogn. Comput. Eng. 2022, 3, 98–105. [Google Scholar] [CrossRef]
- Mohammed, R.; Rawashdeh, J.; Abdullah, M. Machine learning with oversampling and undersampling techniques: Overview study and experimental results. In Proceedings of the 2020 11th International Conference on Information and Communication Systems (ICICS), Irbid, Jordan, 7–9 April 2020; pp. 243–248. [Google Scholar]
- Al Zaatari, A.; El Ballouli, R.; ELbassouni, S.; El-Hajj, W.; Hajj, H.; Shaban, K.; Habash, N.; Yahya, E. Arabic corpora for credibility analysis. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC’16), Portoroz, Slovenia, 23–28 May 2016; pp. 4396–4401. [Google Scholar]
- Haouari, F.; Hasanain, M.; Suwaileh, R.; Elsayed, T. ArCOV19-rumors: Arabic COVID-19 twitter dataset for misinformation detection. arXiv 2020, arXiv:2010.08768. [Google Scholar]
Tweet Label | Tweet Text |
---|---|
Real | القوات الروسية تسيطر على مفاعل #تشيرنوبل النووي. #عاجل #اوكرانيا #روسيا Russian forces take control of the #Chornobyl nuclear reactor. #Urgent #Ukraine #Russia |
Fake | .خبر عاجل جداً #على حسب المرصد الإخباري، إصطدام قمر صناعي هندي بمحطة الفضاء الدولية أثناء صيانتها astronaut Michael Collins https://t.co/6ucb9alyWO Very urgent news #according to the news observatory, an Indian satellite collided with the International Space Station during its maintenance. astronaut Michael Collins https://t.co/6ucb9alyWO |
Real |
لاعبو منتخب #إنجلترا يثيرون حالة من الجدل خلال مواجهة #سويسرا.. بسبب خوض نجوم ”الأسود الثلاثة“ لجزء من المباراة بدون أسماء على قمصانهم #عينك_على_العالم https://t.co/EOPcdUhz1X The players of the #England national team raise a state of controversy during the confrontation with #Switzerland. because the “Three Lions” stars played part of the match without names on their shirts #YourEye_On_theWorld https://t.co/EOPcdUhz1X |
Fake | الجزائر تفرض حظر جوي وبحري على المملكة المغربية، بداية من الغد، ممنوع مرور السفن والطائرات النقل المغربية بالمجال الجوي والبحري الجزائري. Algeria has imposed an air and sea embargo on the Kingdom of Morocco; starting tomorrow, it is forbidden for Moroccan ships and planes to pass through Algerian air and sea space. |
Raw Tweet | Processed Tweet |
---|---|
Parameter | Value |
---|---|
Epoch | 40 |
Batch size | 64 |
Data split | Cross validation (Five folds) |
Validation split | 0.1 |
Early stopping | Monitor = validation loss, patience = 10 |
Processing Layer | Parameter | Experimental Range |
---|---|---|
CNN | Kernel size | [1, 2, 3, 4, 5, 6, 7] |
No of filters | [16, 32, 64, 128] | |
BiLSTM | No of LSTM units | [16, 32, 64, 128] |
Dense | No of neurons in the dense layer | [128, 256, 512, 1024] |
Dropout rate | [0.1–0.5] | |
Optimizer | [‘Adamax’, ‘Adam’, ‘SGD’] | |
Learning rate | [ 0.001, 0.002, 0.003, 0.004] |
Parameter | Value |
---|---|
Kernel size | 1, 2, 3 |
No of filters | 64 |
No of neurons in the dense layer | 512, 256 |
Dropout rate | 0.5 |
Optimizer | Adamax |
Learning rate | 0.004 |
Parameter | Value |
---|---|
No of LSTM units | 128 |
No of neurons in the dense layer | 512, 256 |
Dropout rate | 0.3 |
Optimizer | Adamax |
Learning rate | 0.002 |
Embedding Model | Embedding Level Method | Dim | CNN | |||
---|---|---|---|---|---|---|
Accuracy | Precision | Recall | F1-Score | |||
Keras Embedding Layer | Word_level | 100 | 0.9458 | 0.9462 | 0.9458 | 0.9459 |
200 | 0.9502 | 0.9500 | 0.9502 | 0.9497 | ||
300 | 0.9498 | 0.9500 | 0.9498 | 0.9496 | ||
Word2vec | Word_level | 100 | 0.9330 | 0.9324 | 0.9330 | 0.9315 |
300 | 0.9322 | 0.9311 | 0.9322 | 0.9307 | ||
FastText | Word_level | 100 | 0.9410 | 0.9407 | 0.9410 | 0.9405 |
200 | 0.9376 | 0.9331 | 0.9376 | 0.9370 | ||
300 | 0.9484 | 0.9482 | 0.9484 | 0.9481 | ||
ARBERT | Word_level | 768 | 0.9538 | 0.9534 | 0.9538 | 0.9531 |
Sentence_level | 0.9380 | 0.9375 | 0.9380 | 0.9368 | ||
MARBERT | Word_level | 768 | 0.9564 | 0.9564 | 0.9564 | 0.9563 |
Sentence_level | 0.9446 | 0.9462 | 0.9446 | 0.9448 |
Embedding Model | Embedding Level Method | Dim | BiLSTM | |||
---|---|---|---|---|---|---|
Accuracy | Precision | Recall | F1-Score | |||
Keras Embedding Layer | Word_level | 100 | 0.9478 | 0.9472 | 0.9478 | 0.9474 |
200 | 0.9444 | 0.9435 | 0.9444 | 0.9437 | ||
300 | 0.9454 | 0.9457 | 0.9454 | 0.9454 | ||
Word2vec | Word_level | 100 | 0.9470 | 0.9464 | 0.9470 | 0.9464 |
300 | 0.9460 | 0.9455 | 0.9460 | 0.9456 | ||
FastText | Word_level | 100 | 0.9300 | 0.9305 | 0.9300 | 0.9300 |
200 | 0.9336 | 0.9336 | 0.9336 | 0.9335 | ||
300 | 0.9356 | 0.9354 | 0.9356 | 0.9353 | ||
ARBERT | Word_level | 768 | 0.9548 | 0.9543 | 0.9548 | 0.9542 |
Sentence_level | 0.9476 | 0.9470 | 0.9476 | 0.9470 | ||
MARBERT | Word_level | 768 | 0.9548 | 0.9546 | 0.9548 | 0.9545 |
Sentence_level | 0.9426 | 0.9437 | 0.9426 | 0.9428 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Alyoubi, S.; Kalkatawi, M.; Abukhodair, F. The Detection of Fake News in Arabic Tweets Using Deep Learning. Appl. Sci. 2023, 13, 8209. https://doi.org/10.3390/app13148209
Alyoubi S, Kalkatawi M, Abukhodair F. The Detection of Fake News in Arabic Tweets Using Deep Learning. Applied Sciences. 2023; 13(14):8209. https://doi.org/10.3390/app13148209
Chicago/Turabian StyleAlyoubi, Shatha, Manal Kalkatawi, and Felwa Abukhodair. 2023. "The Detection of Fake News in Arabic Tweets Using Deep Learning" Applied Sciences 13, no. 14: 8209. https://doi.org/10.3390/app13148209
APA StyleAlyoubi, S., Kalkatawi, M., & Abukhodair, F. (2023). The Detection of Fake News in Arabic Tweets Using Deep Learning. Applied Sciences, 13(14), 8209. https://doi.org/10.3390/app13148209