Next Article in Journal
Assisting Glaucoma Screening Process Using Feature Excitation and Information Aggregation Techniques in Retinal Fundus Images
Previous Article in Journal
Evaluation of the Fast Synchrophasors Estimation Algorithm Based on Physical Signals
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Multitask Learning Framework with a Fire Hawk Optimizer for Arabic Fake News Detection

1
Department of Mathematics, Faculty of Science, Zagazig University, Zagazig 44519, Egypt
2
Faculty of Computer Science and Engineering, Galala University, Suez 435611, Egypt
3
Artificial Intelligence Research Center (AIRC), Ajman University, Ajman 346, United Arab Emirates
4
Department of Electrical and Computer Engineering, Lebanese American University, Byblos 13-5053, Lebanon
5
Mathematics and Computer Science Department, University of Ahmed DRAIA, Adrar 01000, Algeria
6
Faculty of Media Production, Galala University, Suez 435611, Egypt
7
Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
8
Department of Computer, Damietta University, Damietta 34517, Egypt
*
Authors to whom correspondence should be addressed.
Mathematics 2023, 11(2), 258; https://doi.org/10.3390/math11020258
Submission received: 17 November 2022 / Revised: 22 December 2022 / Accepted: 29 December 2022 / Published: 4 January 2023
(This article belongs to the Section Mathematics and Computer Science)

Abstract

:
The exponential spread of news and posts related to the COVID-19 pandemic on social media platforms led to the emergence of the disinformation phenomenon. The phenomenon of spreading fake information and news creates significant concern for the public health and safety of the population. In this paper, we propose a disinformation detection framework based on multi-task learning (MTL) and meta-heuristic algorithms in the context of the COVID-19 pandemic. The developed framework uses an MTL and a pre-trained transformer-based model to learn and extract contextual feature representations from Arabic social media posts. The extracted contextual representations are fed to an alternative feature selection technique which depends on modified version of the Fire Hawk Optimizer. The proposed framework, which aims to improve the disinformation detection rate, was evaluated on several datasets of Arabic social media posts. The experimental results show that the proposed framework can achieve accuracy of 59%. It obtained, at best, precision, recall, and F-measure of 53%, 71%, and 53%, respectively, on all datasets; and it outperformed the other algorithms in all measures.

1. Introduction

The presence of digital technology has led to changing of habits among people who previously used conventional media. Now, they are turning to use digital media to meet their needs, especially the need to access information. Clear evidence of this change can be seen from the rapid annual increase in the consumption of digital media [1]. In emergencies and urgent situations, such as the COVID-19 pandemic, many are disseminating information through social media channels; this has an impact on billions of persons worldwide. One of the most relevant characteristics of social media platforms in this pandemic has been the rapid dissemination of protocols at regional, national, and international levels. Sharing protocols about treatment, personal protective equipment, or the vaccine have now become the new normal [2]. A number of official organizations and governments used social media as a platform to disseminate information to the public [1]. Meanwhile the results of research about news consumption on social media in 2021 showed that more than half of Twitter user get news on the site regularly [3].
In the meantime, with the growth of using social media platforms as an information source, the significance of fact-checking journalism is increasing, especially with the diffusion of fake news [4]. Fake news is a dangerous and widespread issue in the news industry [5]. There has been a rise in public concern since the Brexit referendum in the United Kingdom—most especially after the 2016 US elections—that social media spreads misinformation [5]. Fake news is as old as one can remember. However, its growth has increased in recent times because of the broad usage of the Internet and cheap access to social media platforms [6]. With the characteristics of social media—low cost, easy access, and prompt dissemination of information—it has become the primary platform for interaction and exchanging information [7]. Due to the attractiveness of social media platforms, they became the ideal grounds for spreading fake news, so they gained the attention of researchers and academics. In recent years, to help online users identify correct and factual information, there have emerged automatic frameworks and online fact-checking resources to peform online fake news detection [8]. There is a growing body of research that focuses on fake news. The authors of [9] analyzed stories that were labeled as fake news, and they found out that these stories used moderate levels of sensationalism and misinformation, whereas complete fabrications were uncommon and did not go well with audiences. The authors of [10] investigated the spread of all the verified true and false news stories on Twitter from 2006 to 2017. The results show that false news is more novel than trustworthy news, and that it spreads more because humans are more likely to spread it.
After the global outbreak of COVID-19, the WHO characterized it as a pandemic in March 2020 (https://www.who.int/europe/emergencies/situations/COVID-19, accessed on 9 November 2022). This outbreak was accompanied by increased use of social media and the explosion of misinformation about the disease. In conclusion, the WHO stated that they are fighting an infodemic, not just a pandemic, due to the levels of rumors and misinformation and the wide spread of fake news (https://www.who.int/director-general/speeches/detail/munich-security-conference, accessed on 9 November 2022). Reference [11] concludes that fake news is pervasive on social media platforms and is characterized by false claims and conspiracy theories regarding the treatment, prevention, origin, and spread of the virus. Reference [6] analyzed a dataset of 10,700 social media posts on COVID-19. The top results show that nearly 50% of the sample was fake news, and real news was more prolonged than fake news in terms of the average number of words per post. In a comparison between Arabic and English tweets originating from Qatar, the authors of [12] stated that Arabic tweets contain much false information, but English tweets are more propagandistic. Furthermore, Arabic tweets adopted a health and safety perspective, whereas English ones focused on the economic consequences of COVID-19. As for engagement and interaction of posts related to COVID-19, reference [13] found that the highest numbers of posts and comments were on mainstream platforms, such as YouTube and Twitter. As for the main topics, debates included comparisons to other viruses, requests for God’s blessing, and racism; the most attention was given to the shutting down of flights. The authors of [14] studied the main channel Spanish people depend on to get information about COVID-19. Results indicated that the most credible sources of information were the online press and television, followed by institutional websites. WhatsApp and Facebook were considered more likely to present false news. To understand the motives behind sharing fake news, reference [15] concluded that the main reasons include low awareness, a lack of knowledge, and low trust in government news media.
Moreover, several feature selection (FS) methods based on metaheuristic (MH) techniques have been proposed to enhance the disinformation detection methods [16]. For example, the authors of [17] proposed a benchmarking studying fake news. In addition, FS has been used to enhance the detection of fake news within online social media based on artificial intelligence methods [18]. Ksieniewicz et al. [19] presented an alternative FS method to detect the fake news from data streams. However, these methods still need improvements to avoid limitations such as attraction to local points that affect the convergence towards to optimal solution. This motivated us to proposed an alternative FS method that depends on a modified version of the Fire Hawk Optimizer (FHO) [20]. According to the behavior of the FHO, it has been applied to various applications, including engineering problems and the BIM-Based Resource Tradeoff in Project Scheduling [21]. This indicates the strong ability of FHO to balance between exploration and exploitation during the search for the optimal subset of features.
Multi-task learning (MTL) has been used in different applications to simultaneously leverage and learn shared information from related tasks. In addition, MTL has been widely used to tackle natural language processing tasks and improve performance on multiple shared tasks, including sentiment analysis, event detection, and hate-speech detection [22,23]. The learning processes of MTL models can differ based on the model’s architecture, the data, and the task. In addition, MTL models can benefit from the recently introduced pre-trained language models that rely on transformer-based architectures, such as AraBERT for Arabic [24], which has been used for various NLP applications [25,26]. Meanwhile, the training data can be shared for multiple different tasks. In addition, MTL models can benefit from multimodality and be trained for different tasks and data types [27]. Moreover, the combination of evolutionary algorithms and multi-task learning has established its performance in several works, such as [28,29,30].
The proposed method depends on integration of deep learning and evolutionary methods to improve the prediction performance on the misinformation detection task. We implemented a transformer-based multi-task learning (MTL) model to tackle the misinformation detection task. The MTL model consists of two types of layers, shared and task-specific layers, to learn and extract contextual text representations. The MTL model can be trained on several datasets corresponding to different related tasks simultaneously, and share learned knowledge from each dataset across the shared layers. Meanwhile, the task-specific layers are used to perform feature extraction and classification in different datasets and tasks used to train the model. The extracted features corresponding to each task/dataset are fed to the feature selection algorithms to boost the performance and reduce the feature-representation space. Then, we use a modified version of the FHO to select the optimal features. Then, the selected subset of features which correspond to the ones in the best solution are used to reduce the features in the testing set. We evaluated the quality of the reduced testing set using different performance measures.
To summarize, the contributions of this study can be given as:
  • Developing a misinformation detection method for the social media data by integrating deep learning and meta-heuristic techniques.
  • Proposing a transformer-based MTL architecture for learning and extracting contextual text representations.
  • Proposing an alternative FS approach based on a modified version of FHO.
  • Evaluating the performance of the developed method using a set of real-world collected datasets with comparison with other well-known methods.
The rest of this paper is organized as follows: Section 2 presents the background of the Fire Hawk Optimizer. Section 3 introduces the steps of the proposed misinformation detection framework. In Section 4, the experimental results and discussion are given. Finally, the conclusion and future works are illustrated in Section 5.

2. Background

In this section, we introduce the basic steps of the Fire Hawk Optimizer (FHO). In general, the FHO, similarly to other MH techniques, starts by assigning an initial value for a set of N agents using the following formula.
X i j = r a n d × ( U j L j ) + L j , j = 1 , 2 , , D
where X i j refers to the ith agent at dimension j. U j and L j are the boundaries of the parameters at the jth dimension. r a n d [ 0 , 1 ] is a random value, and D is the dimension of each X i .
After that, the objective function is used to assess the performance of each X i . Then, we determine the fire hawks ( F H l , l = 1 , 2 , , n ), which represent the best solutions; the other solutions are referred to as prey ( P R k , k = 1 , 2 , , m ). We compute the distance between F H and P R using the following equation.
D l k = ( x 2 x 1 ) 2 + ( y 2 y 1 ) 2 , l = 1 , 2 , , n , k = 1 , 2 , , m
where m and n are the numbers of F H and P R , respectively. Then, the next task is to compute identify the territory of F H by dispersing P R . Then, we update each F H according to the following formula.
F H l ( t + 1 ) = F H l ( t ) + ( r 1 × X b r 2 × F H n ( t ) ) , l = 1 , 2 , , n
where X b refers to the best solution and F H n ( t ) is one fire hawk. r 1 and r 2 are random values generated from [0, 1].
The next step is to determine the safe area in which the prey meet together to remain safe and sound during a hazard. This can be formulated by computing S P l and S P as follows:
S P l = q = 1 r P R q r , q = 1 , 2 , , r , l = 1 , 2 , , n
Thereafter, the movement of P K inside the area of F H is used to simulate the animal behavior. According to this behavior, the prey can update its position using the following formula.
P R q ( t + 1 ) = P R q ( t ) + ( r 3 × F H l r 4 × S P l ( t ) ) , l = 1 , 2 , , n , q = 1 , 2 , , r
S P l denotes the safe position under the area of the lth fire hawk.
The next step is to update the safe location outside the lth F H , and this is formulated as:
S P = k = 1 m P R k r , k = 1 , 2 , , m
Then, update the position of prey using the following formula.
P R q ( t + 1 ) = P R q ( t ) + ( r 5 × F H a r 6 × S P ( t ) ) , l = 1 , 2 , , n , q = 1 , 2 , , r
The process of updating the solutions is repeated until reaching the stop conditions, and the best solution X b is returned to.

3. Proposed Misinformation Detection Framework

In this section, we present the steps of the developed method, which depends on the integration between the DL and modified FHO for feature selection. In general, the developed method consists of two stages. The first one is to extract the features from the text dataset. The second stage is to select the relevant features.

3.1. Feature Extraction Model

This section presents the implemented model based on multi-task learning (MTL) and a BERT pre-trained model to tackle the disinformation classification task on Arabic text data. In our study, we exploited the usage of an MTL model with a pre-trained AraBERT model on COVID-19 data. In addition, we relied on additional datasets to boost the model’s performance and generalization via adding datasets from similar domains, including the COVID-19 disinformation dataset, COVID-19 rumors dataset, fake news detection dataset, and offensive and hate speech datasets. We focused in our study on the task of disinformation classification using Twitter data related to COVID-19 covering multiple Arabic dialects. Thus, we relied on an AraBERT version 2 [24] model trained on 1.5 million COVID-19 multi-dialect Arabic tweets (https://huggingface.co/moha/arabert_arabic_covid19, accessed on 9 November 2022) as a shared representation for the proposed MTL model and to learn multiple related tasks jointly. In the upcoming paragraphs, we will detail the architecture and training process of the proposed MTL model for COVID-19 disinformation classification.
The implemented MTL model relies on a multi-task multi-corpora learning objective where the main objective is to improve the model’s performance on disinformation classification tasks with the help of additional tasks and corpora. In addition, the selected transformer-based model trained using AraBERT v2 on COVID-19 multi-dialect Arabic tweets can help to learn contextual representations related to the tackled task instead of training the MTL model from scratch using random initialization. The MTL model architecture is shown in Figure 1. The core component of the MTL model is the shared layers, initialized using a pre-trained version of AraBERT v2 on COVID-19 data. The shared layers benefit from the training on different tasks to learn general contextual representations simultaneously. In addition, the shared-layers-learned representations could help the task-specific layers to learn specific representations related to the tackled task and improve the model’s overall performance. Thus, the MTL model can overcome several single-task model training problems, including data shortage, context variation, class imbalance, and the lack of generalization on unseen data or tasks.

3.1.1. The MTL Shared Layers

The input data for the MTL model were processed similarly to the pre-trained multilayer bidirectional transformer encoder (BERT) model input using the SentencePiece [31] algorithm to perform input segmentation and relying on a lexicon encoder (pre-trained neural-based tokenization vocabulary). The SentencePiece is a multi-layered RNN (recurrent neural network) with the objective of mapping each token x from the input sentence X = x 1 , , x s of length s to three embedding vectors representing the word, the segment, and the positional (ID). At this stage, the lexicon encoder pads special tokens at the beginning and the end of the tokenized input sentence, which are [CLS] and [SEP], respectively. Finally, the word, segment, and positional embeddings are summed up for each token x to form the embedding vectors for X. The shared layer consists of 12 transformer encoders similar to the canonical BERT encoders used to fine-tune the weights of the pre-trained AraBERT v2 used in this study to learn shared global contextual representations in different tasks. For instance, each transformer encoder consists of a self-attention mechanism to gather contextual information for each input token x. Rather than learning a single objective function, as in single-task training, the MTL model shares representations learned by the transformer encoders with a multi-task objective. At the last transformer encoder, all the learned contextual embeddings for the sentence X are concatenated to generate a single semantic embedding vector stored in the [CLS] token and fed to the task-specific layer.

3.1.2. The MTL Task-Specific Layers

The task-specific layers are fully-connected layers randomly initialized and placed on top of the shared task layer corresponding to each learned task. The task-specific layers were used as single-sentence classifiers to perform sentence classification tasks followed by a softmax output layer and trained using the categorical cross-entropy loss. We perform feature extraction by placing a fully-connect layer before each task-specific layer to extract a representation vector of size 128 for each input sample. Later, the extracted representation vectors are fed to a feature selection algorithm to enhance the performance on specific tasks by removing nonrelevant features and reduce the feature space.
The Algorithm 1 demonstrates the training process of the proposed MTL model using a mini-batch Adamax optimizer with a 5 × 10 5 learning rate to update the model’s weights. All datasets are merged in mini-batches as dataset D, and in each epoch, a random mini-batch b a t c h T corresponding to task T is selected for fine-tuning the model’s weights using the task-specific objective, similarly to [32].
Algorithm 1 MTL learning process.
Require: D
for  e p o c h in 1,2,…, e p o c h m a x  do
      for  b a t c h T in D do
          Compute loss for task T
          Compute gradient
          Update model’s weights
      end for
end for

3.2. Feature Selection

Within this stage, the steps of presented FS method based on the modified version of FHO are introduced. The developed FS approach named BFHO starts by dividing the dataset into training and testing sets. Then, it uses the training set to find the relevant features, and this process begins by building the population of N solutions to compute the fitness value of each one. After that. it allocates the best solutions and uses it with the operators of FHO to update the population X. The next step is to select the relevant features from the testing set based on the binary version of the best solution. The details of these steps are given in the following subsections.

3.2.1. First Stage: Generating Population

The first step in this stage is to split the social data into training and testing samples, which represent 80% and 20%, respectively. After that, the population X with N solutions is generated, and each solution X i , i = 1 , 2 , , N has D dimensions, and the value of each X i is given as:
X i j = L j + r 1 × ( U j L j ) , i = 1 , 2 , , N , j = 1 , 2 , , D
In Equation (8), U j and L j denote the maximum and minimum values in the search space at dimension j.

3.2.2. Second Stage: Updating Solutions

In this section, the population X is updated through using the operators of modified FHO. This stage begins by converting each X i into Boolean form using the following formula.
B X i j = 1 i f X i j > 0.5 0 o t h e r w i s e
Thereafter, the features of training set that corresponding to ones in B X i j are selected and evaluated using the following fitness value ( F i t i ).
F i t i = ρ × γ + 1 ρ × | B X i j | D
where ρ [ 0 , 1 ] is a weight parameter used to balance between the two parts of Equation (10). γ denotes the classification error obtained using the KNN classifier with the training set. | B X i , j | and D refer to the number of selected features and the total number features in the dataset, respectively.
The next process is to determine the best solution X b which has best F i t b . After that, the solutions X are updated using the operators of the FHO defined in Equations (2)–(7). The updating steps are conduced until stop conditions are reached.

3.2.3. Third Stage: Evaluation of Selected Features ( X b )

In this stage, the features of the testing set which correspond to ones in the best solution X b are selected to evaluate their quality. This is achieved using the same fitness value defined in Equation (10). Then, the quality of predicted output using different performance measures is computing. Algorithm 2 provides the steps of the developed FS approach based on the BFHO.

3.3. Implementation of Disinformation Detection Framework

As described in the above-mentioned section, the proposed disinformation detection framework for Arabic language consists of two main phases, which are feature extraction (text representation learning) and feature selection. In this section, a summary of the proposed implementation is given. We started by learning and extracting contextual text representations using an MTL model which benefits from training on several data sources at the same time. In addition, the MTL model learned to generate contextual representations from different tasks while relying on the training mechanism introduced in the shared layers. The shared layers were initialized using a pre-trained language model trained on a large amount of Arabic text. Later, the learned contextual representations were fine-tuned based on their associated tasks using the MT task-specific layers; a feature extraction layer was added to extract the fine-tuned representations and feed them to the BFHO. Then, the steps of BFHO method that are defined in Algorithm 2 were used to determine the relevant features and evaluate their quality using various performance measure.
Algorithm 2 The FS based on the BFHO method.
  Input: number of solutions (N), social data contains D features, number of iterations ( t m a x ), and parameters of FHO.
  First Stage
  Split the data into two sets training and testing sets.
  Generate population X based on Equation (8).
  Second Stage
  Assign t = 1 .
  while ( t < t m a x ) do
      Generate Boolean form of X i based on Equation (9).
      Calculate fitness value of X i based on training samples as in Equation (10).
      Find the best solution X b which has the smallest fitness value.
      update X using Equations (2)–(7).
       t = t + 1 .
  end while
  Third Stage
  Select the features of testing set which corresponding to ones in X b .
  Compute the quality of predicted output using performance metrics.

4. Experimental Results

4.1. Description of Datasets

In our experiments, we used a multilingual dataset from [33] having 16K annotated tweets related to COVID-19 used for disinformation analysis collected from January 2020 to March 2021. Moreover, the dataset can be used for other tasks, such as fact checking. The dataset contains seven class labels, including questions Q1–Q7. Table 1 lists the questions and their corresponding tweets. We used this dataset in our primary task to optimize focusing on Arabic tweets while only tackling binary and multi-class classification tasks. For instance, each class question (e.g., Q1) contains a question such as “Does the tweet contain a verifiable factual claim?” and the corresponding tweets are annotated as “Yes” or “No” to indicate whether the tweet contains a harmful factual claim related to COVID-19 or not. The question classes can be used for binary and multi-class classification based on the question type and the answers provided by the annotators. Meanwhile, OSACT, ArCOV19, and FKD datasets were used in our experiments to improve the contextual representations learned by the proposed MTL model and refine the learned embeddings based on data from a similar domain as the main task (Q1–Q7). The OSACT [34] dataset was created in a shared task called Open-Source Arabic Corpora and Corpora Processing Tools, which contains two sub-tasks, offensive detection (OSACT-OFF) and hate-speech detection (OSACT-HS), for the Arabic language. The OSACT-OFF is a binary classification of Arabic tweets as containing or not containing offensive speech (OFF or NOT OFF), and with OSACT-HS, the task is to detect hate speech in a tweet (HS or NOT HS). The ArCOV19 [35] dataset was used for the Arabic COVID-19 rumors tweet verification task; the dataset was collected for misinformation detection from 27th January till the end of April 2020. The FKD [36] dataset was used for the Arabic COVID-19 fake news detection task. Table 2 and Table 3 list the statistics of each dataset used in our experiments.

4.2. Results and Discussion

This section evaluates the proposed FHO method in classifying a set of datasets. The FHO method received the datasets in a text form; then, it extracted the features from these data as described in Section 3. Therefore, the output of this stage produced a numerical dataset. After that, the FHO method started a new stage to select the most important features within the dataset. The produced features were evaluated using binary and multi-class classification, as described in this section.
The experimental results of the FHO are compared to those of nine methods: HBA, bGWO, QChamRSA, RunEFO, RSA, LSHADE_cnEpSin, SaDE, LSHADE_SPACMA, and CHCLPSO. Four performance measures were used: accuracy, precision, recall, and F-measure. The experiments were divided into two parts. The first used binary-class datasets, whereas the second used multi-class datasets. The results are presented in Table 4, Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10.
For the binary-class experiment, Table 4 shows the classification accuracy results for FHO and other methods. In this measure, the FHO achieved the best accuracy results and was ranked first for 6 out of 11 datasets (i.e., HS, OFF, Q2, Q4, Q6, and Q7); FHO, HBA, and bGWO obtained the best accuracy on the FKD dataset. The HBA came in second; it obtained the best accuracy on both ArCR and Q3 datasets. LSHADE_SPACMA and RSA obtained the third and fourth ranks, respectively, followed by SaDE, bGWO, LSHADE_cnEpSin, CHCLPSO, and EFO; QChamRSA shows the worst results. Figure 2 illustrates the average accuracy. The results of the precision measure are listed in Table 5. In this table, we can see that the FHO outperformed the other methods in 5 out of 11 datasets: ArCR, FKD, Q3, Q4, and Q6. It shows promising results on the other datasets. The HBA obtained the second rank, achieving the best precision results on three datasets HS, Q1, and Q2. The RSA had the best results for two datasets (OFF and Q7); therefore, it was ranked third, followed by LSHADE_SPACMA, SaDE, bGWO, CHCLPSO, and EFO, respectively. Figure 3 illustrates the average precision.
The results of the recall measure are listed in Table 6. In this table, the FHO shows the best recall results for 72% of all datasets, whereas for FKD datasets, the FHO, HBA, and bGWO obtained the same recall values equal 0.948. The HBA and RSA are ranked second and third, respectively, followed by EFO, bGWO, SaDE, LSHADE_SPACMA, LSHADE_cnEpSin, and CHCLPSO. Figure 4 illustrates the average of recall measures. In addition, the results of the F-measure are also considered and recorded in Table 7. This measure is used as the weighted mean of both precision and recall. In this measure, the FHO is ranked first; it obtained the best F-measure in 4 out of 11 datasets and showed the best results, equal with those of the HBA and bGWO, for the FKD dataset. The second and third-best algorithms were the HBA and RSA, followed by LSHADE_SPACMA, SaDE, EFO, bGWO, and LSHADE_cnEpSin. The QChamRSA had the worst F-measure results.
In terms of the multi-class experiment, six datasets were used in this experiment (Q2, Q3, Q4, Q5, Q6, and Q7). The results are listed Table 8, Table 9 and Table 10. The results of the classification accuracy measure are shown in Table 8. In this table, the FHO is shown to have achieved the best accuracy results and is ranked first for 4 out of 6 datasets (i.e., Q2, Q3, Q4, and Q7). The HBA obtained the second rank and achieved the best accuracy in two datasets (Q5 and Q6). The third and fourth-best methods were LSHADE_SPACMA and CHCLPSO, respectively, followed by bGWO, LSHADE_cnEpSin, SaDE, EFO, and RSA. The QChamRSA failed to obtain the best value in any dataset. Figure 5 illustrates the average accuracy measure.
The same results are shown for the recall measure listed in Table 10. For this measure, the FHO had the best recall results in 66% of all datasets, whereas the HBA and LSHADE_ cnEpSinwere were ranked second and third, followed by LSHADE_SPACMA, CHCLPSO, SaDE, and bGWO, respectively.
These results indicate that the proposed FHO can correctly classify different types of datasets in both experiments (binary-class and multi-class classification) and can obtain better classification accuracy compared to the other methods.

5. Conclusions and Future Works

This paper presents a framework using three methods, namely, multi-task learning, a transformer-based model, and an optimization algorithm, to tackle the misinformation detection problem and help decrease the COVID-19 infodemic in the Arabic community on social media. The framework combines several tasks to learn contextual feature representations from Arabic social media posts. The proposed framework relies on two core phases: feature extraction and selection. The feature extraction is performed using a pre-trained AraBERT model via fine-tuning and a multi-task learning approach. Several tasks and datasets were incorporated to enhance this phase to extract meaningful feature representations. For the feature selection phase, we developed a metaheuristic algorithm to select the most relevant features from the contextual feature representations and boost the detection accuracy of the framework. The algorithm used in this phase was the Fire Hawk Optimizer (FHO). After applying these phases, the proposed method was evaluated in classifying the obtained features, and its results were compared to the state-of-the-art optimization algorithms. The results showed that the proposed framework outperforms the compared approaches in terms of accuracy and F1-score, and other existing optimization algorithms. In future work, we plan to extend the proposed framework with other languages and natural language processing tasks, such as hate-speech detection and sentiment analysis. We also plan to investigate the proposed framework using multimodal data. In terms of optimization algorithms, we plan to modify the performance of the FHO using chaotic maps or opposite-based optimization techniques.

Author Contributions

Conceptualization, M.A.E., A.D., D.A.O., S.A., E.M.S. and A.A.E.; methodology, M.A.E., A.D., D.A.O., S.A., E.M.S. and A.A.E.; software, A.D., A.A.E. and M.A.E.; validation, A.D., A.A.E. and M.A.E.; formal analysis, M.A.E., A.D., D.A.O., S.A., E.M.S. and A.A.E.; investigation, M.A.E., A.D., D.A.O., S.A., E.M.S. and A.A.E.; writing—original draft preparation, D.A.O., S.A., A.D., A.A.E. and M.A.E.; writing—review and editing, M.A.E., A.D., D.A.O., S.A., E.M.S. and A.A.E.; visualization, M.A.E., A.D., D.A.O., S.A., E.M.S. and A.A.E.; project administration, M.A.E., A.D., D.A.O., S.A., E.M.S. and A.A.E.; and funding acquisition, S.A. All authors have read and agreed to the published version of the manuscript.

Funding

Princess Nourah bint Abdulrahman University Researchers Supporting Project (PNURSP2023R197), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Data Availability Statement

It is available upon request.

Conflicts of Interest

The authors declare that there are no conflict of interest regarding the publication of this paper.

References

  1. Arisanty, M.; Wiradharma, G.; Fiani, I. Optimizing Social Media Platforms as Information Disemination Media. J. ASPIKOM 2020, 5, 266–279. [Google Scholar] [CrossRef]
  2. González-Padilla, D.A.; Tortolero-Blanco, L. Social media influence in the COVID-19 Pandemic. Int. Braz J. Urol. 2020, 46, 120–124. [Google Scholar] [CrossRef] [PubMed]
  3. Walker, M.; Matsa, K.E. News Consumption across Social Media in 2021. Available online: https://www.pewresearch.org/journalism/2021/09/20/news-consumption-across-social-media-in-2021 (accessed on 9 November 2022).
  4. Birks, J. Fact-Checking Journalism and Political Argumentation: A British Perspective; Springer: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
  5. Valenzuela, S.; Halpern, D.; Katz, J.E.; Miranda, J.P. The paradox of participation versus misinformation: Social media, political engagement, and the spread of misinformation. Digit. J. 2019, 7, 802–823. [Google Scholar] [CrossRef]
  6. Pate, U.A.; Gambo, D.; Ibrahim, A.M. The impact of fake news and the emerging post-truth political era on Nigerian polity: A review of literature. Stud. Media Commun. 2019, 7, 21–29. [Google Scholar] [CrossRef] [Green Version]
  7. Shu, K.; Sliva, A.; Wang, S.; Tang, J.; Liu, H. Fake news detection on social media: A data mining perspective. ACM SIGKDD Explor. Newsl. 2017, 19, 22–36. [Google Scholar] [CrossRef]
  8. Zhang, X.; Ghorbani, A.A. An overview of online fake news: Characterization, detection, and discussion. Inf. Process. Manag. 2020, 57, 102025. [Google Scholar] [CrossRef]
  9. Mourão, R.R.; Robertson, C.T. Fake news as discursive integration: An analysis of sites that publish false, misleading, hyperpartisan and sensational information. J. Stud. 2019, 20, 2077–2095. [Google Scholar] [CrossRef]
  10. Vosoughi, S.; Roy, D.; Aral, S. The spread of true and false news online. Science 2018, 359, 1146–1151. [Google Scholar] [CrossRef]
  11. Naeem, S.B.; Bhatti, R.; Khan, A. An exploration of how fake news is taking over social media and putting public health at risk. Health Inf. Libr. J. 2021, 38, 143–149. [Google Scholar] [CrossRef]
  12. Nakov, P.; Alam, F.; Shaar, S.; Martino, G.D.S.; Zhang, Y. A second pandemic? Analysis of fake news about COVID-19 vaccines in Qatar. arXiv 2021, arXiv:2109.11372. [Google Scholar]
  13. Cinelli, M.; Quattrociocchi, W.; Galeazzi, A.; Valensise, C.M.; Brugnoli, E.; Schmidt, A.L.; Zola, P.; Zollo, F.; Scala, A. The COVID-19 social media infodemic. Sci. Rep. 2020, 10, 16598. [Google Scholar] [CrossRef] [PubMed]
  14. Fernández-Torres, M.J.; Almansa-Martínez, A.; Chamizo-Sánchez, R. Infodemic and fake news in Spain during the COVID-19 pandemic. Int. J. Environ. Res. Public Health 2021, 18, 1781. [Google Scholar] [CrossRef] [PubMed]
  15. Balakrishnan, V.; Zhen, N.W.; Chong, S.M.; Han, G.J.; Lee, T.J. Infodemic and fake news—A comprehensive overview of its global magnitude during the COVID-19 pandemic in 2021: A scoping review. Int. J. Disaster Risk Reduct. 2022, 78, 103144. [Google Scholar] [CrossRef] [PubMed]
  16. Tompkins, J. Disinformation Detection: A review of linguistic feature selection and classification models in news veracity assessments. arXiv 2019, arXiv:1910.12073. [Google Scholar]
  17. Gravanis, G.; Vakali, A.; Diamantaras, K.; Karadais, P. Behind the cues: A benchmarking study for fake news detection. Expert Syst. Appl. 2019, 128, 201–213. [Google Scholar] [CrossRef]
  18. Ozbay, F.A.; Alatas, B. Fake news detection within online social media using supervised artificial intelligence algorithms. Phys. A Stat. Mech. Its Appl. 2020, 540, 123174. [Google Scholar] [CrossRef]
  19. Ksieniewicz, P.; Zyblewski, P.; Choraś, M.; Kozik, R.; Giełczyk, A.; Woźniak, M. Fake news detection from data streams. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; pp. 1–8. [Google Scholar]
  20. Azizi, M.; Talatahari, S.; Gandomi, A.H. Fire Hawk Optimizer: A novel metaheuristic algorithm. Artif. Intell. Rev. 2022, 1–77. [Google Scholar] [CrossRef]
  21. Shishehgarkhaneh, M.B.; Azizi, M.; Basiri, M.; Moehler, R.C. BIM-Based Resource Tradeoff in Project Scheduling Using Fire Hawk Optimizer (FHO). Buildings 2022, 12, 1472. [Google Scholar] [CrossRef]
  22. Chen, S.; Zhang, Y.; Yang, Q. Multi-task learning in natural language processing: An overview. arXiv 2021, arXiv:2109.09138. [Google Scholar]
  23. Aldjanabi, W.; Dahou, A.; Al-qaness, M.A.; Elaziz, M.A.; Helmi, A.M.; Damaševičius, R. Arabic Offensive and Hate Speech Detection Using a Cross-Corpora Multi-Task Learning Model. Informatics 2021, 8, 69. [Google Scholar] [CrossRef]
  24. Antoun, W.; Baly, F.; Hajj, H. Arabert: Transformer-based model for arabic language understanding. arXiv 2020, arXiv:2003.00104. [Google Scholar]
  25. Chennafi, M.E.; Bedlaoui, H.; Dahou, A.; Al-qaness, M.A. Arabic Aspect-Based Sentiment Classification Using Seq2Seq Dialect Normalization and Transformers. Knowledge 2022, 2, 388–401. [Google Scholar] [CrossRef]
  26. Adel, H.; Dahou, A.; Mabrouk, A.; Abd Elaziz, M.; Kayed, M.; El-Henawy, I.M.; Alshathri, S.; Amin Ali, A. Improving crisis events detection using distilbert with hunger games search algorithm. Mathematics 2022, 10, 447. [Google Scholar] [CrossRef]
  27. Hu, R.; Singh, A. Unit: Multimodal multitask learning with a unified transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 1439–1449. [Google Scholar]
  28. Ong, Y.S.; Gupta, A. Evolutionary multitasking: A computer science view of cognitive multitasking. Cogn. Comput. 2016, 8, 125–142. [Google Scholar] [CrossRef]
  29. Li, H.; Ong, Y.S.; Gong, M.; Wang, Z. Evolutionary multitasking sparse reconstruction: Framework and case study. IEEE Trans. Evol. Comput. 2018, 23, 733–747. [Google Scholar] [CrossRef]
  30. Chandra, R.; Gupta, A.; Ong, Y.S.; Goh, C.K. Evolutionary multi-task learning for modular knowledge representation in neural networks. Neural Process. Lett. 2018, 47, 993–1009. [Google Scholar] [CrossRef]
  31. Kudo, T.; Richardson, J. Sentencepiece: A simple and language independent subword tokenizer and detokenizer for neural text processing. arXiv 2018, arXiv:1808.06226. [Google Scholar]
  32. Liu, X.; He, P.; Chen, W.; Gao, J. Multi-task deep neural networks for natural language understanding. arXiv 2019, arXiv:1901.11504. [Google Scholar]
  33. Alam, F.; Shaar, S.; Dalvi, F.; Sajjad, H.; Nikolov, A.; Mubarak, H.; Martino, G.D.S.; Abdelali, A.; Durrani, N.; Darwish, K.; et al. Fighting the COVID-19 infodemic: Modeling the perspective of journalists, fact-checkers, social media platforms, policy makers, and the society. arXiv 2020, arXiv:2005.00033. [Google Scholar]
  34. Husain, F. OSACT4 Shared Task on Offensive Language Detection: Intensive Preprocessing-Based Approach. arXiv 2020, arXiv:2005.07297. [Google Scholar]
  35. Haouari, F.; Hasanain, M.; Suwaileh, R.; Elsayed, T. ArCOV19-rumors: Arabic COVID-19 twitter dataset for misinformation detection. arXiv 2020, arXiv:2010.08768. [Google Scholar]
  36. Mahlous, A.R.; Al-Laith, A. Fake news detection in Arabic tweets during the COVID-19 pandemic. Int. J. Adv. Comput. Sci. Appl. 2021, 12, 776–785. [Google Scholar] [CrossRef]
Figure 1. The feature extraction model architecture based on MTL and AraBERT.
Figure 1. The feature extraction model architecture based on MTL and AraBERT.
Mathematics 11 00258 g001
Figure 2. Average of the accuracy measure in the binary case.
Figure 2. Average of the accuracy measure in the binary case.
Mathematics 11 00258 g002
Figure 3. Average of the precision measure in the binary case.
Figure 3. Average of the precision measure in the binary case.
Mathematics 11 00258 g003
Figure 4. Average of the recall measure in the binary case.
Figure 4. Average of the recall measure in the binary case.
Mathematics 11 00258 g004
Figure 5. Average of the accuracy measure for the multi-class experiment.
Figure 5. Average of the accuracy measure for the multi-class experiment.
Mathematics 11 00258 g005
Table 1. The main datasets’ descriptions and their corresponding samples.
Table 1. The main datasets’ descriptions and their corresponding samples.
QuestionClassesBinary InstancesMulti-Class InstancesTask
Q1: Does the tweet contain a verifiable factual claim?24966-Binary (Yes or No)
Q2: To what extent does the tweet appear
to contain false information?
434173439Binary/Multi-class
Q3: Will the tweet’s claim have an impact on
or be of interest to the general public?
434253439Binary/Multi-class
Q4: To what extent does the tweet appear to be
harmful to the society, a person(s), a company(s) or a product(s)?
428703439Binary/Multi-class
Q5: Do you think that a professional fact-checker
should verify the claim in the tweet?
434393439Binary/Multi-class
Q6: Is the tweet harmful to the society and why?949544966Binary/Multi-class
Q7: Do you think that this tweet should get
the attention of a government entity?
1049544966Binary/Multi-class
Table 2. Descriptions and statistics for OSACT, ArCOV19, and FKD datasets.
Table 2. Descriptions and statistics for OSACT, ArCOV19, and FKD datasets.
DatasetLanguageTotal SamplesLabelTraining SetDevelopment SetTest Set
OSACT-HSArabic MSA
Arabic DA
10KHS
NOT HS
361
6639
44
956
101
1899
OSACT-OFFArabic MSA
Arabic DA
10KOFF
NOT OFF
1410
5590
179
821
402
1598
ArCOV19Arabic MSA
Arabic DA
3584False
True
1220
1288
261
277
272
266
FKDArabic MSA
Arabic DA
1537Fake
NOT Fake
593
482
113
118
129
102
Table 3. Descriptions and statistics for the tested dataset.
Table 3. Descriptions and statistics for the tested dataset.
DatasetLanguageLabelsTotal SamplesTraining SetDevelopment SetTest Set
Q1Arabic MSA
Arabic DA
218091172320317
Q2Arabic MSA
Arabic DA
51355891220244
Q3Arabic MSA
Arabic DA
51360893223244
Q4Arabic MSA
Arabic DA
51444892219243
Q5Arabic MSA
Arabic DA
41365896223246
Q6Arabic MSA
Arabic DA
918031171317315
Q7Arabic MSA
Arabic DA
1018051169319317
Table 4. Accuracy results obtained by FHO and other methods in the binary case.
Table 4. Accuracy results obtained by FHO and other methods in the binary case.
AccFHOHBAbGWOQChamRSAEFORSALSESSaDELSSCCHCLPSO
ArCR0.95540.96280.95540.95540.95540.95170.95350.95910.95540.9554
FKD0.94830.94830.94830.89660.91380.91380.93100.91380.91380.9138
HS0.98200.97600.97400.97400.97400.98000.97400.97600.97600.9760
OFF0.94800.94600.94000.93000.94400.94200.94200.94200.94200.9440
Q10.88010.88010.87700.84860.87380.88330.87380.87700.87380.8644
Q20.87300.86890.84840.81150.84840.84430.86070.85250.85660.8484
Q30.97130.97540.96310.96310.96720.97130.97130.96720.96720.9672
Q40.95930.93420.91360.90530.91360.91360.91360.91770.91770.9136
Q50.72360.69510.72360.64630.69920.67890.70730.70330.72760.6992
Q60.96980.93330.91430.88250.89840.90790.90480.90480.90480.9048
Q70.81070.80760.77600.76030.78860.77920.78550.79180.79500.7886
Table 5. Precision results obtained by FHO and other methods in the binary case.
Table 5. Precision results obtained by FHO and other methods in the binary case.
FHOHBAbGWOQChamRSAEFORSALSESSaDELSSCCHCLPSO
ArCR0.96280.95550.95540.95540.95190.9610.95360.95910.95540.9554
FKD0.95310.95310.95310.91430.92650.93940.93940.91830.91830.9265
HS0.86670.88950.87430.85740.86980.88360.85740.88360.88360.8836
OFF0.91220.9170.91240.89050.90830.92320.91430.90830.91120.9195
Q10.83570.84470.8320.78160.83930.82950.8310.83470.82820.81
Q20.81410.82350.78140.71150.78210.81730.81160.78760.7970.7787
Q30.98770.73760.48550.48550.73760.73760.73760.65420.48560.4856
Q40.93380.92020.89670.87820.9030.91010.89670.90050.90660.8967
Q50.66530.69770.69770.62330.64930.69350.68150.6810.70240.6738
Q60.89760.88760.86750.82940.86350.8710.85990.86350.85990.8675
Q70.76460.75010.72040.69790.7240.76480.73440.74570.75120.7386
Table 6. Recall results obtained by FHO and other methods in the binary case.
Table 6. Recall results obtained by FHO and other methods in the binary case.
FHOHBAbGWOQChamRSAEFORSALSESSaDELSSCCHCLPSO
ArCR0.96290.95550.95540.95540.95180.9610.95360.95910.95550.9554
FKD0.94830.94830.94830.89660.91380.9310.9310.91380.91380.9138
HS0.94830.87080.94830.89660.91380.9310.9310.91380.91380.9138
OFF0.90880.87990.88430.8740.8980.90170.88970.8980.89390.8909
Q10.79580.79250.74690.77080.86980.77190.77080.77190.77190.7719
Q20.80750.76820.80050.78720.81460.80550.78840.79550.79350.7874
Q30.76560.56930.72380.67890.69960.74140.73160.73360.73620.731
Q40.87380.86210.84770.84240.84120.85940.84770.85680.85030.8477
Q50.66270.69770.69770.63010.65010.69450.68510.68980.70340.6788
Q60.89290.88290.86310.78370.84130.82940.83330.82740.83330.8214
Q70.73390.72730.6910.67660.71420.71770.70110.7020.70410.7068
Table 7. F-measure results obtained by FHO and other methods in the binary case.
Table 7. F-measure results obtained by FHO and other methods in the binary case.
FHOHBAbGWOQChamRSAEFORSALSESSaDELSSCCHCLPSO
ArCR0.96280.95540.95540.95540.95170.9610.95350.95910.95540.9554
FKD0.94810.94810.94810.89540.91310.93070.93070.91360.91360.9131
HS0.82710.91340.79630.80750.86980.81730.80750.81730.81730.8173
OFF0.91050.910.89750.88190.9030.91190.90130.9030.90220.9043
Q10.82020.8140.81450.78440.82590.81650.80650.81240.80860.7977
Q20.78570.79060.74580.69190.72680.76920.760.75480.75950.7501
Q30.61880.60380.52170.63170.60380.60380.60380.59160.61380.5529
Q40.89930.88670.86880.85830.86680.88130.86880.87590.87410.8688
Q50.66390.69770.69770.62480.64970.6940.68310.68380.70290.6759
Q60.89520.88520.86530.80320.85170.84780.84560.84360.84560.8415
Q70.74610.73680.7020.68490.71870.73410.71340.7170.720.7188
Table 8. Accuracy results for the multi-class experiment.
Table 8. Accuracy results for the multi-class experiment.
FHOHBAbGWOQChamRSAEFORSALSESSaDELSSCCHCLPSO
Q20.79270.78050.78460.76420.76420.76020.77240.76830.77640.7805
Q30.66260.64230.63010.58540.61790.60980.63010.63410.63820.6179
Q40.54070.52030.48780.46750.49190.47970.50.49590.50410.5041
Q50.60980.64230.60570.57320.60570.61380.63010.61380.61790.622
Q60.84540.87070.85490.82970.83910.83280.84540.8360.84540.8486
Q70.60580.60320.59870.55560.58730.60.57320.58920.59740.5924
Table 9. Precision results for the multi-class experiment.
Table 9. Precision results for the multi-class experiment.
FHOHBAbGWOQChamRSAEFORSALSESSaDELSSCCHCLPSO
Q20.79270.78050.78460.76420.76420.76020.77240.76830.77640.7805
Q30.66260.64230.63010.58540.61790.60980.63010.63410.63820.6179
Q40.54070.52030.48780.46750.49190.47970.50.49590.50410.5041
Q50.60980.64230.60570.57320.60570.61380.63010.61380.61790.622
Q60.84540.64230.60570.57320.60570.61380.63010.61380.61790.622
Q70.59620.87070.85490.82970.83910.83280.84540.8360.84540.8486
Table 10. Recall results for the multi-class experiment.
Table 10. Recall results for the multi-class experiment.
FHOHBAbGWOQChamRSAEFORSALSESSaDELSSCCHCLPSO
Q20.79270.78050.78460.76420.76420.76020.77240.76830.77640.7805
Q30.66260.64230.63010.58540.61790.60980.63010.63410.63820.6179
Q40.54070.52030.48780.46750.49190.47970.50.49590.50410.5041
Q50.60980.64230.60570.57320.60570.61380.63010.61380.61790.622
Q60.84540.64230.60570.57320.60570.61380.63010.61380.61790.622
Q70.59620.87070.85490.82970.83910.83280.84540.8360.84540.8486
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Abd Elaziz, M.; Dahou, A.; Orabi, D.A.; Alshathri, S.; Soliman, E.M.; Ewees, A.A. A Hybrid Multitask Learning Framework with a Fire Hawk Optimizer for Arabic Fake News Detection. Mathematics 2023, 11, 258. https://doi.org/10.3390/math11020258

AMA Style

Abd Elaziz M, Dahou A, Orabi DA, Alshathri S, Soliman EM, Ewees AA. A Hybrid Multitask Learning Framework with a Fire Hawk Optimizer for Arabic Fake News Detection. Mathematics. 2023; 11(2):258. https://doi.org/10.3390/math11020258

Chicago/Turabian Style

Abd Elaziz, Mohamed, Abdelghani Dahou, Dina Ahmed Orabi, Samah Alshathri, Eman M. Soliman, and Ahmed A. Ewees. 2023. "A Hybrid Multitask Learning Framework with a Fire Hawk Optimizer for Arabic Fake News Detection" Mathematics 11, no. 2: 258. https://doi.org/10.3390/math11020258

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop