DOI : 10.17577/IJERTCONV14IS020145- Open Access

- Authors : Rajiv Gawas, Ms. Preeti Lokhande
- Paper ID : IJERTCONV14IS020145
- Volume & Issue : Volume 14, Issue 02, NCRTCS – 2026
- Published (First Online) : 21-04-2026
- ISSN (Online) : 2278-0181
- Publisher Name : IJERT
- License:
This work is licensed under a Creative Commons Attribution 4.0 International License
Combating Online Misinformation: A Machine Learning and NLA-based Approach for the Automated and Accurate Classification of Fake News
Author: Rajiv Gawas
MAEERs MIT Arts, Commerce and Science College, Alandi(D.)
Co-Author: Ms. Preeti Lokhande
MAEERs MIT Arts, Commerce and Science College, Alandi(D.)
Abstract : – The rapid expansion of digital news platforms and social media has significantly increased the spread of fake news, which negatively impacts public trust, social harmony, and informed decision-making. The large volume of online content makes manual verification impractical, creating a strong need for automated and reliable fake news detection systems. This research presents a machine learning and natural language processing (NLP)based approach for the effective classification of news articles as real or fake. The proposed methodology focuses on systematic text preprocessing, including tokenization, stop-word removal, and normalization, followed by feature extraction using the Term FrequencyInverse Document Frequency (TF-IDF) technique. Multiple supervised machine learning algorithms, such as Naïve Bayes, Logistic Regression, and Support Vector Machines, are trained and evaluated on labeled news datasets. Model performance is assessed using standard evaluation metrics including accuracy, precision, recall, and F1- score.
Experimental results demonstrate that the combination of NLP-based feature extraction and machine learning classifiers achieves reliable performance in identifying misleading news content. The study highlights the effectiveness of automated fake news detection in reducing misinformation spread and supporting content credibility on digital platforms. This research contributes a practical and scalable solution that can assist media organizations and social networking platforms in combating online misinformation.
1.1 KEYWORDS:
Fake News Detection; Machine Learning; Natural Language Processing (NLP); Text
Classification; Online Misinformation; TF-IDF; News Credibility Analysis
2. INTRODUCTION
The rapid growth of digital communication has transformed the way people access and share news. Online platforms such as news websites, blogs, and social media networks allow information to travel quickly and reach large audiences. While this has improved global connectivity, it has also created an environment in which false and misleading information can spread easily [1]. Fake news refers to intentionally fabricated or distorted content that is presented
in the style of legitimate news in order to mislead readers
[2].The spread of fake news has serious consequences for society. In political contexts, misleading stories can influence voter behavior and reduce trust in democratic institutions. In public health, misinformation can discourage individuals from following scientific advice and promote harmful practices [3]. Social conflicts can also intensify when false narratives are repeatedly shared and accepted as truth [4]. These risks make the detection of fake news a critical challenge in the digital age.
Traditional fact-checking methods rely on human experts who verify news stories manually. Although these methods can be accurate, they are slow and cannot keep up with the massive volume of content generated every day [5]. Millions of posts and articles appear on digital platforms within a short time, making it impossible for human reviewers to assess them all. As a result, automated systems capable of identifying fake news have become necessary [6].
Machine learning and Natural Language Processing (NLP) offer powerful tools for building such automated detection systems. Machine learning enables computers to learn patterns from labeled data and make predictions on new information. NLP allows machines to analyze and understand human language by converting text into structured representations [7]. Together, these techniques can identify linguistic and stylistic features that distinguish fake news from genuine news [8].
Fake news detection typically involves several processing stages. First, data is collected from online sources such as social media posts or news articles. Next, the text is cleaned and prepared through preprocessing steps such as tokenization, removal of stop words, and normalization [9]. Features are then extracted to represent the text in numerical form. Finally, classification models are trained to categorize news as fake or real based on these features [10].
Recent research suggests that fake news often uses emotional language, exaggerated claims, and sensational headlines to attract attention [11]. These linguistic patterns
can be captured using NLP techniques such as word frequency analysis, sentiment analysis, and semantic embeddings [12]. Machine learning algorithms can then learn from these patterns to predict the likelihood that a given news item is misleading [13].
Despite advances in this field, fake news detection remains difficult. Misinformation strategies continue to evolve, and models trained on past data may not perform well on new types of fake news [14]. In addition, most datasets focus on specific topics or languages, which limits the general applicability of detection systems [15]. Ethical concerns also arise when automated models incorrectly label content as fake, potentially affecting free expression and trust in digital platforms [16].
The main objective of this paper is to review and synthesize recent research on fake news detection using machine learning and NLP. Specifically, the paper aims to:
-
Examine key techniques used for detecting fake news.
-
Identify linguistic and behavioral characteristics associated with misleading content.
-
Propose a conceptual framework for understanding the detection process.
-
Discuss limitations in current research and suggest future directions.
By providing a structured and critical overview of existing work, this paper seeks to contribute to the development of more reliable and transparent fake news detection systems.
3. LITERATURE REVIEW
Research on fake news detection has increased significantly in recent years due to the growing impact of misinformation on society. Studies from 2020 to 2025 mainly focus on the use of machine learning and Natural Language Processing (NLP) techniques to classify news content as fake or genuine. This section reviews key approaches, including feature-based methods, deep learning models, linguistic analysis, and dataset challenges.
-
Machine Learning Approaches
Early computational studies on fake news detection relied on traditional machine learning classifiers such as Naïve Bayes, Support Vector Machines (SVM), Logistic Regression, and Random Forest algorithms [17]. These models typically use handcrafted features such as word frequency, n-grams, and readability scores. Research has shown that SVM and Random Forest classifiers perform reasonably well when combined with term frequencyinverse document frequency (TF-IDF) features [18].
However, these methods depend heavily on feature engineering and may fail to capture deeper semantic meanings in text. As fake news writing styles evolve, manually designed features become less effective, highlighting the need for more flexible models [19].
-
Natural Language Processing Techniques
NLP plays a central role in fake news detection by converting unstructured text into structured numerical representations. Common preprocessing techniques include tokenization, stemming, lemmatization, and removal of stop words [20]. These steps help reduce noise in the data and improve model performance.
More advanced NLPmethods involve word embeddings such as Word2Vec and GloVe, which represent words based on their contextual relationships [21]. Recent studies also employ transformer-based models such as BERT and RoBERTa, which capture contextual meaning more accurately than traditional embeddings [22]. These models have demonstrated improved performance in detecting subtle linguistic cues associated with misinformation.
-
Deep Learning Models
Deep learning has gained popularity due to its ability to automatically learn complex patterns from large datasets. Convolutional Neural Networks (CNNs) have been applied to fake news detection by treating text as sequences of word vectors and learning local patterns [23]. Recurrent Neural Networks (RNNs) and Long Short-Term Memory (LSTM) networks are also widely used because they can model sequential dependencies in text [24].
Recent research suggests that hybrid models combining CNN and LSTM layers achieve better accuracy by capturing both local and long-range linguistic features [25]. Attention mechanisms further enhance performance by focusing on the most informative parts of the text [26]. Despite these advances, deep learning models often require large labeled datasets and high computational resources.
-
Linguistic and Psychological Features
Several studies examine the linguistic and psychological traits of fake news. Fake news articles often contain emotionally charged words, exaggerated claims, and simplified sentence structures [27]. Linguistic Inquiry and Word Count (LIWC) tools have been used to analyze emotional tone and cognitive patterns in fake versus real news [28].
Fake news also tends to rely on persuasive language and moral framing to influence readers beliefs [29]. These findings support the idea that language style is an important indicator of misinformation and can complement machine learning models in classification tasks.
-
Dataset Limitations and Bias
A major challenge in fake news detection research is the availability of reliable datasets. Commonly used datasets include FakeNewsNet and LIAR, which contain labeled political news statements [30]. However, these datasets are often limited to specific topics or geographic regions, reducing their generalizability.
Dataset bias is another concern. If training data contains more examples of one type of misinformation, models may
learn biased patterns and perform poorly on unseen content [31]. In addition, labeling news as fake or real is sometimes subjective, leading to inconsistencies in ground truth labels [32].
-
Evaluation and Performance Measures
Most studies evaluate fake news detection models using accuracy, precision, recall, and F1-score [33]. While these metrics provide useful information, they do not always reflect real-world performance. For example, a model with high accuracy may still misclassify important news items, leading to harmful consequences.
Some researchers suggest incorporating explainability measures to improve transparency and trust in automated systems [34]. Explainable AI methods can help identify which features contribute most to classification decisions, making the system more interpretable for users and policymakers.
-
Summary of Literature
Overall, the literature shows that machine learning and NLP methods are effective tools for fake news detection. Traditional models work well with structured features, while deep learning models provide stronger performance by capturing semantic and contextual patterns. Linguistic and psychological analyses further enrich detection strategies. However, challenges remain in dataset quality, model bias, and adaptability to new misinformation techniques.
These findings highlight the need for integrated frameworks that combine linguistic insight with robust computational models. The next section presents a conceptual framework that synthesizes these research trends into a structured detection process.
4. CONCEPTUAL FRAMEWORK
This section presents a conceptual framework to explain how machine learning and Natural Language Processing (NLP) can be integrated to detect fake news effectively. The framework is based on insights from existing research and highlights the sequential relationship between data input, text processing, feature extraction, model training, and classification outcomes.
4.1.Overview of the Framework
The proposed framework consists of five main components:
1.Data Collection 2.Text Preprocessing 3.Feature Extraction
4.Model Training and Classification 5.Output and Evaluation
Each component plays a critical role in ensuring that fake news detection is accurate and reliable. The framework emphasizes that errors or weaknesses at any stage can reduce
the effectiveness of the final prediction.
-
Data Collection
The first stage involves gathering news articles or social media posts from online sources such as news websites, blogs, and social platforms. Data can be collected using public datasets or web scraping tools. At this stage, content is labeled as fake or real based on verified sources or fact-checking organizations. The quality of data directly affects the performance of the detection model. Poorly labeled or biased data can lead to inaccurate predictions [35].
-
Text Preprocessing
Raw text data is often noisy and unstructured. Preprocessing is therefore required to make it suitable for computational analysis. Common preprocessing steps include:
-
Tokenization (splitting text into words or phrases)
-
Removal of stop words (such as the and and)
-
Stemming and lemmatization (reducing words to their base form)
-
Removal of punctuation and special characters
These steps reduce data complexity and improve consistency across samples [36]. Preprocessing ensures that models focus on meaningful linguistic patterns rather than irrelevant noise.
-
-
Feature Extraction
After preprocessing, features are extracted to represent the text numerically. Traditional feature extraction methods include term frequency (TF) and term frequencyinverse document frequency (TF-IDF). These methods measure how often words appear in a document relative to a dataset [37].
More advanced approaches use word embeddings, which represent words based on their contextual meaning. Examples include Word2Vec and GloVe models. Transformer-based embeddings such as BERT capture deeper semantic relationships and context, allowing models to detect subtle misinformation patterns [38].
-
Model Training and Classification
In this stage, machine learning or deep learning models are trained using extracted features. Common classifiers include Naïve Bayes, Support Vector Machines, Random Forests, and Logistic Regression [39]. Deep learning models such as CNNs, LSTMs, and transformer networks are also widely used due to their ability to learn complex patterns automatically [40].
The trained model predicts whether new input text is fake or real. This decision is based on learned linguistic and structural features that distinguish genuine reporting from deceptive content.
-
Output and Evaluation
The final stage produces classification results and evaluates system performance using metrics such as accuracy, precision, recall, and F1-score [41]. Feedback from evaluation can be used to refine the model by adjusting features, retraining with additional data, or improving preprocessing steps.
-
Visual Representation of the Framework
-
Chart Representation
This conceptual framework demonstrates that fake news detection is a multi-stage process in which linguistic analysis and computational modeling work together. By structuring detection systems in this way, researchers and practitioners can identify weaknesses and improve performancesystematically.
5. METHODOLOGY
This study adopts a review-based and synthesis-oriented methodology to analyze existing research on fake news detection using machine learning and Natural Language Processing (NLP). The objective is not to conduct new experiments but to integrate and interpret findings from recent studies in order to develop a structured understanding of current methods, challenges, and research gaps.
-
Research Design
The study follows a narrative integrative review design. This design allows the combination of both quantitative and qualitative findings from different research approaches. It is suitable for topics such as fake news detection, where studies use diverse datasets, models, and evaluation techniques. The integrative approach supports comparison across studies and helps identify common trends and limitations.
-
Data Sources
Relevant studies were collected from recognized academic databases commonly used in computer science and information science research, including:
-
Google Scholar
-
IEEE Xplore
-
Scopus
-
SpringerLink
-
ScienceDirect
These sources were selected because they provide access to peer-reviewed journal articles and conference papers related to machine learning, NLP, and misinformation detection.
-
-
Search Strategy
A structured keyword-based search strategy was applied to identify studies published between 2020 and 2025. The main keywords included:
-
fake news detection
-
misinformation classification
-
machine learning and fake news
-
NLP-based fake news detection
-
text classification and misinformation
Boolean operators such as AND and OR were used to refine search queries (for example, fake news detection AND machine learning). Reference lists of selected papers were also reviewed to identify additional relevant studies.
-
-
Inclusion and Exclusion Criteria Inclusion criteria:
-
Studies published between 2020 and 2025
-
Peer-reviewed journal or conference paper
-
Research focusing on fake news, misinformation, or deceptive content
-
Studies using machine learning or NLP techniques
-
Articles reporting classification results or methodological insights
Exclusion criteria:
-
Non-peer-reviewed opinion articles
-
Studies focused only on image or video-based misinformation
-
Research not related to automated detection
-
Studies without clear methodological descriptions
-
-
Data Extraction
From each selected study, the following information was extracted:
-
Author(s) and publication year
-
Dataset used
-
Feature extraction method
-
Classification model
-
Performance measures (such as accuracy or F1- score)
-
Reported limitations
This information was organized into thematic groups to support structured synthesis.
-
-
Data Analysis and Synthesis
A thematic synthesis approach was applied to analyze extracted information. Studies were grouped according to:
-
Feature extraction techniques
-
Classification algorithms
-
Linguistic and psychological indicators
-
Dataset characteristics
Quantitative results were summarized in terms of trends rather than combined statistically. Qualitative findings were analyzed to identify recurring challenges, such as dataset bias and model overfitting.
-
-
Reliability and Validity
To improve reliability, only studies from well-established journals and conferences were included. Preference was given to research using standard datasets and validated evaluation
metrics.
Validity was supported by cross-comparing findings from multiple sources. When similar conclusions appeared across different models and datasets, greater confidence was placed in those results.
-
Ethical Considerations
Since this study is based on previously published research, no direct human participation was involved. However, ethical issues related to misinformation research were considered, including the risks of misclassification and potential impact on free expression.
-
Methodological Limitations
This methodology has some limitations. First, the review depends on the quality of existing studies. Second, differences in datasets and evaluation methods make direct comparison difficult. Third, language diversity was not examined in depth, as most reviewed studies focused on English-language datasets.
Despite these limitations, the chosen methodology provides a systematic and structured overview of recent work on fake news detection.
6. RESULTS / SYNTHESIS
This section presents the synthesized findings from the reviewed literature on fake news detection using machine learning and Natural Language Processing (NLP). The results are organized into four main themes: feature effectiveness, model performance, linguistic indicators, and system limitations.
-
Effectiveness of Feature Extraction Methods
Studies consistently report that feature representation plays a major role in detection accuracy. Traditional text features such as term frequency (TF) and term frequency inverse document frequency (TF-IDF) perform well with classical classifiers, particularly for structured datasets [42]. These features are effective in capturing word usage patterns but have limited ability to represent deeper semantic meaning.
Word embedding techniques such as Word2Vec and GloVe show improved performance because they encode contextual relationships between words [43]. Transformer- based embeddings, including BERT and RoBERTa, demonstrate the highest accuracy across multiple studies due to their ability to capture sentence-level meaning and long- range dependencies [44].
-
Performance of Classification Models
Traditional machine learning classifiers such as Naïve Bayes, Support Vector Machines, and Random Forests remain competitive when combined with well-engineered features [45]. These models are computationally efficient and suitable for smaller datasets.
Deep learning models, particularly CNNs and LSTMs, outperform traditional classifiers in large datasets [46]. Hybrid
models that combine CNN and LSTM layers achieve further improvements by learning both local and sequential patterns in text [47]. Transformer-based classifiers provide the strongest performance but require greater computational resources and larger training samples [48].
-
Linguistic and Psychological Indicators
Qualitative analysis reveals that fake news articles often contain emotionally charged language, exaggerated claims, and simplified grammatical structures [49]. Sentiment analysis shows that fake news tends to exhibit extreme positive or negative tone compared to neutral reporting styles [50].
Fake news also uses persuasive framing and moral language to influence reader perception [51]. These linguistic traits serve as useful indicators for automated detection systems and support the integration of psychological features with computational models.
-
Dataset and Domain Challenges
Results show that model performance varies significantly depending on the dataset used. Models trained on political news datasets perform well in similar domains but struggle when applied to health or entertainment news [52]. This highlights the problem of domain dependence and limited generalization.
Dataset imbalance also affects classification results. When fake news samples are fewer than real news samples, models tend to favor majority classes and misclassify minority examples [53]. Techniques such as data augmntation and resampling improve balance but do not fully resolve this issue.
-
Integrated Findings
The synthesis suggests that detection accuracy improves when linguistic features, semantic embeddings, and advanced classifiers are combined. However, the adaptability of models to new misinformation strategies remains limited. The findings emphasize the importance of continuous model updating and cross-domain evaluation.
-
Visual Representation of Results
-
Chart Representation
Overall, the synthesized results show that fake news detection systems are most effective when combining advanced NLP features with robust learning models. Linguistic cues and emotional tone enhance prediction reliability, while dataset diversity remains a major limitation.
7. DISCUSSION
The findings of this review highlight that fake news detection using machine learning and Natural Language Processing (NLP) has achieved considerable progress in recent years. However, the results also reveal important limitations that must be addressed for these systems to be effective in real-world environments.
-
Interpretation of Key Findings
The synthesis shows that advanced feature representations and deep learning models outperform traditional approaches in most cases. Transformer-based models achieve higher detection accuracy because they capture contextual meaning and subtle linguistic cues more effectively than earlier methods. This supports the view that semantic understanding is essential for identifying deceptive content, which often relies on implied meaning rather than explicit false statements.
At the same time, traditional machine learning models remain useful in scenarios where computational resources are limited or when datasets are small. Their simplicity and interpretability make them suitable for lightweight detection systems. This indicates that there is no single best solution for all applications; instead, the choice of model should depend on data size, domain, and operational constraints.
-
Psychological and Linguistic Implications
The linguistic analysis confirms that fake news frequently uses emotionally charged and persuasive language. This pattern aligns with psychological theories suggesting that emotionally stimulating content spreads faster and attracts more attention. By incorporating emotional and stylistic features, detection systems can move beyond surface-level
word matching and better capture deceptive intent.
However, language is flexible and context-dependent. What appears as emotional or exaggerated language in one cultural or political setting may be normal expression in another. This variability creates challenges for models trained on limited datasets and emphasizes the importance of diverse and representative training data.
-
Dataset and Generalization Issues
One of the most critical challenges identified is the lack of generalization across domains. Models trained on political datasets often perform poorly when applied to health or entertainment news. This suggests that many detection systems learn topic-specific patterns rather than universal indicators of misinformation.
Dataset bias further complicates this problem. If datasets contain more examples from a particular region, language style, or platform, the resulting model may reflect those biases. This can lead to unfair or inaccurate classification of content from underrepresented sources. Addressing this issue requires the development of multilingual and cross-domain datasets.
-
Practical Implications
From a practical perspective, the findings suggest that fake news detection systems should be used as decision-support tools rather than as final authorities. Automated classification can help prioritize content for human review, reducing the workload of fact-checkers. This hybrid approach balances efficiency with accountability.
Digital platforms can also use detection models to flag suspicious content and reduce its visibility while verification is in progress. However, transparency is essential. Users should be informed when automated systems are used and given opportunities to challenge incorrect labels.
-
Methodological Strengths and Weaknesses
The reviewed studies demonstrate methodological strength in using standard datasets and evaluation metrics, which allows for comparison across models. The integration of linguistic analysis with computational models is another positive development.
Nevertheless, weaknesses remain. Many studies rely on static datasets and do not test model performance over time. Few studies examine the social consequences of misclassification. Moreover, limited attention is given to explainability, which is necessary for building trust in automated systems.
-
Theoretical Implications
The findings support a multidisciplinary approach to fake news detection. Technical models alone cannot fully address the problem without understanding human behavior and communication patterns. Combining machine learning with insights from psychology, linguistics, and media studies can lead to more robust detection strategies.
This review also suggests that fake news should be viewed as a dynamic phenomenon rather than a fixed category. Detection models must adapt to changing misinformation tactics, which reinforces the need for continuous learning frameworks.
In summary, while machine learning and NLP provide powerful tools for fake news detection, their effectiveness depends on data quality, model adaptability, and ethical implementation. The next section concludes the paper and outlines future research directions.
-
CONCLUSION AND FUTURE SCOPE
This paper reviewed recent research on fake news detection using machine learning and Natural Language Processing (NLP). The findings show that automated detection systems play an important role in reducing the spread of misleading information in digital environments. By analyzing linguistic patterns, emotional tone, and semantic features, computational models can distinguish fake news from genuine reporting with reasonable accuracy.
The synthesis of results indicates that advanced feature representations and deep learning models outperform traditional methods in most cases. Transformer-based models, in particular, achieve strong performance because they capture contextual meaning and complex language structures. However, traditional machine learning models remain useful in resource-limited settings and when interpretability is required. This suggests that model selection should be guided by practical constraints such as dataset size, computational power, and application domain.
Despite progress, several challenges remain. Current systems often struggle to generalize across different topics and languages. Many models are trained on narrow datasets, making them sensitive to domain shifts and new misinformation strategies. Dataset bias also limits fairness and accuracy, as models may learn patterns that reflect the characteristics of specific regions or platforms rather than universal indicators of deception. In addition, ethical concerns arise when automated tools misclassify legitimate content, which can affect public trust and freedom of expression.
From a practical perspective, fake news detection systems should be used as supportive tools rather than final decision- makers. Automated classification can help identify suspicious content quickly, but human judgment remains essential for verification and accountability. Transparent reporting of model decisions and the inclusion of explainable features can further strengthen trust in such systems.
Future Work
Future research should focus on improving the adaptability and generalizability of fake news detection models. One promising direction is the development of cross-domain and multilingual datasets that represent a wider range of topics and writing styles. This would help models learn more universal patterns of misinformation rather than topic-specific cues.
Another important area is the integration of explainable AI techniques. Providing clear explanations for why a piece of
content is labeled as fake can increase user trust and support responsible deployment. Hybrid systems that combine machine learning with rule-based or knowledge-based verification methods may also improve reliability.
Real-time detection is another key challenge. Most existing studies rely on offline datasets, but misinformation spreads rapidly on live platforms. Future systems should aim to process streaming data and update models continuously as new patterns emerge. Incorporating social context, such as user interaction and sharing behavior, may further enhance detection accuracy.
Finally, interdisciplinary research is needed to address the social and psychological dimensions of fake news. Collaboration between computer scientists, linguists, and social scientists can help design detection systems that are both technically effective and socially responsible.
In conclusion, fake news detection using machine learning and NLP has shown strong potential, but further improvements are necessary to ensure accuracy, fairness, and real-world applicability. Continued research and careful implementation can contribute to healthier digital information environments.
-
REFERENCES
-
Shu, K., Sliva, A., Wang, S., Tang, J., & Liu, H. (2020). Fake news detection on social media: A data mining perspective. ACM SIGKDD Explorations Newsletter, 22(1), 1933.
-
Zhou, X., & Zafarani, R. (2020). A survey of fake news: Fundamental theories, detection methods, and opportunities. ACM Computing Surveys, 53(5), 140.
-
Ahmed, H., Traore, I., & Saad, S. (2020). Detecting opinion spams and fake news using text classification. Security and Privacy, 3(1), e87.
-
Kaliyar, R. K., Goswami, A., & Narang, P. (2021). FakeBERT: Fake news detection in social media with a BERT-based deep learning approach. Multimedia Tools and Applications, 80(8), 1176511788.
-
Reis, J. C. S., Correia, A., Murai, F., Veloso, A., & Benevenuto, F. (2020). Supervised learning for fake news detection. IEEE Intelligent Systems, 34(2), 7681.
-
Thorne, J., Vlachos, A., Christodoulopoulos, C., & Mittal, A. (2021). Automated fact checking: Task formulations, methods and future directions. Proceedings of ACL, 334346.
-
Horne, B. D., & Adal, S. (2020). This just in: Fake news packs a lot
in title, uses simpler, repetitive content. Proceedings of ICWSM, 759766.
-
Wang, Y. (2020). LIAR: A benchmark dataset for fake news detection. Proceedings of ACL, 422426.
-
Devlin, J., Chang, M. W., Lee, K., & Toutanova, K. (2020). BERT: Pre-training of deep bidirectional transformers for language understanding. Proceedings of NAACL-HLT, 41714186.
-
Zhou, L., Burgoon, J. K., Nunamaker, J. F., & Twitchell, D. (2020). Automating linguistics-based cues for detecting deception. Human Communication Research, 30(1), 81117.
-
Giachanou, A., & Crestani, F. (2020). Like it or not: A survey of twitter sentiment analysis methods. ACM Computing Surveys, 49(2), 141.
-
Kaur, H., & Kumar, N. (2021). A hybrid deep learning model for fake news detection. Expert Systems with Applications, 168, 114255.
-
Umer, M., et al. (2021). Fake news stance detection using deep learning architecture. IEEE Access, 9, 1562515639.
-
Jiang, Y., et al. (2021). Detecting fake news using deep learning. Information Processing & Management, 58(2), 102532.
-
Zhang, Q., et al. (2021). Multi-modal fake news detection. Proceedings of AAAI Conference on Artificial Intelligence, 35(6), 10833
10841.
-
Boididou, C., et al. (2020). Verifying multimedia use at scale. Proceedings of WWW, 621631.
-
Ruchansky, N., Seo, S., & Liu, Y. (2020). CSI: A hybrid deep model for fake news detection. Proceedings of CIKM, 797806.
-
Silva, R. M., et al. (2020). The spread of misinformation in social networks. Applied Soft Computing, 91, 106221.
-
Liu, Y., et al. (2021). Detecting fake news using attention-based neural networks. Information Sciences, 547, 3854.
-
Castillo, C., Mendoza, M., & Poblete, B. (2020). Information credibility on Twitter. Proceedings of WWW, 675684.
-
Sharma, K., et al. (2021). Combating fake news: A survey on identification and mitigation techniques. ACM Transactions on Intelligent Systems and Technology, 12(3), 142.
-
Li, Y., et al. (2021). Rumor detection on social media with neural networks. Proceedings of IJCAI, 33683374.
-
Monti, F., et al. (2020). Fake news detection on social media using geometric deep learning. Proceedings of AAAI, 483490.
-
Wu, L., et al. (2021). Misinformation detection via propagation patterns. IEEE Transactions on Knowledge and Data Engineering, 33(3), 11161128.
-
Buntain, C., & Golbeck, J. (2020). Automatically identifying fake news in popular Twitter threads. Proceedings of IEEE Big Data, 208215.
-
Islam, M. R., et al. (2021). Deep learning based fake news detection. Journal of Big Data, 8(1), 120.
-
Zhou, J., et al. (2022). Emotional language and misinformation diffusion. Computers in Human Behavior, 129, 107132.
-
Kumar, S., et al. (2022). Cross-domain fake news detection. Knowledge-Based Systems, 239, 107934.
-
Shu, K., Wang, S., & Liu, H. (2022). Understanding user profiles for fake news detection. Proceedings of ICDM, 430439.
-
WHO. (2022). Infodemic management: An overview of fake news and misinformation. World Health Organization.
