1. Introduction
In recent years, the proliferation of fake news has become a significant concern in the digital age. With the rise of social media platforms and the increasing ease of information dissemination, the spread of misinformation and disinformation poses serious threats to public opinion, democratic processes, and societal harmony. Fake news and fake profiles, often used in conjunction with sophisticated astroturfing techniques, manipulate public perception and can lead to widespread misinformation [
1].
Several deep learning models have been recommended in the recent past to answer the challenges of detecting fake news and their dissemination patterns, such as Long-short term memory (LSTM) and Convolutional Neural Networks (CNN) for fake news content [
2,
3] or fake profile [
4] detection, as well as Graph Convolution Networks (GCNs) for the analysis of fake news diffusion paths [
5].
The advent of Large Language Models (LLMs) and generative Artificial Intelligence (AI) has added a new dimension to this challenge. These technologies possess the dual capability to both generate and detect fake news, creating a paradox in their application. On one hand, LLMs can be harnessed to create convincingly realistic but false content, making the identification of such fake news increasingly difficult [
6]. On the other hand, these same models offer advanced methods for detecting and combating the spread of fake news [
7], presenting a potential solution to the problem they help create.
The importance of detecting fake news cannot be overstated. Reliable detection mechanisms are crucial for maintaining the integrity of information ecosystems, ensuring that the public can trust the information they consume. Effective detection not only helps in identifying false information but also in curbing its spread, thereby mitigating its impact.
This survey paper aims to explore the multifaceted role of LLMs and generative AI in the realm of fake news. It delves into the background of fake news, fake profiles, and the mechanisms of fake news dissemination and astroturfing. Furthermore, it examines the dual role of LLMs as both generators and detectors of fake news, highlighting the importance of detection technologies. The paper also sets forth the objectives of this survey, aiming to provide a comprehensive overview of the current landscape and emerging trends in this field.
The structure of this survey is organized as follows: First, in
Section 2 we provide a detailed background on fake news, fake profiles, and the state-of-the-art approaches used before LLMs for their detection.
Section 3 describes the methodology used in this study for searching and selecting the most relevant articles in the literature.
Section 4 provides more details on the different ways that LLMs can be used to generate and detect fake content and profiles in social networks.
Section 5 presents the different datasets, metrics, and techniques used to assess the performance of LLMs in fake news detection, and summarizes the main performance results in this task.
Section 6 presents case studies and practical applications of the use of LLM in the fake news domain.
Section 7 discusses the main challenges and limitations and
Section 8 concludes the paper by providing references to future directions in the field.
2. Overview of Fake News and Fake Profiles
2.1. Disinformation/Fake News, Spreading
Fake news and disinformation have become pervasive issues in the digital age, exacerbated by the rise of social media and advanced AI technologies. Understanding the origins, methods, and impacts of fake news is crucial for developing effective detection and mitigation strategies. In the following we explore the evolution of fake news creation and detection, highlighting the role of both human and AI-driven approaches.
Fake news is defined as false news deliberately spread to deceive people, as stated by the Oregon Library [
8]. In the pre-LLM era, the creation of fake news was purely the result of human manipulation [
9]. Prior to the appearance of LLMs, research in AI-generated misinformation detection focused mainly on Smaller Language Models (SLMs) [
10], such as BERT [
11], GPT-2 [
12], and T5 [
13].
Disinformation, which implies an intent to mislead, refers to false information disseminated tactically to bias and manipulate facts, often associated with the term âpropagandaâ. The prefix dis- indicates a reversal or negative instance of information [
14].
With the rise of LLMs, fake news can now be both automatically generated and detected [
9]. The complexity of disinformation detection has dramatically escalated with the advent of LLMs, which possess billion-scale parameters [
10]. In recent years, the popularization of the internet and social networks has facilitated the rapid spread of fake news, causing severe personal, social, and economic damage [
15]. Fake news poses significant risks in various areas [
6], like public health and healthcare, politics, and e-commerce. The threat of the automatic generation of fake news through large language models (LLMs) is considered one of the most significant risks to the further development of LLMs [
16]. Therefore, it is urgent to update detection systems with knowledge of LLM-generated fake news [
17].
Fake profiles in Online Social Networks (OSNs) are another significant concern. These profiles are used in Advanced Persistent Threat cases to spread malware or links to it. They are also used in other malicious activities like junk emails, spam, or artificially increasing the number of users in some applications to promote them [
18]. Additionally, fake accounts are created to boost the number of fans for celebrities or politicians on social media.
The widespread use of social networking has led to various problems, including the exposure of users to incorrect information through fake accounts, which results in the spread of malicious content [
19].
Table 1 presents a comparison of disinformation and/or fake news detection studies.
2.2. Detection Methods
Internet and social media made access to news information much easier and more comfortable. On social media platforms, every user from around the world can publish and disseminate any kind of statement, or set of statements, to spread fake news to achieve different goals, which may be fraudulent or illegal [
3]. Aside from individuals spreading fake news, some sources are considered to be authentic and are popular sources for informational services, such as Wikipedia, which are also prone to false information or fake news [
21]. Mass media may also manipulate the information in different ways to achieve their goals. As the sources of spreading fake news are too many, the detection of fake news is essential. One important detection mechanism is the fact-checking sites, but unfortunately, they are insufficient to combat the amount of disinformation disseminated [
22]. Fake news is increasingly sophisticated, adopting complex characteristics that differ between different types of social networks. Therefore, it is essential to develop detection mechanisms using artificial intelligence or machine learning algorithms. The paragraphs that follow present the two main fake news detection and mitigation types: (i) Rule-based Approaches and (ii) Human Fact-Checking and Moderation and the key solutions of each type.
2.2.1. Rule-Based Approaches
Rule-based approaches for fake news detection generally fall into two main methods: Keyword Analysis and Heuristic Rules.
The
Keyword Analysis method explores the Positive and Unlabeled Learning (PUL) combined with a network-based Label Propagation (PU-LP) algorithm for detecting fake news [
23]. This combination reduces labeling costs while improving the classification task by exploiting relationships between news and terms using a few labeled fake news examples. The PU-LP algorithm is integrated with a Graph Attention Neural Event Embedding (GNEE) to further refine the relevance of terms used in fake news detection. Experiments show that such approaches, especially when using specialized keyword extraction tools, achieve high accuracy with minimal labeled data. The potential incorporation of attention mechanisms is also expected to improve effectiveness in fake news detection.
Other studies in this group [
3,
24] introduce hybrid deep learning models that combine Convolutional Neural Networks (CNN) and Recurrent Neural Networks (RNN) for fake news classification. The models leverage CNN’s ability to extract local features and the Long Short-Term Memory (LSTM) network’s capability to learn long-term dependencies. By processing input data to capture both local and sequential features, the models classify text as fake or not. Validation on popular fake news datasets (ISO and FA-KES) shows that hybrid models significantly outperformed other non-hybrid baseline methods.
Furthermore, the Interpretable Knowledge-Aware (IKA) model uses historical evidence in the form of graphs for fake news detection [
25]. It constructs positive and negative evidence graphs by collecting signals from historical news. IKA compares the given piece to be detected with these evidence graphs to generate a matching vector, enhancing prediction accuracy. Semantic, stylistic, and emotional elements are extracted from news articles and comments and combined with graph-based data for prediction. The model was tested on English and Chinese datasets, achieving high accuracy. IKA is versatile and can be integrated with other existing models to improve their interpretability.
The
Heuristic Rules approach utilizes artificial intelligence algorithms like the Naive Bayes classifier to detect fake news on social media [
26]. Each user post represents a news item, with each word processed independently to calculate the probability of the post being fake. The classification process evaluates the likelihood that the post is fraudulent based on the presence of certain words. To ensure numerical stability, the system uses default probabilities for unknown words and logarithmic computations. This method achieves a classification accuracy of approximately 74%, which is acceptable for the simplicity of the approach.
More advanced deep learning techniques exploit the power of Graph Convolution Neural Networks (GCNs) to analyze the text of user posts and/or user profiles and detect fake news or fake profiles [
5]. They also take advantage of the news diffusion paths to further detect hidden attempts to disseminate fake content, by engaging fake and legitimate user accounts [
27].
2.2.2. Human Fact-Checking and Moderation
Human fact-checking and moderation approaches play a crucial role in the fight against fake news by leveraging the expertise of individuals and organizations to verify information. These approaches can be broadly categorized into Manual Verification Processes and Fact-Checking Organizations methods.
Manual Verification Processes involve human experts who assess the credibility of information through thorough investigation and cross-referencing with reliable sources. One of the seminal works on the automatic detection of fake news was conducted by Vlachos and Riedel [
28]. The authors defined the task of fact-checking, collected a dataset from two popular fact-checking websites, and used k-Nearest Neighbors classifiers to handle fact-checking as a classification task.
Fact-Checking Organizations are dedicated entities that specialize in verifying the validity of information across various domains. Several well-known organizations include:
PolitiFact [
29]: A well-known non-profit fact-checking website in the United States [
30]. It mainly rates the accuracy of claims made by US political news and politicians’ statements using classifications such as true, mostly true, half true, mostly false, false, and “pants on fire” (i.e., clearly fake news) [
31].
Snopes.com [
32]: One of the first online fact-checking sites, launched in 1994 [
31]. The site addresses political, social, and other current issues, primarily confirming news stories spread on social media.
Suggest (also known as GossipCop) [
33]: This site investigates fake stories about US celebrities and entertainment published in magazines and web portals. Each story is rated on an authenticity scale from 0 to 10, where 0 denotes a completely false rumor and 10 indicates a report that is entirely true.
FactCheck.org [
34]: A non-profit site that describes itself as a âconsumer advocate for votersâ aiming to reduce the level of deception and confusion in US politics [
31]. It primarily assesses assertions made by politicians in speeches, debates, and TV commercials.
FullFact [
35]: An online fact-checking site supported by a British charity that verifies and corrects news and assertions circulating on UK social media.
2.3. Fake Profile Detection
Online social networks such as Facebook, X (former Twitter), RenRen, LinkedIn, and Google+ have become increasingly popular over the last few years [
36]. OSNs dramatically impact people’s lives [
37], as they are used to interact with others, exchange information, organize events, engage in e-commerce, and even run their own e-businesses. However, one of the ways people misuse social media is by creating fake profiles for various unethical purposes, such as targeting specific users, spreading misinformation, or engaging in other cybercrimes [
38]. Therefore, the identification and elimination of fake accounts on OSNs are essential for maintaining the platforms’ integrity and security. In the following, we present the three main methods in this task: Profile picture verification, Network analysis, and Behavior and interaction patterns.
2.3.1. Profile Attribute Analysis
Profile attribute analysis is crucial for distinguishing between genuine and fake social media accounts. This analysis employs two main approaches: Profile Picture Verification and Consistency of Profile Information. Each method involves various ML techniques and algorithms to ensure the authenticity of profiles, thereby enhancing the security and reliability of social networking platforms.
The
Profile Picture Verification approach leverages several methods to verify the authenticity of profile pictures. One such method is the Photo-Response Non-Uniformity (PRNU) technique, which can be successfully used as a watermarking approach for authorship attribution and verification of pictures on social networks. According to [
39], PRNU employs a unique noise pattern for each camera that withstands the compression algorithms applied by various social platforms, unlike traditional watermarking techniques. This method enables accurate detection of fake accounts without altering the original image’s appearance.
Additionally, a new algorithm named SVM-NN is proposed by [
36] to provide efficient detection for fake Twitter accounts and bots. This machine learning classification algorithm utilizes fewer features than traditional support vector machine (SVM) or neural network (NN) algorithms. The SVM-NN model involves training an NN model using SVM decision values, effectively distinguishing between Sybil and real accounts.
Moreover, a hybrid model described in [
40] uses machine learning and skin detection algorithms to identify fake accounts. This model integrates five supervised machine learning algorithms—K-nearest Neighbor, Support Vector Machine, Naïve Bayes, Decision Tree, and Random Forest—to analyze user account features. It achieves up to 80% accuracy in detecting fake accounts by calculating the percentage of skin exposure in profile images, finding that fake accounts often exhibit higher skin percentages.
The
Consistency of Profile Information approach focuses on the textual and metadata consistency of profile information. One significant method is the detection of fake LinkedIn profiles immediately after creation, as described by [
41]. Utilizing textual data provided during the registration process, the system employs section and subsection tags (SSTE) to detect fake accounts before they establish connections with real users.
Furthermore, a data mining approach for fake profile identification, as detailed by [
37], includes four well-known techniques: Neural Network (NN), Support Vector Machine (SVM), Principal Component Analysis (PCA), and Weighted Average (WA). PCA is used for dimensionality reduction, selecting the most explanatory profile features. NN, SVM, and WA predict the authenticity of profiles, with SVM using a Polynomial Kernel achieving the highest accuracy at 87.34%.
Finally, a proposed framework combines the Random Forest Classifier model with pipelining to identify fake profiles across social communities, as described in [
42]. The classifier, optimized through grid search, evaluates features like profile picture, bio length, external URL, privacy status, number of posts, followers, and follows. The pipelining process enhances model accuracy by automating the workflow and evaluating the model’s performance using a confusion matrix for precise and reliable results.
2.3.2. Network Analysis
In network analysis, various methods are utilized to detect and analyze fake accounts on social media platforms. These methods focus on examining relationships and communities within the network, leveraging graph-based algorithms and community detection techniques to ensure the authenticity of user profiles.
Relationship Graphs is a fundamental aspect of detecting fake users. An enhanced graph-based semi-supervised learning algorithm (EGSLA) has been introduced to detect fake users on Twitter by examining user activity over a prolonged period [
43]. EGSLA employs dynamic feature changes, which are challenging for spammers to manipulate over time since spammers typically mimic real users temporarily due to the associated costs. By representing user activity through a graph model, EGSLA classifies, predicts, and makes recommendations using scalable streaming for real-time data processing. The algorithm’s performance, tested against existing game theory, k-nearest neighbor (KNN), support vector machine (SVM), and decision tree techniques, achieves a 90.3% accuracy in identifying fake users, outperforming other methods and underscoring its effectiveness.
Another approach proposes a hybrid method for detecting malicious profiles on Twitter by utilizing machine learning techniques. Initially, a Petri net structure analyzes user profiles and features [
44], which are subsequently used to train various classifiers to distinguish between malicious and genuine users. The framework processes baseline data through different classifiers in machine learning environments, creating separate datasets for authentic and fraudulent accounts. By applying reachability checks, the system identifies a harmful activity, involving interaction anomaly detection and addressing low accuracy issues through a trackback mechanism. A comparative analysis demonstrates the framework’s efficiency, showing improved performance over state-of-the-art techniques.
Furthermore, a real-time Sybil detector for the Renren online social network has been developed, operating based on a graph where nodes represent user profiles and edges denote connections between profiles through interactions and relationships [
45], taking into account the interaction anomalies that may exist. The detector monitors account features such as friend-request frequency, outgoing request acceptance rates, and clustering coefficients to identify suspicious behavior. By analyzing the graph structure, particularly patterns of link creation and clustering coefficients, it distinguishes between genuine and Sybil accounts. If an account exceeds predefined limits, it is flagged as a potential Sybil by a threshold-based system. An adaptive feedback scheme dynamically tunes the threshold parameters, leading to the identification and banning of over 100,000 Sybil accounts.
Community Detection is another critical method for identifying fake accounts. The Communal Influence Propagation Framework (CIPF) identifies fake accounts by analyzing essential features from individual user profiles, linguistic characteristics, and group profiles (network features) [
46]. The framework scrutinizes these features to generate a feature vector of the user space, utilizing Social Media Mining (SMM) theories such as Influence, Homophily, and Balance to enrich the malicious user space as an influential index. The Jaccard coefficient evaluates similarity indices, detects Sockpuppet nodes, and generates a negative propagation belonging matrix. CIPF integrates two influence-based verification models—the Sockpuppet Detection Phase (IB-SPD) and the Convolutional Neural Network-based Crowdturfing Community (IB-CFC)—to robustly identify and verify fake accounts.
Additionally, a majority voting approach leverages machine learning algorithms to identify and categorize counterfeit social media profiles [
47]. This approach employs various algorithms, including Decision Trees, XG-Boost, Random Forest, Extra Trees, Logistic Regression, Ada-Boost, and K-Nearest Neighbors, each capturing distinct aspects of user behavior and profile attributes. By combining several techniques, a group of classifiers is formed, and a majority vote process determines the legitimacy of a social media profile. This methodology effectively distinguishes authentic user profiles from fake ones by utilizing the decision-making capabilities of multiple classifiers.
Lastly, the Grey Wolf Optimization Algorithm (GWO), inspired by the hunting behavior of grey wolves, finds optimal solutions for complex problems by balancing exploration and exploitation [
48]. In this context, each online social network (OSN) user is classified as either “Real” or “Fake”. Compared to conventional machine learning algorithms such as K-Nearest Neighbor, Support Vector Model, Decision Tree, and Random Forest, GWO outperforms them in a Facebook dataset, achieving the highest accuracy (0.98), precision (0.96), recall (0.97), and F1-score (0.96). This demonstrates its efficacy in optimizing classification tasks and identifying fake profiles.
2.3.3. Behavior and Interaction Patterns
In the realm of behavior and interaction patterns, various methods are utilized to detect and analyze fake accounts on social media platforms. These methods focus on examining the frequency of posts, the content of tweets, and the interaction patterns of users, employing machine learning algorithms to identify anomalies and ensure the authenticity of user profiles.
Posting Frequency/
Tweets are critical indicators in detecting fake accounts. A study presents a classification method for detecting fake accounts on Twitter using the Naïve Bayes algorithm [
19]. The data include features provided by the Twitter API, such as description, profile image, number of followers, users, and public lists in which the user participates. Additionally, they consider the number of hashtags, mentions, and URL links used in the last 20 tweets, as well as the number of tweets liked by the user. The dataset was preprocessed using a supervised discretization technique named Entropy Minimization Discretization (EMD) on numerical features. In the Naïve Bayes learning algorithm, preprocessing the data with EMD without any feature selection increases the success rate, as Naïve Bayes performs better with discrete values than continuous values. The success rate of the Naïve Bayes algorithm is 90.9% after discretization.
In another study, Facebook user feeds were collected using the Facebook Graph API, which was fine-grained based on privacy settings [
49]. The most commonly used supervised machine learning classification techniques were applied to a basic dataset comprising the user’s node, their friends in their social neighborhood, and a set of manually identified spam accounts. The performance of the classifiers was evaluated to determine their ability to detect fake accounts on Facebook. Results show that the best-performing classifiers achieved a high classification accuracy of 79%. Furthermore, the study investigated the feature set to determine the attributes that most successfully detect fake profiles on Facebook. It was found that user activities related to likes, comments, and, to some extent, shares on Facebook contribute the most to detecting fake accounts. Therefore, the study represents a significant step towards a profile-feature-based detection of fake accounts on Facebook.
3. Methodology
To determine the related scope of the current work, we utilized the PRISMA 2020 guidelines [
50]. These guidelines outline the systematic approach adopted for reviewing literature on the usage of LLMs for generating fake news, and fake profiles, and also the methods of detection based on LLMs.
Search Strategy: The initial search was conducted across multiple academic sources, including Google Scholar and IEEE Xplore, but mainly focused on Scopus due to the better quality of related results and improved control during the filtering process. In Scopus, we used a set of search parameters that look for specific keywords in the titles, abstracts and keywords, and limit the search range to works published from 2018 onwards. We targeted published works in journals, conferences, and review papers. The query used for Scopus is presented below.
Search Criteria: To gather relevant works, we extended our search to include Google Scholar and IEEE Xplore. In addition to the initial keywords, we used composite phrases such as “LLM in fake news generation”, “fake profile detection using LLM”, and “security challenges in detecting fake news with LLM”. We conducted several queries and manually validated the results to find more relevant papers, including pre-prints. By screening abstracts, assessing the credibility of sources, and prioritizing peer-reviewed articles, we selected works that address the various aspects of using LLMs for generating and detecting fake news and profiles.
We also employed a snowballing technique, which involves examining the references of initially identified papers to discover additional relevant works. This method allowed us to uncover papers that might not have surfaced during our initial search. By exploring citations within these papers, we could track the development and interconnections of research topics over time. We also utilized forward snowballing, looking at newer papers that cited our core set of articles. Iterative refinement of our search strategies further helped us in this process. The methodology we followed in selecting papers for inclusion in the survey is depicted in
Figure 1. The results of our literature review are detailed in the following sections.
Inclusion and Exclusion Criteria: The search was restricted to English-language articles. We limited the search to works published in the years after 2018 and further filtered the papers using additional criteria. The inclusion criteria were as follows: peer-reviewed articles focusing on using large language models (LLMs) to generate fake news, detect fake profiles, and address security challenges associated with these activities. We also considered papers that discuss various techniques for improving the effectiveness and accuracy of LLMs in these contexts. Exclusion criteria included non-peer-reviewed articles and publications not specific to LLMs or their application in fake news and profile detection. In total, 117 publications met the inclusion criteria. The distribution of these publications over the years is presented in
Figure 2, while
Figure 3 shows the types of the accepted publications.
Quality Assessment: The quality of the selected papers was meticulously evaluated based on several criteria: their relevance to the research topics, methodological rigor, the impact factor of the journals in which they were published, and their citation counts. This comprehensive evaluation ensured that only the most credible and significant studies were included in our survey, thereby enhancing the reliability and depth of our findings.
Data Extraction and Synthesis: Data extraction and synthesis focused on identifying categories and methods related to the use of LLMs for generating fake news, creating fake profiles, and detecting these activities.
Figure 4 depicts the most popular topics mentioned in the selected articles. We specifically considered survey works that illustrate different approaches to using LLMs, studies discussing aspects such as detection accuracy, robustness of fake news generation, and user identification techniques. Given the recent broad access to LLM technologies, the number of surveys and studies available was limited. This limitation was reflected in the literature, which mostly consists of recent publications. The limited literature on LLMs and their application in fake news and profile detection underscores the emerging nature of this field and the need for further research and exploration.
Results of the Literature Search: The findings are organized in alignment with the document’s structure, providing a detailed analysis of the use of large language models (LLMs) for generating fake news, creating fake profiles, and detecting such activities. This includes examining the existing literature that highlights the importance and various dimensions of LLMs in the context of fake news and profile detection. Given the limited literature due to the recent broad access to LLM technologies, our survey covers the most relevant and recent studies in this emerging field. We analyze the different methodologies employed, the effectiveness of various detection techniques, and the security and privacy considerations associated with using LLMs for these purposes.
4. Methods That Use Large Language Models
LLMs are sophisticated AI models designed to understand and generate human language. They have evolved significantly, enabling a wide range of applications, including the potential for malicious uses such as generating fake news and fake profiles, as well as spreading misinformation. However, these unique abilities also provide powerful tools for detecting such misuse. In
Section 4.1, we explore some of the most prominent LLMs and their contributions to the field.
The advancements in LLMs have led to their application in both the generation and detection of fake news and fake profiles. The following sections will delve into these specific applications, discussing how LLMs are used to generate fake news and profiles, and how they can be leveraged to detect such deceptive content effectively. These advancements will be explored in the
Section 4.2,
Section 4.3 and
Section 4.4.
4.1. Introduction to Large Language Models (LLMs)
During the last few years, an abundance of pre-trained models has been developed and used in various linguistic tasks, boosting the performance of traditional methods, and using only fine-tuning in most cases. In this section, we present the details of these models, whereas
Table 2 provides a summary of the most frequently used models.
The evolution of large language models began with BERT (Bidirectional Encoder Representations from Transformers), introduced by Google researchers in 2018. BERT is a language representation model that learns language in a bidirectional manner. It has two variants: BERT-BASE (12 layers, 110 M parameters) and BERT-LARGE (24 layers, 340M parameters). BERT was trained on the BooksCorpus (800 million words) and English Wikipedia (2.5 billion words), excluding lists, tables, and headers [
11]. Building on BERT, M-BERT (Multilingual BERT) was released in 2019 as a single language model pre-trained on Wikipedia pages of 104 languages. It has a 12-layer transformer architecture and comes in several variants: BERT-Base Multilingual Cased (104 languages, 110 M parameters), BERT-Base Multilingual Uncased (102 languages, 110M parameters), and BERT-Base Chinese (110 M parameters) [
51].
Another significant model is T5 (Text-to-Text Transfer Transformer), released in 2019, which is an encoder-decoder model that converts each task into a text-to-text format. It was trained on the Colossal Clean Crawled Corpus (C4) dataset, a 750 GB dataset created from Common Crawl’s web-extracted text. The model architecture is similar to the original Transformer, with both encoder and decoder consisting of 12 blocks, and it has 220 million parameters [
13]. Following T5, RoBERTa is an enhanced version of BERT with more data, dynamic masking, and byte-pair encoding. RoBERTa improves performance by training longer, using larger batches, removing the next sentence prediction objective, and dynamically changing the masking pattern. RoBERTa was trained on a large dataset called CC-NEWS [
52].
In the realm of models that are fine-tuned for specific tasks, FN-BERT stands out as a BERT-based model fine-tuned on a fake news classification dataset in 2023. The model is based on distilBERT and was trained using a specific fake news dataset [
53]. Meanwhile, Grover, developed by the Allen Institute for AI, is a GAN-based model designed to generate and detect fake news. Grover has three sizes: Grover-Base (117 M parameters), Grover-Large (345 M parameters), and Grover-Mega (1.5 B parameters). It was trained on the RealNews dataset, created from Common Crawl news articles [
54].
Recent advancements include Llama2, a collection of pre-trained and fine-tuned LLM models from Meta with varying parameter sizes. It includes Llama 2 Chat, optimized for dialogue, and Code Llama, specialized for code generation. Llama2 was trained on 2 trillion tokens, with increased context length and grouped-query attention [
55]. Additionally, Meta has developed the Llama3 models, which show significant improvements in performance and scalability compared to their predecessors.
Microsoft’s Phi3 [
56] also presents significant advancements in performance and scalability. It emphasizes improvements in computational efficiency and model adaptability, making it a strong contender in the evolving landscape of large language models. Phi-3 comes in different versions ranging from 3.8 B (Phi-3 mini) to 14 B (Phi-3 medium) parameters, which offer two context lengths, 128 K and 4 K, and outperform bigger models. They have been trained on 3.3 million tokens.
OpenAI’s GPT-3 demonstrated remarkable abilities in translation, question answering, and content generation, comparable to human-written texts. Furthermore, ChatGPT-3.5, also developed by OpenAI, is widely used for its conversational abilities. It was trained using Reinforcement Learning from Human Feedback (RLHF) and fine-tuned from a model in the GPT-3.5 series. ChatGPT can answer follow-up questions, admit mistakes, and reject inappropriate requests [
57]. Continuing this progress, OpenAI’s GPT-4, released in March 2023 [
58], opened new horizons by enabling the understanding and answering of questions from both text and image inputs, fostering a more detailed and comprehensive understanding of information. This progress in the GPT family has raised expectations and broadened the range of their applications.
Alibaba’s Qwen2 [
59], another advanced model, focuses on enhancing accuracy and efficiency in natural language processing. Mistral AI’s Mistral models [
60] and Google’s Gemini models [
61] also contribute to expanding the capabilities of open standards by providing advanced solutions for complex learning problems.
Table 2.
Overview of Key Large Language Models (LLMs).
Table 2.
Overview of Key Large Language Models (LLMs).
Model | Tokens | Parameters | Training Corpus Size | Architecture | Reference |
---|
BERT | 3.3 B | BERT-BASE: 110 M, BERT-LARGE: 340 M | BooksCorpus (800 M words), English Wikipedia (2.5 B words) | Transformer (12/24 layers) | [11] |
M-BERT | 300 M | 110 M | Wikipedia pages of 104 languages | Transformer (12 layers) | [51] |
T5 | 500 M | 220 M | C4 (750 GB) | Encoder-Decoder (12 blocks) | [13] |
RoBERTa | 500 M | 355 M | CC-NEWS, Web texts (160 GB) | Enhanced BERT | [52] |
FN-BERT | 300 M | 66 M | Fake news dataset | Fine-tuned DistilBERT | [53] |
Grover | 400 M | Grover-Base: 117 M, Grover-Large: 345 M, Grover-Mega: 1.5 B | RealNews dataset (120 GB) | GAN-based Transformer | [54] |
Llama2 | 2T | Llama2-7B, Llama2-13B, Llama2-70B | Various datasets | Transformer | [55] |
Llama3 | 15 T | 8 B, 70 B | Massive dataset (incl. Common Crawl, code, data, books) | Improved Transformer | - |
Phi3 | 3.3 M | Small 3.8 B, Medium 14 B | Large-scale dataset | Improved Transformer | [56] |
GPT-3 | 500 B | 175 B | Diverse datasets (web, books, Wikipedia) | Multimodal Transformer | [58] |
GPT-4 | 10 T | 1.7 T | Diverse datasets (web, books, Wikipedia) | Multimodal Transformer | [58] |
Qwen2 | 500 B | 200 B | Massive dataset | Transformer | [59] |
Mistral | 500 B | 175 B | Large-scale dataset | Transformer | [60] |
Gemini | 1 T | 1.4 T | Massive dataset | Advanced Transformer | [61] |
4.2. LLM-Based Fake News and Fake Profiles Generation
The generation of fake news and fake profiles has evolved significantly with the advent of LLMs. Traditionally, the automatic creation of fictitious news pieces relied on simplistic methods such as word jumbles and random replacements within actual news articles [
6,
54,
62]. These rudimentary techniques often resulted in content that was easily recognizable as fake due to its lack of coherence.
With the introduction of LLMs, there has been a substantial shift in the research focus towards creating more coherent fake news. Initial efforts involved using prompting techniques to generate fake news, which aimed to produce more believable narratives [
63,
64]. However, these generated stories were still detectable by other language models because they often lacked sufficient details or consistency.
To address these shortcomings, researchers introduced methodologies that combined real news, factual information, and intentionally false information provided by humans. For example, Su [
20] tasked LLMs with fabricated articles from human-collected summaries of fake events. Similarly, Wu and Hooi refined the generation of fake news articles using LLMs to enhance their believability [
64].
Other innovative approaches include Jiang’s method, which uses fake events combined with real news articles to generate fake news [
10], and Pan’s approach, which employs a question-answer dataset based on real news to manipulate answers and create fake news content [
6,
65]. These strategies, particularly those involving manually crafted fake news, help prevent the automated mass production of fake news articles. Nonetheless, techniques that rely on fabricated summaries often produce content lacking in detail, and any alterations to specific events or elements can lead to issues with contextual coherence.
4.3. LLM-Based Fake News Detection
Detecting fake news has become increasingly sophisticated with the integration of large language models (LLMs). Traditional fake news detection methods often rely on auxiliary data in addition to the article’s text. For instance, mainstream detection programs use metadata such as authors, publishers, and publication dates to determine the authenticity of an article. A notable example is Grover, which leverages these metadata for validation [
54]. Similarly, DeClare compares statements to data collected from web-searched articles to verify their credibility [
66]. Another approach involves extracting emotional and semantic characteristics from texts, as demonstrated by Zhang and colleagues in 2021.
Moreover, platforms like Defend examine the responses garnered by publications on social media to aid in fake news detection [
67]. However, these methods have limitations in practical applications. Auxiliary data are not always accessible, and relying on online fact-checking services like PolitiFact and Snopes is labor-intensive, making it difficult to keep up with the rapid production of misinformation [
68].
Recent studies have shown that fine-tuned pre-trained language models (PLMs) such as BERT, and LLMs like GPT-3.5, can achieve commendable results in fake news detection without the need for extensive auxiliary data [
6,
20,
63,
64]. The primary approaches for LLM-based fake news detection include text classification, fact-checking and verification, and contextual analysis, which will be analyzed in the following subsections.
4.3.1. Fake Text Classification Using LLMs
In recent years, the fine-tuning of pre-trained models on fake news datasets has emerged as a pivotal approach in enhancing the accuracy of fake news detection. This method leverages the robust capabilities of pre-trained models and adapts them specifically for identifying misinformation, leading to significant improvements in performance metrics.
Boissonneault et al. fine-tuned the ChatGPT and Google Gemini models using the LIAR benchmark dataset [
69]. The training set from the LIAR dataset was employed, utilizing high-performance computing resources, including GPUs, to meet the computational demands and ensure efficient processing. During the fine-tuning process, hyperparameters such as learning rate, batch size, and the number of training epochs were systematically adjusted to achieve optimal performance. The test set of the LIAR dataset was used for evaluating the fine-tuned models, comparing their predictions to true labels to calculate performance metrics like accuracy, precision, recall, and F1 score. Google Gemini exhibited slightly better results compared to ChatGPT, as evidenced by higher accuracy, F1 score, and superior performance on the AUC-ROC curve.
Al Esawi et al. explored deep learning techniques for detecting Arabic misinformation by leveraging the contextual features of news article content [
70]. This approach employed a combination of BiLSTM and the attention mechanism inherent in the transformer architecture. A pre-trained AraBERT model was used to extract features from Arabic text, generating contextual embeddings. These embeddings were then fed into a BiLSTM layer, and the outputs were processed through an attention layer to create a context vector. This context vector was further refined through dense layers and a dropout layer to prevent overfitting, enabling the model to classify each article as fake or real.
Zhang et al. applied fine-tuning of the BERT model to create a health misinformation detection classifier [
71]. This classifier processes input text through a transformer architecture comprising multiple layers of bidirectional self-attention mechanisms. The model was trained, validated, and evaluated using three publicly available datasets: CMU-MisCOV19, CoAID (COVID-19 Health Misinformation Dataset), and FakeHealth. The training data were labeled, and the pre-trained weights were adjusted to optimize performance for the task of distinguishing between real and fake health news.
Liu et al. proposed a novel framework named FakeNewsGPT4, which augments Large Vision-Language Models (LVLMs) with forgery-specific knowledge for manipulation reasoning while inheriting extensive world knowledge as a complement [
72]. This framework requires two types of forgery-specific knowledge: semantic correlations between visual and textual data and artifact traces indicating manipulations. It employs a multi-level cross-modal reasoning module and a dual-branch fine-grained verification module to extract this knowledge. The generated knowledge is then converted into embeddings compatible with LVLMs using candidate answer heuristics and soft prompts to enhance input informativeness. Experiments have demonstrated that FakeNewsGPT4 outperforms previous methods in cross-domain fake news detection tasks.
4.3.2. Fact-Checking and Verification
Automated fact-checking and verification using large language models (LLMs) has shown considerable promise in improving the accuracy and efficiency of identifying misinformation. Various innovative approaches have been developed to leverage the capabilities of LLMs in this domain.
One approach, named Fine-grained Feedback with Reinforcement Retrieval (FFRR), aims to enable more informed and accurate evidence retrieval for LLM-based fact-checking on real-world news claims [
73]. This method involves generating intermediate questions from various perspectives of a claim through prompting to retrieve relevant documents. FFRR then collects fine-grained feedback from the LLM on the retrieved documents at both document-level and question-level, using this feedback as rewards to refine the list of retrieved documents and optimize the retrieval policy for intermediate questions. The FFRR approach, evaluated on two public datasets, significantly outperforms state-of-the-art LLM-enabled and non-LLM baselines.
Another method, known as Hierarchical Step-by-Step (HiSS) prompting, directs LLMs to decompose a claim into several subclaims and verify each of them step-by-step through multiple question-answering sessions [
68]. This technique addresses two main challenges in news claim verification: omission of necessary details and fact hallucination. The process includes claim decomposition, subclaim verification, and final prediction. Initially, the LLM decomposes the claim into subclaims to ensure every explicit point is addressed. It then verifies each subclaim by generating and answering a series of insightful questions, consulting external sources like Google Search when necessary. After verifying the subclaims, the LLM makes a final prediction and classifies the original claim. Experiments on two public misinformation datasets show that HiSS prompting outperforms state-of-the-art fully-supervised approaches and strong few-shot ICL-enabled baselines.
Another innovative approach uses LLMs to produce weak labels for 18 credibility signals through prompting [
74]. Initially, an open-ended prompt generates answers to questions regarding the news article text, which are then mapped to predefined classes using a task-agnostic restrictive prompt. For credibility signals, an instruction prompt is used alongside the article text, followed by a credibility signal prompt. This process is repeated for each credibility signal. The zero-shot prompt is used to detect misinformation as a binary classification task. If simple string matching rules fail to map the model’s answer to a label, a task-agnostic category mapping prompt is employed. This method, which combines zero-shot LLM credibility signal labeling and weak supervision, outperforms state-of-the-art classifiers on two misinformation datasets without using any ground-truth labels for training.
4.3.3. Contextual Analysis
Understanding the context of news articles is crucial for effective fake news detection. Various innovative approaches have been developed to leverage the capabilities of large language models (LLMs) for contextual analysis in this domain.
In the work by Hu et al., the Adaptive Rationale Guidance (ARG) network is proposed for fake news detection by combining the advantages of small and large language models [
75]. The goal of ARG is to enable small language models (SLMs) to select useful rationales as references for final judgments. To achieve this, ARG encodes the inputs using SLMs, which utilize LLM-generated informative rationales on news content from various perspectives, such as textual description, commonsense, and factuality, to improve performance. The small and large LMs collaborate via news-rationale feature interaction, LLM judgment prediction, and rationale usefulness evaluation, allowing for rich interaction between news and rationales. The obtained interactive features are aggregated to make a final judgment. Additionally, a rationale-free version, ARG-D, was developed for cost-sensitive scenarios without querying LLMs, demonstrating the superiority of both ARG and ARG-D in experiments.
Wu et al. introduced SheepDog, a style-agnostic fake news detector robust to news writing styles [
76]. This detector achieves adaptability through LLM-empowered news reframing, which customizes each article to match different writing styles using style-oriented reframing prompts. The style-agnostic training enhances the detector’s resilience to stylistic variations by maximizing prediction consistency across these diverse reframings, focusing on content rather than styling.
Wan et al. developed DELL, a framework that identifies three key stages in misinformation detection [
77]. First, LLMs generate synthetic reactions to new articles to represent diverse perspectives and simulate user-news interaction networks. Second, LLMs generate explanations of news articles for proxy tasks, refining feature embeddings with these explanations. Finally, DELL employs three LLM-based strategies to selectively merge the predictions of task-specific experts, enhancing calibration. Extensive experiments on seven datasets with three LLMs demonstrate that DELL outperforms state-of-the-art baselines by up to 16.8% in macro F1-score.
Table 3 presents a comprehensive comparison of studies that employ LLMs for Fake News Detection.
4.4. LLM-Based Fake Profile Detection
Profile Content Analysis
MedGraph is a multi-head attention-based graph neural network (GNN) approach to detect malicious behaviors, i.e., edges from a temporal reciprocal graph [
85] in online dating applications, considering both the user’s features and their interactive behaviors. The method is based on four components. The first component is the motif-based GNN component that defines reciprocal user features through a bipartite graph and uses a motif-based random walk algorithm to sample neighbors based on predefined motifs. A temporal user behavior embedding component discovers abnormal user behaviors based on feature representations of historical interaction data. Then, a co-attention component captures the interactive behavior attributes between users, especially focusing on distinguishing the behavior of abnormal users. Finally, there is the prediction layer to predict if an interaction (edge) is malicious.
4.5. Multi-Modal Approaches
Ayoobi et al. [
41] proposed an approach for detecting fake and Large Language Model (LLM)-generated profiles in the LinkedIn Online Social Network immediately upon registration, and before establishing connections with other users. The approach focuses on the textual and metadata consistency of profile information provided during the registration process and leverages advanced word embeddings from models like BERT or RoBERTa.
The approach proposed in [
86] integrates a profanity detector and gender identification module for fake profile detection, where the profanity detector identifies the level of profanity in the given text -in this case in the comments posted from the user account-, while the gender detector enhances feature accuracy. It uses pre-trained BERT in combination with Logistic Regression to classify profiles as fake or real. The proposed framework considers an optimized set of 12 features in total, which combinate user features, network-based features, and attribute-based features [
86].
5. Comparative Analysis of Traditional and Modern Methods
5.1. Performance Metrics
There is a large variety of evaluation metrics used in the task of fake news or fake profile detection. The metrics consider fake news or profile detection as a binary classification task.
Table 4 summarizes the most widely used evaluation metrics.
In the subsections that follow, we briefly present the performance of various methods discussed in the previous sections in the fake news detection task in various datasets. Where available, the links to the datasets are provided, to allow researchers to further evaluate or improve the methods’ performance. Although the comparison of different methods across different datasets is not straightforward, the performance metrics give us an idea of the difficulty of the binary classification task.
5.2. Traditional Methods
5.2.1. Fake News Detection: Rule-Based Approaches—Keyword Analysis and Heuristic Rules
KAFN-PULP [
23] is a method that operates with few labeled samples (news items). It extracts keywords from the text, constructs a network that contains a few fake and real labeled nodes among all other nodes, and then applies the Graph Attention Neural Event Embedding (GNEE) method to classify unlabeled nodes. The method outperformed its competitors on the FakeNewsCorpus dataset
The authors in [
24] proposed an attention-based ANN with textual, social and image information sources and applied it to the Twitter and Weibo datasets, achieving an accuracy of 75%. The authors in [
25] conducted experiments on both Chinese (using the Weibo21 dataset) and English real-world datasets, with the IKA model. They achieved an accuracy of 90.06% on the English and 93.17% on the Chinese dataset.
The performance of the Naive Bayes classifier in fake news detection [
26] on the BuzzFeed News dataset was lower, with a precision of 71% and a recall of 13% due to the skewness of the data in the test dataset.
5.2.2. Fake Profile Detection Using Profile Attribute Analysis
In the task of user profile classification, the SVM-NN method introduced in [
36] has achieved a classification accuracy of 98% in detecting fake Twitter accounts and bots, using the MIB dataset.
Moreover, the hybrid model described in [
40] achieved an accuracy of 80% in a balanced dataset of 400 Facebook profiles. The accuracy was almost the same when Random Forest, Naive Bayes and Decision Tree classifiers were used, and only the KNN algorithm had a lower accuracy score (at 60%). The authors calculated the percentage of skin exposure in profile images, finding that fake accounts often exhibit higher skin percentages.
In [
37], four well-known data mining techniques were applied, Neural network (NN), Support vector machine (SVM), and Weighted average (WA) to the problem of fake profile identification in LinkedIn. The accuracy for NN was 79.75% when all the features were used and 82.83% using selected features. Respectively, SVM with RBF Kernel achieved an accuracy of 79.24% and 81.83% before and after feature selection. SVM using a Polynomial Kernel achieved the highest accuracy at 87.34%. Finally, simple WA achieved an average accuracy of 76.16% when the datasets contain all features, and 74.68% when the datasets contain selected features.
The authors in [
42] used two fake account datasets, one from Weibo and one from freelancer.in. Using the Random Forest Classifier model combined with pipelining they identified fake profiles in the two datasets with an F1-score of 93% and 95%, respectively.
5.2.3. Fake Profile Detection Using Network Analysis
The authors in [
43] used an enhanced graph-based semi-supervised learning algorithm (EGSLA) and graph-based classification techniques to distinguish fake users from a pool of Twitter users. The performance of their approach was tested on a Twitter dataset comprising 21,473 users and 2,915,147 tweets and they achieved an average F1-measure of 88.95%, outperforming k-nearest neighbor (KNN), support vector machine (SVM), and decision tree (DT) classifiers.
The performance of a fake profile detector framework that builds upon PetriNets [
44] is evaluated using a crawled Twitter dataset comprising 6824 Twitter profiles and 59,153,788 tweets. The method used various classifiers at the final classification step like Random forest, Bagging, JRip, J48, PART, Random tree, and Logistic regression. Logistic regression achieved the highest precision score, which is 99.2%, and Random Forest and Bagging followed with 98.9%.
In [
47], the majority voting approach and other traditional classification methods were tested once again on the MIB dataset of fake profiles. The AdaBoost classifier outperformed all other methods (i.e., 99.22%), whereas the majority voting approach came second with a performance that compared to that of the Random Forest classifier (i.e., 99.12%).
The gray wolf optimization (GWO) method in [
48] was tested on a Facebook profile dataset, which was handcrafted to comprise 1043 real and 201 fake accounts and 14 attributes in total (name, profile picture, likes, groups joined, number of friends, educational status, work, living place, check-ins, etc.). The proposed GWO method achieved an accuracy of 98%, compared to the 97% achieved by DTs, SVMs and KNN, making it a robust choice for identifying fake profiles on social networks. It also achieved an F1-score of 96%, outperforming conventional machine learning algorithms such as K-Nearest Neighbor, Support Vector Model, Decision Tree, and Random Forest.
5.2.4. Fake Profile Detection Using Behavior and Interaction Patterns
The performance of all seven baseline classifiers (J48, Random Forest, Random Tree, REP Tree, OneR, JRip) was evaluated in [
49] on a dataset collected using the Facebook API. The dataset comprised 549 real users who are the direct friends of the authors and 2672 more users, friends of the authors’ friends, who are assumed to be real users. The dataset also contained 230 Facebook spammers and black market users, and another 1257 users from black market services who are also assumed as fake users. The methods were evaluated on a subset of the original dataset and the proposed method achieved an accuracy score in the range of 75–80% and an error rate of 20–25%. Only Naive Bayes gave a lower accuracy score (58%) but could detect 92% of the actual fake profiles, which was the best performance among all models.
A work on Twitter accounts [
19] collected a dataset of 501 fake and 499 real accounts, each account described with 16 features, including: description length, protected account, number of followers, number of following, number of tweets and retweets, number of favorite tweets, number of public lists participating in, verified account, profile background image, contributor mode enabled, etc. The success rate of the Naïve Bayes algorithm tested on this dataset is 80.6% before and 90.9% after discretization.
5.3. LLM-Based Methods
5.3.1. Fake News Classification Using LLMs
In [
69], Boissonneault, et al. fine-tuned the ChatGPT and Google Gemini models using the LIAR benchmark dataset. Google Gemini exhibited slightly better results compared to ChatGPT, as evidenced by higher accuracy, F1-score and superior performance on the AUC-ROC curve.
The authors in [
70] used a pre-trained BERT model and BiLSTM to detect fake content on the ArCOV19-Rumors dataset from Twitter and on the AraNews dataset and achieved an accuracy of 93.0% and 88.1%, respectively. They also applied the Att-BiLSTM model, which boosted the accuracy to 96.9% and 90.5%, respectively, in the two classification tasks.
In [
71] the BERT model was trained on different pre-processed data of the CMU-MisCOV19 dataset. The first case involved using the raw tweet data with an accuracy of 84.1% and the second using tweets without emojis with an accuracy of 87%. The third case uses tweets with emojis and achieves an increased accuracy of 87.5%. The authors also combined the BERT model with different socio-contextual information and normalization of numerical data in association with cleaned textual tweets, achieving an accuracy above 92%. The results demonstrate that incorporating socio-contextual information, including user and tweet information, enhances the model’s performance.
In the case of multi-modal content, the vision-enhanced framework of FakeNewsGPT4 [
72] outperformed all the methods it compared against, such as MFND and HAMMER, showing a 7.7% increase in AUC when tested across different subsets, in the single-domain. In multi-domain the performance of FakeNewsGPT4 remains higher than PandaGPT and HAMMER. The authors employed the DGM benchmark dataset for their experiments.
5.3.2. Fake News Detection via Using LLMs for Fact-Checking and Verification
The FFRR method of [
73] was evaluated on two public datasets (the RAWFC and the LIAR-RAW, significantly outperforming other state-of-the-art models. FFRR on the document and question level achieves the highest F1-score (57%) on the RAWFC dataset and the second-best F1-score (33.5%) on the LIAR-RAW dataset. The HiSS method, presented in [
68], achieved a lower F1-score (53.9%) on the RAWFC dataset, but achieved a better F1-score (37.5%) on the LIAR dataset.
In [
74] different LLMs were used to produce weak labels for 18 credibility signals through prompting, achieving different scores. The highest results on the EuvsDisinfo dataset were achieved using GPT-3.5-Turbo-FULL (99.0% F1-score), whereas the OpenAssistant-30B-FULL model achieved 54.8% F1-score on the FA-KES dataset.
5.3.3. Fake News Detection Using LLMs and Contextual Analysis
The ARG method proposed in [
75] achieved a 78.6% accuracy in the Chinese dataset Weibo21 and an accuracy of 87.8% in the GossipCop English dataset from FakeNewsNet. When ARG is combined with the LLM Judgement predictor, it achieves an accuracy of 77.4% and 88% in the Chinese and English datasets retrospectively. Finally, ARG-D achieves 77.2% and 87% accuracy scores in the Chinese and English datasets, respectively.
The SheepDog method [
76], which extracts content-focused veracity attributions from LLMs, has been tested on the Politifact dataset and achieved an accuracy of 88.44% and on the GossipCop dataset and got an accuracy of 75.77%. On the LUN dataset, the SheepDog method achieved its highest accuracy (93.05%). In all datasets, SheepDog outperformed all other methods that rely on pre-trained models like BERT, RoBERTa, DeBerta, GPT3.5, or InstructGPT.
The Diverse reaction generation Explainable proxy tasks and LLM-Based expert ensemble method, also known as DELL [
77] has been tested on seven datasets related to fake news detection (Pheme and llm-mis), framing detection and propaganda tactic detection. DELL achieved state-of-the-art performance against baselines in all benchmarks. This shows the success of integrating LLMs in multiple stages of news veracity evaluation.
5.3.4. Fake Profile Detection Using Language Patterns and Anomalies in Profile Content Analysis
In a different context than the previous works, the authors in [
85] introduced MedGraph a multi-head attention-based graph neural network (GNN) for detecting malicious behaviors (friendship edge creation) on information exchange networks. The experiments on the UCI message and the Digg datasets show that MedGraph outperformed all other baselines (i.e., Node2Vec, GraphSAGE, DeepWalk, etc.) with an AUC of 0.85 in the UCI message graph and 0.96 in the Digg graph.
5.3.5. Multi-Modal Approaches for Fake Profile Detection
The authors in [
41] developed a dataset of 1800 legitimate LinkedIn profiles (LLPs), 600 fake LinkedIn profiles (FLPs) and 1200 profiles generated by ChatGPT (CLPs). They used word embeddings and trained a classifier to evaluate the set of LLM-generated user profiles. Their proposed method, SSTE achieved an accuracy of 72.26% when using BERT and 76.12% when using RoBERTa. They also used CLPs as fake profiles in the training phase to identify FLPs in the testing phase and achieved 65.39% accuracy with BERT embeddings and 60.38% accuracy with RoBERTa. Finally, when the task was to discriminate between LLPs and FLPs the accuracy of SSTE was 96.33% and 95% with BERT and ROBERTa, respectively.
The approach proposed in [
86] used pre-trained BERT in combination with Logistic Regression for classifying profiles as fake or real. They considered a rich set of impactful features: attribute-based, network-based, and user features and used LLMs to extract a few more features such as gender. The method achieved an accuracy score of 94.65% with the additional extracted features. In contrast, the accuracy with the original features was only 87.32%.
Table 5 provides a summary of all the traditional and LLM-based methods presented in the previous subsections in the two tasks of detecting Fake user profiles and Fake content.
5.4. Datasets
Table 6 summarizes the main datasets found in the literature and used by the studies analyzed in the previous sub-sections.
6. Case Studies and Practical Applications
Various instances of using Large Language Models (LLMs) to combat fake news are available across different platforms, showing their practical uses and how effective they are. The LLMs deployed on social media like Facebook and Twitter help in analyzing the user-generated content in real-time. These models find Fake information by scanning posts, comments, and shared links. Social media platforms can use natural language processing and machine learning techniques to spot fake news patterns such as sensational phrases or unverified sources thus helping them take quick action faster than before hence reducing the spread of misinformation [
88].
The evaluation results of existing implementations have been promising. For example, performance comparisons between rule-based traditional detection systems and LLM-based ones indicated that the latter have higher accuracy of detection for fake news. The spread of misinformation on social media platforms can be reduced using language model machine learning techniques, an improvement from previous methods. User impact and feedback also highlight the effectiveness of LLMs in real-world applications. Users have reported higher trust in platforms that actively utilize LLMs for fake news detection, appreciating the increased reliability of the information provided [
89].
In short, their practical use against misinformation across social media platforms, news aggregator sites, and online marketplaces shows how beneficial large language models can be in reducing falsehoods. This examination points out a better performance by LLMs as compared to traditional methods with significant improvements in terms of precision and user confidence rates. The continuous development and more extensive application of LLMs are critical as we continue to face up to fake news and look forward to digital platforms where dependable information is shared among users.
7. Challenges and Limitations
The use of LLMs for fake news detection comes with many challenges. These challenges arise from both the complexities of the technology and the tricky nature of fake news. In the following, we examine the various limitations of LLMs, which include issues with data quality and bias, difficulties in understanding context, and the high computational power needed. We also discuss the problems that arise when these models have to scale in order to handle large amounts of data, the need for better interpretability, and the ethical and privacy concerns. Furthermore, we explore how these models struggle to adapt to new types of fake content. By examining these aspects in detail, we aim to provide a clear understanding of what needs to be improved to make LLMs more effective in detecting fake news.
Data Quality and Bias: LLMs are trained on large datasets that may contain biases reflecting the prejudices and inaccuracies present in the source data. These biases can influence the model’s outputs, making it difficult to reliably detect fake news. Certain topics, events, or viewpoints may be overrepresented or underrepresented in the training data, leading to skewed detection capabilities [
4].
Contextual Understanding: Fake news often involves nuanced language and contextual subtleties that are challenging for LLMs to fully comprehend. The lack of deep understanding of the context can lead to misclassification. Ambiguous language and the use of irony, sarcasm, or satire in fake news can confuse LLMs, which might struggle to differentiate between genuine and deceptive content [
90].
Computational Efficiency: LLMs are computationally intensive. They require substantial amounts of processing power and memory to function effectively. This high computational demand poses several challenges. The sheer volume of computations needed for inference makes it difficult to deploy these models on devices with limited resources, such as mobile phones or edge devices. Furthermore, running LLMs at scale consumes significant energy, which is both costly and environmentally unsustainable. This high energy consumption can limit the practical application of LLMs in real-time fake news detection. Additionally, ensuring low latency responses is critical for real-time applications like fake news detection. However, the computational complexity of LLMs can introduce delays, reducing their effectiveness in time-sensitive scenarios [
91].
Scalability: Scalability is a major challenge when deploying LLMs for fake news detection across vast and diverse data sources. Processing massive volumes of text data from social media, news outlets, and other online sources requires a highly scalable infrastructure. LLMs struggle to efficiently scale up to handle this volume without significant computational overhead. Implementing LLMs in distributed systems to manage scalability introduces complexities in synchronization, consistency, and data distribution, which can affect the model’s performance and reliability. The financial cost of scaling up LLMs is considerable. The need for powerful hardware and continuous model retraining makes it expensive to maintain large-scale deployments, limiting accessibility for smaller organizations or projects [
74].
Interpretability: Interpretability refers to the ability to understand and explain how LLMs arrive at their decisions. This is particularly important in sensitive applications like fake news detection. LLMs are often described as “black boxes” because their internal decision-making processes are opaque. This lack of transparency makes it difficult to understand why a particular piece of news is classified as fake, reducing trust in the model’s outputs. Providing clear and understandable explanations for the model’s decisions is challenging. Without explainability, it is hard for users to verify the accuracy and fairness of the model’s predictions. In many jurisdictions, regulations require that automated decision-making systems be explainable. The interpretability challenges of LLMs make it difficult to comply with these regulations, posing legal and ethical risks [
92].
Adaptability to New Types of Fake Content: Fake news constantly evolves, with new tactics and forms emerging to evade detection. LLMs face significant challenges in adapting to these changes. LLMs are typically trained on large datasets that do not update in real time. This static nature means they may not be equipped to recognize and respond to new types of fake content that arise after the training period. Implementing continuous learning mechanisms to keep LLMs up-to-date with the latest fake news tactics is complex and resource-intensive. Without this, the models can become outdated quickly. While LLMs are powerful, they can struggle to generalize from known types of fake news to novel forms. The subtle nuances and evolving strategies used in fake content require constant adaptation, which is challenging for models primarily designed for static learning [
72].
Ethical and Privacy Concerns: The deployment of Large Language Models (LLMs) in fake news detection brings significant ethical and privacy challenges that must be meticulously addressed. LLMs require vast amounts of data for training, often including personal and sensitive information. Ensuring these data are protected and handled in compliance with regulations like the General Data Protection Regulation (GDPR) is critical to safeguard individual privacy rights. The potential for data breaches or misuse poses substantial risks, necessitating robust data security measures and transparency in data handling practices [
89]. Besides, the ethical landscape of LLM usage is complex, as these models themselves can be weaponized to generate and disseminate misinformation, exacerbating the very issue they are intended to combat. This dual-use nature of LLMs underscores the need for stringent ethical guidelines and controls to prevent misuse. Ethical considerations also extend to ensuring fairness and avoiding biases within the models, which can inadvertently perpetuate harmful stereotypes or misinformation [
93]. A responsible approach to LLM deployment involves ongoing ethical oversight, the establishment of clear accountability frameworks, and the implementation of rigorous standards to guide the ethical use of these powerful technologies. By addressing these privacy and ethical concerns, the integration of LLMs in fake news detection can be managed more responsibly, ensuring they contribute positively to the integrity of information in the digital age [
94].
8. Conclusions and Future Research Directions
Large language models (LLMs) have a dual role in the fake news world, though they can be misused to produce extremely believable false news articles and create deceptive online profiles. However, their potential as a tool for detection cannot be ignored. Using enhanced natural language processing skills, LLMs are capable of examining massive amounts of text information, detecting inconsistencies and identifying patterns that are indicative of misinformation.
There is no doubt that LLMs’ for fake news detection have limitations and challenges that need to be overcome in order to increase their effectiveness and trustworthiness. One of the most promising areas to focus on is computational modeling. This can be conducted by using a combination of traditional methods and modern deep learning approaches such as hybrid techniques. As an example, we can improve detection accuracy by fusing rule-based systems with neural networks to allow for more detailed analyses of fake content. Furthermore, transformer models, which are among the latest developments in deep learning architectures, offer new possibilities for refining and optimizing LLM performance so they become better equipped to deal with issues relating to fake news complexity.
The inclusion of multimodal data into the fight against fake news presents another huge challenge. Most existing LLM models are centered on text, but fake news is usually composed of other forms of media such as images and videos. We can obtain a better understanding of the content by developing approaches that would allow us to carry out analysis and correlation across different modalities using models. In this way, LLMs will be able to detect more inconsistencies and lies than when they take into account a single modality only. To create more powerful systems to counteract complex strategies employed by contemporary disinformation; it will be necessary to combine textual, video and image analyses.
Another thing that is important in the fight against fake news is working with domain experts. This partnership can help to improve the accuracy of detecting false information through better fact-checking. Additionally, journalists and people from other fields who are involved in producing LLMs can work together to ensure their models are more accurate and take into account context. By doing so, they will be able to create datasets that capture the richness of real-world situations, which would reduce bias and make LLMs more reliable in general. Additionally, this kind of cooperation encourages cross-disciplinary study that combines expertise from fields like computer science, linguistics, psychology and media studies to develop holistic approaches for fighting misinformation.
The use of more durable and transparent models is very significant in terms of improving trust and usability in fake news detection technologies. With the application of explainable AI techniques, it will be possible to understand the reasons behind the models’ decisions thereby addressing the ‘black box’ problem and enhancing interpretability. Such a level of openness on a compliance basis will enable users to have confidence in these tools. Moreover, simplified tools for detecting false information that is understandable by laypersons can be instrumental in empowering wider population groups. As such, we can enhance LLMs’ efficiency and reliability as well as their positive contribution towards ensuring the truthfulness of information in this digital era if we center on these future developments.
In conclusion, this survey paper has highlighted the state of the art in the use of LLMs in the field of fake news, shedding light on both fake profiles and fake content and using the LLMs both as generators and as detectors. It also discussed the challenges and limitations associated with using Large Language Models (LLMs) for fake news detection, including issues related to data quality and bias, contextual understanding, computational efficiency, scalability, interpretability, adaptability, and ethical and privacy concerns. Nevertheless, the potential of LLMs to enhance fake news detection is significant as long as they are addressed by future developments. Therefore, continuing to investigate this issue is paramount so that more accurate models that are faster and open can be created. Secondly, interdisciplinary collaboration should be fostered in order to encourage innovation which will result in holistic solutions blending expertise from different fields. We, therefore, ask researchers, technologists, as well as experts in various fields to join hands in refining LLMs so that their responsible deployment can be ensured thus reinforcing collective effort in fighting against false information that stains the digital landscape’s reputation.