Open access peer-reviewed article

Combating Digital Forgeries: Advanced AI Techniques for Detecting Forgeries of Diverse Data

Aziz Makandar

Shilpa Kaman

Syeda Bibi Javeriya

This Article is part of Communications and Security Section

Article metrics overview

168 Article Downloads

Article Type: Research Paper

Date of acceptance: November 2024

Date of publication: November 2024

DoI: 10.5772/acrt.20240042

copyright: ©2024 The Author(s), Licensee IntechOpen, License: CC BY 4.0

Download for free

Table of contents


Introduction
Related Work
Materials and Methods
Proposed Methodology
Results and discussion
Conclusion
Acknowledgments
Author’s contribution
Funding
Ethical statement
Data availability statement
Conflict of interest

Abstract

Detecting forgeries in diverse datasets involves identifying alterations or manipulations of various digital media, such as images, videos, and documents. There are unique challenges associated with each type of media for efficient forgery detection. This study investigated the effectiveness of both traditional and AI-based feature extraction techniques, combined with machine learning classifiers, in detecting forgeries across three distinct datasets: document images, the CASIA 2.0 standard image forgery dataset, and a video dataset containing real and forged videos. Traditional feature extraction methods such as Local Binary Patterns (LBP), Gray-Level Co-occurrence Matrix (GLCM), color histograms, and RGB to HSV conversion were utilized alongside machine learning classifiers like Support Vector Machines (SVM), Random Forest (RF), and Gradient Boosting Machine (GBM). Additionally, AI-based techniques leveraging pretrained models like ResNet50, VGG16, EfficientNetB0 and ensemble methods with Principal Component Analysis (PCA) were employed. This study evaluated the performance of these methods using metrics such as precision, recall, F1 score, and accuracy across different datasets. The results demonstrate the effectiveness of using ensembled features with PCA in improving the detection capabilities of the models. This study underscores the importance of leveraging traditional and advanced AI methodologies to effectively combat digital forgeries in various media types.

Keywords

  • ensemble methods

  • forgery detection

  • image classification

  • pretrained models

  • video forgery

Author information

Introduction

The digital era has made information a vital tool for communication and dissemination in a variety of fields, including business, science, law, education, politics, media, military, medical imaging and diagnosis, art, digital forensics, intelligence, sports and scientific publications. The authenticity of this digital information is paramount, as it often influences significant decisions and public perception. Images can convey far more information than text, and people are more likely to believe what they see, significantly influencing their judgment. This visual credibility can lead to a series of unintended reactions and consequences. Image forgery is primarily carried out for malicious purposes, such as distorting information, spreading immoral and fake news, fraudulently obtaining money from unsuspecting audiences, damaging the reputations of celebrities or public figures, and exerting adverse political influence on digital platform users. Therefore, clear authentication of images and videos before they are used or shared on digital media platforms is essential. This precaution makes it more difficult for users to unwillingly spread falsified information. However, the advent of sophisticated editing tools has led to an increase in digital forgeries, posing a serious threat to the integrity of information [13].

Digital forgery

Forgery is the act of deliberately altering, creating, or imitating documents, signatures, artworks, or other items with the intent to deceive, mislead, or defraud. The rise of powerful editing tools and Artificial Intelligence (AI)-generated content has made it easier than ever to create sophisticated forgeries that can deceive even trained professionals. Among the most concerning types of digital forgeries are those involving identity documents, general images, and deepfake videos as shown in the following Figure 1. These forgeries pose significant risks, including identity theft, misinformation, financial fraud, and lack of trust in digital media.

  • Identity Document Forgeries: Forged identity documents can facilitate a wide range of criminal activities, from illegal immigration to financial fraud. The ability to detect these forgeries accurately and swiftly is crucial for maintaining security and trust in digital and physical identification processes.

  • General Forgery Images: Manipulated images are prevalent in digital platforms, often used to spread false information, manipulate public opinion, or tarnish reputations. Detecting these forgeries is essential to uphold the integrity of visual information and prevent the harmful consequences of misinformation.

  • Deepfake Videos: Deepfakes represent one of the most challenging types of digital forgeries. These AI-generated videos are convincingly altered appearances and speeches of individuals, creating realistic but entirely fake content. Deepfakes have been used for malicious purposes, including political manipulation, blackmail, and misinformation campaigns. As deepfake technology continues to advance, the need for effective detection methods becomes increasingly urgent.

Given the diverse nature and serious implications of these forgeries, there is an immediate need to develop advanced AI techniques capable of detecting forgeries across different types of data. Traditional methods are often inadequate against the sophistication of modern forgeries. Thus, leveraging state-of-the-art AI models offers a promising solution.

Figure 1.

Types of digital forgeries.

Forgery detection system

Structured workflow for detecting and managing digital forgeries, enhancing the reliability of digital media is shown in Figure 2. The major steps involved are

  • Collecting digital media (images or videos) from various sources such as standard databases, social media platforms, news websites, or personal archives.

  • Converting data to a consistent format to carry out experiments, i.e., normalization and standardization of data. The common preprocessing techniques involve remove noise and artifacts, resizing images, and standardizing video frames.

  • Identifying and extracting relevant features from the media.

  • Applying machine learning or deep learning models to the extracted features to detect anomalies or inconsistencies that indicate forgery.

  • Cross-validating the results using multiple models or techniques to ensure robustness.

Figure 2.

Workflow of digital forgery detection system.

Motivation

The motivation for this research is rooted in the desire to leverage advanced AI techniques to protect against the multifaceted threats of digital forgeries. By addressing the challenges posed by identity document forgeries, general forgery images, and deepfake videos, developing a robust and comprehensive framework for forgery detection, we address the following:

  • Enhance security measures for identity verification processes by accurately detecting forged identity documents.

  • Combat the spread of misinformation by identifying manipulated images.

  • Protect individuals and institutions from the threats posed by deepfake videos.

Related Work

Document image forgery detection

Morphing attacks present a significant threat to face recognition systems, particularly in border control scenarios. These attacks exploit the vulnerabilities of facial recognition systems by altering facial features, making it difficult to distinguish between legitimate and morphed images. To combat this, various morphing attack detection (MAD) algorithms and performance metrics have been developed, offering innovative solutions to detect and reject morphed images during enrollment or verification [4]. One approach utilizes patch-level features and lightweight networks, which significantly improve detection accuracy while maintaining computational efficiency and robustness [5]. Another method enhances detection by analyzing RGB color channels and using error-level analysis combined with selective kernel networks. This approach captures critical feature differences and improves classification accuracy without adding excessive parameters [6].

Despite advances in supervised detection methods, they often struggle with diverse attacks due to limited dataset diversity. To address this, an unsupervised solution using self-paced anomaly detection and convolutional autoencoders has been proposed [7]. This method analyses reconstruction behaviours in attack and bona fide samples, allowing better generalization to unknown attacks. Additionally, new neural network training schemes enhance robustness by modifying training data to improve generality and resistance to morphing attacks [8]. By employing layer-wise relevance propagation, these training schemes ensure more reliable models across all image regions, essential for security applications. Overall, these advancements in detection techniques offer promising solutions for enhancing the security of face recognition systems.

Face morphing attacks compromise face recognition systems (FRS), hence the need for development of MAD techniques. The fairness of six single image-based MAD (S-MAD) methods across four ethnic groups was evaluated in [9] and bias was revealed through fairness discrepancy rate (FDR) underscoring the need for more equitable MAD approaches. A survey [10] reviewed progress in face morphing, covering image generation techniques and state-of-the-art MAD algorithms in the context of public databases, vulnerability assessments, benchmarking, and future challenges in biometrics.

Image forgery detection

Traditional approaches to tampered image detection often involve extracting handcrafted features that capture specific properties of images, such as texture and color. These features are then used to train machine learning classifiers to distinguish between authentic and tampered images. In a review of recent deep learning (DL) techniques [11] for detecting image forgeries, particularly copy-move and splicing attacks, as well as deepfake content, the authors highlighted the superior performance of DL methods on benchmark datasets and possible future research directions in DL architectures, evaluation methods, and dataset development.

Image splicing is detected by analyzing and characterizing texture content within different regions of an image [12]. Detection of copy-move and splicing forgeries has been described in [13] by analyzing texture changes in the Cb and Cr components of YCbCr images using a standard deviation filter and higher-order texture descriptors, followed by SVM classification. A study [14] explored feature extraction and classification techniques for detecting multimedia forgery, using a combination of GLCM, GLDM, GLRLM, and Local Binary Pattern (LBP) features. Transform domain techniques combined with LBP for tampered image detection using machine learning techniques were proposed in [15]. Utilizatiob of Haralick-based textural descriptors and Extreme Learning Machine (ELM) to detect image tampering by applying LBP on YCbCr color bands and classifying features with ELM has been described in [16]. Segmentation-based fractal texture analysis (SFTA) with K-nearest neighbors (KNN) was proposed for copy-move forgery detection in [17]. Extracting and combining SIFT and KAZE features from the HSV color channels for detecting copy-move forgery (CMF) was proposed in [18]. Spliced image forgery detection using adaptive over-segmentation combined with AKAZE, ORB, and SIFT feature descriptors was proposed in [19]. Detecting and localizing CMF by applying SIFT key points on CLAHE-enhanced sub-images in RGB and Lab color spaces was presented in [20]. A CMF detection method using color features and hierarchical feature point matching was presented in [21].

Deepfake forgery detection

Various digital image tamper detection methods, focusing on the challenges posed by advanced editing tools and DL techniques like GANs, analyze both traditional and DL approaches for detecting image forgeries, their strengths and limitations are elucidated in [22]. A comprehensive review [23] of deepfake detection techniques using DL algorithms categorizes the methods of video, image, audio, and hybrid multimedia tamper detection. It also posits how deepfakes are created and detected, recent advancements, weaknesses in current security methods, and areas for further research. The results highlight convolutional neural networks (CNN) as the most commonly used DL approach, with video deepfake detection being the primary focus, particularly on improving accuracy. The use of DL in creating and detecting deepfakes, which are often difficult to distinguish from real images has been explored in [24]. With growing concerns about privacy and security, various deepfake detection methods have emerged and a DL-based image enhancement technique to improve the quality of generated deepfakes has also been proposed.

A novel framework for detecting fake videos using transfer learning with autoencoders and a hybrid model of CNN and recurrent neural networks (RNN) is worth noting [25]. Convolutional vision transformer for detecting deepfakes, combining CNN for feature extraction with vision transformers (ViT) for classification using an attention mechanism has been proposed [26]. Robust methods for detecting deepfake videos by analyzing teeth and mouth movements, which are challenging to replicate accurately are presented in [27] by incorporating multi-transfer learning techniques with models such as DenseNet121, DenseNet169, EfficientNetB0, EfficientNetB7, InceptionV3, MobileNet, ResNet50, VGG16, VGG19, and Xception. Detecting deepfake videos using CNNs with transfer learning has been explored in [28].

Using the recent advancements in ViT for face forgery detection, a parameter-efficient ViT-based detection model that includes lightweight forgery feature extraction modules, enabling the model to extract global and local forgery clues simultaneously was proposed in [29]. Generalized multi-scenario deepfake detection framework (GM-DF) presented in [30] showed strong performance across five datasets on unseen scenarios. GM-DF uses hybrid expert modelling for domain-specific features, CLIP for cross-domain alignment, and a masked reconstruction mechanism to capture detailed forgeries. Latent space data augmentation model proposed in [31], expanded the forgery feature space by simulating diverse forgeries within the latent space. This approach enhanced domain-specific features and smoothed transitions across forgery types, resulting in a robust, generalizable deepfake detector that outperformed state-of-the-art models on multiple benchmarks. A framework that captures broader forgery clues by combining diverse local representations into a global feature, guided by local and global information integrity principles to optimize feature relevance was proposed in [32]. Deepfake detection network combining spatial and noise domains proposed in [33], used a dual feature fusion module and a local enhancement attention module, which captured complementary forgery features, achieving superior performance across mainstream datasets.

Materials and Methods

Dataset description

Dataset1 (Aadhaar Card Images Dataset): The images in the dataset were sourced from roboflow.com platform [34] to create a custom dataset for research purposes [35]. These images were resized from 640 × 640 to 300 × 300 pixels to standardize their dimensions. Additionally, faces within the Aadhaar card images were morphed using faceswapper.ai tool [36], ensuring that the identities of individuals in the images remain unidentifiable. In order to further enhance the dataset’s diversity, various image augmentation techniques were applied to generate 3085 images consisting of 1532 real images and 1553 morphed images for use in our study.

Dataset2 (CASIA 2.0 image tampering detection): This dataset [37] consists of 7492 authentic images; 5125 tampered images and 5123 ground truth images. Input images are of resolution 384 × 256 or 256 × 384. Authentic images are of 8 different categories: animal, architecture, art, characters, indoor, nature, plants, and texture. Tampered images comprise spliced and copy-move forgeries. The database is made publicly available in Kaggle for researchers to compare and evaluate their proposed tampering detection techniques.

Dataset3 (Small-scale Deepfake Forgery Video Dataset-SDFVD): This custom dataset, which includes both actual and deepfake videos in a variety of scenarios, was developed to research and evaluate deepfake detection algorithms [38]. There are 106 videos in all, 53 of which are original and 53 of which are deepfakes. The original videos were obtained from Pexels, a popular source of stock footage and photos [39]. These videos depict a wide range of events with persons of all ages and genders from diverse origins. Remaker AI [40] was used to create the deepfake videos by employing face-swapping techniques.

Traditional techniques

Traditional approaches for forgery detection such as texture-based and color-based, often rely on analyzing the intrinsic properties of digital media, leveraging various statistical and structural characteristics. Texture-based techniques are widely used due to their effectiveness in capturing fine details and patterns that can indicate tampering, whereas color-based techniques are essential for detecting image forgeries as they analyze the color properties and distributions within an image.

Texture-based

Two prominent texture-based techniques used in the proposed work are local binary patterns (LBP) and gray level co-occurrence matrix (GLCM).

Local binary patterns

LBP is a simple yet powerful texture descriptor used for image analysis, which works by comparing each pixel in an image to its neighboring pixels typically in a 3 × 3 grid. A binary value (0 or 1) is assigned to each neighboring pixel based on whether its value is less than or greater than the central pixel’s value. The binary values are concatenated to form a binary number representing the local texture pattern. A histogram of the LBP values is constructed over the entire image or a specific region, capturing the distribution of local textures. LBP is an effective feature for texture categorization and is frequently used for forgery detection in digital images due to its effectiveness in capturing local texture patterns [41, 42]. LBP calculation is expressed as

  • For a given pixel P (x, y) in the image, consider its 3 × 3 neighborhood.

  • The central pixel pixel P (x, y) is compared with each of its 8 neighbors.

  • The result is a binary value pixel bi which is 1 if the neighbor pixel Ni is greater than or equal to the central pixel, otherwise 0.

    where

Gray level co-occurrence matrix

GLCM is a powerful tool used in texture analysis to describe the spatial relationship between pixel intensities in an image. GLCM is created by counting how often pairs of pixels with specific values and spatial relationships (defined by distance and angle) occur in an image. Various statistical properties are extracted from GLCM to quantify the texture of an image [43]. GLCM Calculation of GLCM is as follows.

For an image with gray levels g1, g2, …, gN, the GLCM P (i, j) at distance d and angle 𝜃 is defined as:

  • P (i, j)|d, 𝜃 = Number of times gi and gj at distance d and angle 𝜃.

  • GLCM is a matrix where the element P (i, j) represents the frequency of occurrence of pixel pairs gi, gj in the image.

Texture features that are extracted from GLCM in our study are:

  • Contrast: Measures the intensity difference between a pixel and its neighbour.

  • Homogeneity: Measures how close the distribution of elements in the GLCM is to the diagonal, indicating uniform texture.

  • Energy (ASM): Measures textural uniformity by summing the squared elements of GLCM.

  • Correlation: Assesses how correlated a pixel is to its neighbor, indicating linear dependencies.

Color-based techniques

Color-based techniques are valuable tools for detecting image forgeries by analyzing various color properties and distributions. These methods can identify inconsistencies and anomalies in color, lighting, and patterns that are indicative of tampering, thus enhancing the reliability and accuracy of forgery detection systems.

Colour histograms

Forgeries may introduce sharp edges or unnatural transitions in color, which can be detected by analyzing the histograms of different image regions. It works by calculating the color histograms for different regions of an image, comparing these histograms to identify inconsistencies in color distribution. Significant differences in color histograms between regions indicate possible forgeries [44]. For an image, a histogram H (c) is calculated by counting the number of pixels that fall into each color bin c:

where:
  • I (x, y) is the color value of the pixel at position.

  • H (c) is the count of pixels for color bin c.

RGB to HSV conversion

Converting an image from RGB (Red, Green, Blue) color space to HSV (Hue, Saturation, Value) color space helps to separate color information (hue) from intensity information (value). This is useful in detecting color anomalies that may not be apparent in the RGB space. It works by analysing hue, saturation, and value channels separately to detect inconsistencies in any of these channels that might indicate tampering [45]. RGB to HSV conversion is a color space transformation often used in image processing and is computed using

Machine learning classifiers

In the field of forgery detection, machine learning classifiers play a crucial role in distinguishing between authentic and manipulated media. Among the various classifiers, random forest (RF), support vector machine (SVM), and gradient boosting machine (GBM) are widely used due to their robust performance and ability to handle complex patterns [46, 47].

∙  Random Forest: It is an ensemble learning method that constructs multiple decision trees during training and merges their outputs to make a final prediction. Each tree in the forest is trained on a random subset of the data and a random subset of features, which helps in reducing overfitting and improving generalization.

where ht(x) is the prediction of the tth decision tree.

∙  Support Vector Machine: SVM is a supervised learning algorithm that finds the optimal hyperplane which maximizes the margin between different classes in the feature space. It can perform linear as well as non-linear classifications using kernel functions like polynomial, radial basis function (RBF), etc.

where w is the weight vector and b is the bias.

∙  Gradient Boosting Machine: It is an ensemble technique that builds models sequentially, with each new model attempting to correct the errors of the previous models. It combines the outputs of multiple weak learners (typically decision trees) to create a strong predictive model with stagewise additive modelling and is expressed as

where n is the learning rate, and hm(x) is the mth weak learner.

AI techniques

AI techniques have proven to be highly effective in identifying tampered images. This approach typically involves two primary steps: feature extraction and classification. Pretrained models such as ResNet50, VGG16, and EfficientNetB0 are employed for feature extraction, and CNNs are utilized for classification.

Feature extraction with pretrained models

Pretrained models like ResNet-50, VGG16, and EfficientNetB0 are CNNs trained on large datasets (e.g., ImageNet). When using these models for feature extraction, we remove the final classification layers and use the output of the last convolutional layer or a fully connected layer (before softmax layer) as features. Computations at convolutional layer, max pooling layer and then feature concatenation are expressed as

where:
  • Xi, j, c is the input image pixel at position (i, j) in channel c,

  • Wm, n, c, k is the filter weight at position (m, n) for channel ccc and filter k,

  • bk is the bias for filter k,

  • Zi, j, k is the output feature map at position (i, j) for filter k.

where Pi, j, k is the pooled feature map.

After the convolutional and pooling layers, the feature maps are flattened into a 1D feature vector.

where ∥ denotes concatenation.

ResNet50

ResNet50 has a deep CNN architecture as shown in the following Figure 3 and is widely used for image classification tasks, for its depth and efficiency in learning complex patterns [48]. Some of the key components are given below.

  • Residual Blocks: Introduces skip connections (shortcuts) that allow the network to learn residual mappings instead of directly fitting to desired mappings. This helps in tackling the vanishing gradient problem during training.

  • Convolutional Layers: Uses filters to convolve over the input data, extracting features at multiple levels of abstraction.

  • Pooling Layers: Performs downsampling to reduce the spatial dimensions of the input, aiding in feature extraction and computational efficiency.

  • Fully Connected Layers: Integrate extracted features and produce the final classification output.

Figure 3.

ResNet50 architecture [49].

Visual Geometry Group16 (VGG16)

VGG16 is a deep CNN [50] known for its high accuracy in image classification. The architecture of vgg16 is given in Figure 4 and the key components are as follows.

  • Convolution layer applies filters that perform a dot product with areas of the input image to produce an output.

  • Pooling layer reduces dimensionality and complexity by down-sampling the activation map, which helps to speed up processing and prevent overfitting.

  • Max Pooling selects the maximum value from each region in the input image, similar to the convolution layer, but focusing on reducing the size of the data.

  • Softmax layer outputs a probability distribution over the labels, representing the likelihood of each label being correct.

Figure 4.

VGG16 architecture [51]. [image source: https://iq.opengenus.org/vgg16/].

EfficientNet B0

EfficientNet B0 architecture is designed to achieve state-of-the-art performance while maintaining computational efficiency [52, 53]. Following are the key components of efficientnet B0.

  • Compound Scaling to balance network depth (number of layers), width (number of channels), and resolution (input image size). This scaling strategy optimizes both accuracy and efficiency across different computational constraints.

  • Convolutional Blocks uses a combination of depthwise separable convolutions and standard convolutions to extract features efficiently. Depthwise separable convolutions reduce the number of parameters and computational cost compared to traditional convolutions.

  • Squeeze-and-Excitation (SE) Blocks improve feature representation by adaptively recalibrating the responses of channel-wise features. This enhances the model’s focus on informative features and boosts overall performance.

  • Efficient Scaling refers to scaling the network architecture uniformly in all dimensions (depth, width, and resolution) based on a predefined compound scaling method, which ensures a balanced trade-off between accuracy and computational efficiency.

Principal component analysis (PCA) for dimensionality reduction after feature extraction

PCA is a powerful technique used for dimensionality reduction in machine learning. After extracting features from images or other types of data, the dimensionality of the feature vectors can be quite high, leading to challenges such as increased computational complexity, risk of overfitting, and difficulties in visualization. PCA helps address these issues by transforming the high-dimensional data into a lower dimensional space while preserving as much variance as possible [54]. Following are the steps to perform PCA:

  • Consider a dataset with highly correlated features X1 and X2.

  • Standardize: Subtracting the mean and dividing by the standard deviation for each feature.

Computing the mean of the data and subtracting the mean from the data is as follows:

  • Compute Covariance Matrix: Calculate the covariance between X1 and X2.

  • Eigenvectors and Eigenvalues: Solve for eigenvalues and eigenvectors of the covariance matrix.

  • Select the top k eigenvectors corresponding to the largest eigenvalues to form the projection matrix W:

  • Transform Data: Project the original data onto the first principal component, reducing it to one dimension with most variance retained.

Classification with CNNs

Once the features are extracted, the next step is to classify the image as either authentic or tampered. This is achieved using a CNN designed for binary classification [55]. The architecture of a CNN for binary classification is shown in Figure 5.

  • The CNN consists of multiple convolutional layers, pooling layers, and fully connected layers that further process the feature vectors to learn distinguishing patterns.

  • The final layer of the CNN is sigmoid layer which outputs the probability of the image being authentic or tampered.

  • The CNN is trained using a labelled dataset of authentic and tampered images. The network learns to minimize the classification error by adjusting its weights through backpropagation.

Figure 5.

Architecture of CNN for binary classification [56].

Proposed Methodology

Forgery detection using traditional techniques

This process typically includes texture-based and color-based feature extraction methods, followed by classification using machine learning algorithms as shown in Figure 6 and the steps involved are as follows:

Figure 6.

Workflow of forgery detection system using traditional techniques.

Step 1: Read input data containing authentic and tampered images/videos (Aadhaar card images, CASIA 2.0, SDFVD).

Step 2: Apply preprocessing techniques to the input data such as standardize the size of images, converting them to grayscale (for texture-based techniques) and split input videos into frames.

Step 3: Extract features using following methods

  • LBP: Compute the LBP for each pixel and generate histogram of LBP values.

  • GLCM: Compute GLCM for specified distances and angles, extract properties such as contrast, dissimilarity, homogeneity, energy and correlation.

  • Color histograms: Compute the histogram for each color channel (RGB) and normalize the histograms.

  • RGB to HSV Conversion: Convert the image from RGB to HSV color space. Extract and normalize histograms for each channel (H, S, V).

Step 4: Generate feature vectors from LBP, GLCM, Color histograms and RGB to HSV Conversion separately and use them for further processing.

Step 5: The data is split into training and testing sets with 20% of the data reserved for testing.

Step 6: Use the extracted features for training the classifiers like Random Forest, SVM and GBM.

Step 7: Evaluate the models on the test sets to check the accuracy and performance with the help of metrics like precision, recall, F1 score and accuracy.

Forgery detection using AI techniques

This process typically includes feature extraction from pretrained models, followed by classification using DL algorithms as shown in Figure 7 and the steps are as follows:

Figure 7.

Workflow of forgery detection system using advanced AI techniques.

Step 1: Read input data containing authentic and tampered images/videos (Aadhaar card images, CASIA 2.0, SDFVD).

Step 2: Apply preprocessing techniques to the input data such as resizing images to 224x224 pixels and split input videos into frames.

Step 3: Features are extracted from the images using pretrained models such as ResNet-50, VGG16 and EfficientnetB0 without the top classification layer. Later on, we have also ensembled these pretrained models for feature extraction and concatenated all the features to form a single feature vector for each image. PCA is applied to reduce the number of features.

Step 4: The data is divided into training and testing sets, with 20% of the data used for testing.

Step 5: A simple CNN classifier was constructed with an input layer for the feature vectors, two hidden layers using ReLU activation and dropout for regularization, and an output layer with a sigmoid activation function for binary classification. The model was compiled using the Adam optimizer and binary cross-entropy loss.

Step 6: Custom CNN classifier is built with: convolutional layers with 32, 64 and 128 filters, followed by max pooling layers, flattened the model, followed by dense layer by 256 and 128 units, dropout layers, an output layer with a sigmoid activation function for binary classification.

Step 7: Classification report is generated to evaluate the performance of the classifier, providing precision, recall, and F1 score for each class.

Step 8: Compare the performance of both simple and custom CNN models on individual features extracted as well as on ensembled features from the pretrained models.

Results and discussion

Evaluation metrics

Accuracy, recall, precision, and F1 score are the assessment metrics that are used to gauge how well the suggested algorithm performs.

∙  Accuracy: It is determined by applying the following equation to the percentage of images that can be correctly identified.

where: TP is true positives, TN is true negatives, FP is false positives, FN is false negatives.

∙  Precision: The algorithm’s positive predictions are accurate to a certain degree, which is measured by precision. In other words, it is the proportion to find positivity of the predictions by considering true positives and both true and false positives.

∙  Recall: Recall, also known as sensitivity or true positive rate, is a metric used to assess an algorithm’s ability to find all positive cases. It evaluates appropriately detected positive examples. Along with actual positive values, it also takes into account misleading negative values.

∙  F1 Score: It is calculated as the harmonic mean of recall and precision values. The value of F1 score ranges from 0 to 1, prediction values near to 1 means model is performing really well.

Experimental results

Detailed results and analysis of our forgery detection experiments using both traditional feature extraction techniques combined with machine learning classifiers, and advanced AI-based methods leveraging pretrained convolutional neural networks (CNNs) are summarized in Tables 1, 2 and 3 along with performance metrics Precision, Recall, F1 score and Accuracy for various classifiers SVM, RF, GBM applied to different traditional feature extraction techniques LBP, GLCM, Color Histogram, RGB to HSV conversion evaluated on three different datasets respectively.

PrecisionRecallF1 scoreAccuracy
ClassifiersLBP
SVM 0.92 0.9 0.91 0.89
RF 0.94 0.91 0.92 0.9
GBM 0.93 0.92 0.92 0.91
GLCM
SVM 0.88 0.87 0.87 0.86
RF 0.9 0.88 0.89 0.87
GBM 0.91 0.89 0.9 0.88
Color Histogram
SVM 0.85 0.83 0.84 0.82
RF 0.86 0.84 0.85 0.83
GBM 0.87 0.85 0.86 0.84
RGB to HSV conversion
SVM 0.8 0.78 0.79 0.77
RF 0.82 0.8 0.81 0.79
GBM 0.83 0.81 0.82 0.80

Table 1

Results of forgery detection using traditional feature extraction techniques with machine learning classifiers on Dataset1.

PrecisionRecallF1 scoreAccuracy
ClassifiersLBP
SVM 0.89 0.88 0.88 0.86
RF 0.91 0.89 0.90 0.87
GBM 0.90 0.89 0.89 0.88
GLCM
SVM 0.85 0.83 0.84 0.82
RF 0.87 0.85 0.86 0.84
GBM 0.88 0.86 0.87 0.85
Color Histogram
SVM 0.81 0.79 0.80 0.78
RF 0.83 0.81 0.82 0.80
GBM 0.84 0.82 0.83 0.81
RGB to HSV conversion
SVM 0.78 0.76 0.77 0.75
RF 0.80 0.78 0.79 0.77
GBM 0.81 0.79 0.8 0.78

Table 2

Results of forgery detection using traditional feature extraction techniques with machine learning classifiers on Dataset2.

PrecisionRecallF1 scoreAccuracy
ClassifiersLBP
SVM 0.87 0.85 0.86 0.84
RF 0.88 0.86 0.87 0.85
GBM 0.89 0.87 0.88 0.86
GLCM
SVM 0.83 0.81 0.82 0.80
RF 0.85 0.83 0.84 0.82
GBM 0.86 0.84 0.85 0.83
Color Histogram
SVM 0.79 0.77 0.78 0.76
RF 0.81 0.79 0.80 0.78
GBM 0.82 0.8 0.81 0.79
RGB to HSV conversion
SVM 0.75 0.73 0.74 0.72
RF 0.77 0.75 0.76 0.74
GBM 0.78 0.76 0.77 0.75

Table 3

Results of forgery detection using traditional feature extraction techniques with machine learning classifiers on Dataset3.

Figure 8 illustrates the performance of the GBM classifier when applied to various feature extraction techniques on Dataset1. The comparative analysis highlights the effectiveness of each technique in detecting forgeries and shows that LBP provides good results. Figure 9 also presents a comprehensive comparison of the accuracy of different classifiers; SVM, RF, GBM across three datasets. It demonstrates how each classifier performs with traditional feature extraction techniques, providing insights into their relative strengths and weaknesses.

Figure 8.

Results of various feature extraction techniques with GBM classifier on dataset1.

Figure 9.

Accuracy comparison of different classifiers used with traditional feature extraction on three datasets.

Tables 4, 5 and 6 detail the performance metrics of pretrained CNN models ResNet50, VGG16, EfficientNetB0 and custom CNN architectures, along with ensembled features with PCA, on dataset1, dataset2 and dataset3 respectively. Performance metrics include Precision, Recall, F1 score, and Accuracy. Comparative analysis of these techniques is represented in Figures 10, 11 and 12 respectively.

PrecisionRecallF1 scoreAccuracy
ClassifiersResnet50
CNN 0.92 0.9 0.91 0.9
Custom CNN 0.94 0.92 0.93 0.92
VGG16
CNN 0.91 0.89 0.9 0.89
Custom CNN 0.93 0.91 0.92 0.91
EfficientnetB0
CNN 0.93 0.91 0.92 0.91
Custom CNN 0.95 0.93 0.94 0.93
Ensembled features with PCA
CNN 0.94 0.92 0.93 0.92
Custom CNN 0.96 0.94 0.95 0.94

Table 4

Results of forgery detection using pretrained models and CNNs on Dataset1.

PrecisionRecallF1 scoreAccuracy
ClassifiersResnet50
CNN 0.93 0.91 0.92 0.91
Custom CNN 0.95 0.93 0.94 0.93
VGG16
CNN 0.92 0.9 0.91 0.9
Custom CNN 0.94 0.92 0.93 0.92
EfficientnetB0
CNN 0.94 0.92 0.93 0.92
Custom CNN 0.96 0.94 0.95 0.94
Ensembled features with PCA
CNN 0.95 0.93 0.94 0.93
Custom CNN 0.97 0.95 0.96 0.95

Table 5

Results of forgery detection using pretrained models and CNNs on Dataset2.

PrecisionRecallF1 scoreAccuracy
ClassifiersResnet50
CNN 0.91 0.89 0.9 0.89
Custom CNN 0.93 0.91 0.92 0.91
VGG16
CNN 0.9 0.88 0.89 0.88
Custom CNN 0.92 0.9 0.91 0.9
EfficientnetB0
CNN 0.92 0.9 0.91 0.9
Custom CNN 0.94 0.92 0.93 0.92
Ensembled features with PCA
CNN 0.93 0.91 0.92 0.91
Custom CNN 0.95 0.93 0.94 0.93

Table 6

Results of forgery detection using pretrained models and CNNs on Dataset3.

Figure 10.

Comparison of results of various pretrained models on Dataset1.

Figure 11.

Comparison of results of various pretrained models on Dataset2.

Figure 12.

Comparison of results of various pretrained models on Dataset3.

Table 7 presents a comparative analysis of the proposed system for forgery detection against existing methods described in the literature. The comparison covers different datasets, types of data such as images or videos, methods employed, and their corresponding detection results in terms of accuracy. The proposed system demonstrates superior performance compared to existing methods across different datasets and types of data. Its higher accuracy rates 95.23% for image datasets and 93.34% for video datasets highlight the effectiveness of combining ensembled features with PCA for forgery detection. This approach not only improves detection accuracy but also maintains computational efficiency, making it a robust solution for both image and video forgery detection tasks.

Article DatasetType of dataMethodsResults
Doegar et al. [57] MICC-f220ImagesAlexnet model with SVM93.94%
Saikia et al. [58]FaceForensics++VideosHybrid CNN-LSTM model with optical flow features91.21%
Pokroy and Egorov [59] DFDCVideosPretrained EfficientNet B574.4%
Malolan et al. [60] FaceForensicsImagesLRP, LIME94.3%
Patel et al. [61] DFDCVideosMobileNet and Random Forest90.2%
Proposed System Casia 2.0 and Adhaar card image datasetImagesEnsembled features (Pretrained NN) with PCA 95.23%
Proposed System SDFVD- Small-scale Deepfake Forgery Video DatasetVideosEnsembled features (Pretrained NN) with PCA 93.34%

Table 7

Comparison of proposed system with the existing methods.

Role of AI tools in fraud detection

∙  Pretrained Models (ResNet-50, VGG16, EfficientNetB0) for Deep Feature Extraction: Pretrained models, originally trained on large-scale datasets like ImageNet, can extract deep, high-level features from images. Unlike traditional handcrafted features, these deep features capture complex patterns, textures, and structures within the images, which are critical for identifying subtle manipulations. For images and videos, pretrained models can distinguish between genuine and forged images by recognizing anomalies in the texture or layout that simpler techniques might overlook.

∙  Ensembling of Models for Enhanced Feature Representation: Ensembling different pretrained models, leveraging the strengths of each model, for instance, ResNet-50 may capture detailed spatial hierarchies, VGG16 may focus on edge and texture information, and EfficientNetB0 balances performance and computational efficiency. Combining these features provides a richer and more comprehensive representation of the image. In fraud detection involving normal images, such as detecting tampered photos on social media, an ensemble of these models would provide a robust detection system capable of identifying a wide range of manipulations, from minor to major alterations.

∙  CNN Classifier: After feature extraction, a custom CNN is employed to classify images as “fake” or “authentic.” The CNN architecture is specifically tailored to process the combined feature vector and make accurate predictions. For videos, the custom CNN can detect deepfakes by analyzing subtle inconsistencies in facial expressions, lighting, or shadows that may not align perfectly due to the manipulation.

Advantages over traditional methods

∙  Higher Accuracy and Robustness: AI tools, particularly DL models, have demonstrated superior accuracy in tasks like image recognition and classification compared to traditional methods (LBP, GLCM). They are better equipped to handle the complexity and variety of real-world data. Traditional methods might fail to detect sophisticated forgeries involving small, nuanced changes, but a DL-based approach can pick up on these subtleties due to its ability to learn complex patterns from the data.

∙  Adaptability: DL models can be fine-tuned on specific datasets, making them highly adaptable to different types of fraud detection tasks. This adaptability ensures that the models remain effective even as the nature of forgeries evolves. In a scenario where new types of document forgeries emerge, the AI models can be retrained or fine-tuned on updated datasets to quickly adapt to new fraud patterns.

∙  Automation and Scalability: AI tools automate the process of image identification and fraud detection, which is crucial for large-scale applications like scanning millions of images or videos on social media platforms. This automation reduces the need for manual inspection, making the process faster and more scalable. In the financial sector, AI tools can automatically scan thousands of transaction documents daily to detect any signs of forgery, significantly improving efficiency and accuracy.

Real-time scenarios

∙  Consider a scenario where government-issued IDs are being forged. The AI tools analyze the texture, layout, and fonts in the document images, detect inconsistencies, such as slight deviations in font style or spacing, which are indicative of forgery. This leads to the early detection of forged documents, preventing potential identity theft or fraud.

∙  On social media platforms, AI tools are used to detect fake images that have been manipulated for malicious purposes (e.g., spreading misinformation). Such AI applications can identify signs of tampering including unnatural lighting or shadows, that may not be detectable by the naked eye. The platform can automatically flag or remove these images, maintaining the integrity of the content shared with users.

∙  Deepfakes represent one of the most challenging forms of video manipulation. The AI tools analyze facial movements, expressions, and other subtle cues in the video. This helps in preventing the spread of malicious deepfake videos, which could be used for blackmail or misinformation.

By leveraging advanced AI tools like pretrained DL models and custom CNNs, the proposed system provides a powerful and scalable solution for fraud detection. These tools are well-suited to handle the complexities and evolving nature of digital forgeries, offering significant advantages in terms of accuracy, robustness, and adaptability. The examples provided illustrate how these tools can be effectively applied across different scenarios.

Conclusion

This study comprehensively evaluated the performance of traditional and AI-based feature extraction techniques, along with various machine learning classifiers, in detecting digital forgeries across three diverse datasets. Traditional feature extraction methods such as LBP, GLCM, color histograms, and RGB to HSV conversion, while effective to a certain extent, showed limitations in handling the complexity and diversity of forgery types present in the datasets. The results demonstrate that AI-based feature extraction methods, particularly those leveraging pretrained models like ResNet50, VGG16, and EfficientNetB0, significantly outperform traditional methods in terms of precision, recall, F1 score, and accuracy. AI-based methods, especially when enhanced with ensemble techniques and dimensionality reduction through PCA, demonstrated superior adaptability and accuracy with advanced classifiers such as CNNs and custom CNN architectures. This study underscores the critical role of advanced AI techniques in addressing the sophisticated challenges posed by digital forgeries and offers a robust framework for future developments in this field.

Acknowledgments

The authors acknowledge the use of QuillBot for language polishing of the manuscript. 

Author’s contribution

Makandar, Aziz: Resources, Supervision, Project administration, writing review & editing; Kaman, Shilpa: Conceptualization, Formal analysis, Methodology, Investigation, Writing original draft; Javeriya, Syeda Bibi: Data curation, Software, Validation, Visualization.

Funding

This research did not receive external funding from any agencies.

Ethical statement

Not applicable.

Data availability statement

Source data is not available for this article.

Conflict of interest

The authors declare no conflict of interest.

References

  1. 1.
    Amerini I, Ballan L, Caldelli R, Del Bimbo A, Serra G. A sift-based forensic method for copy–move attack detection and transformation recovery. IEEE Trans Inform Forensics Secur. 2011;6(3):10991110.
  2. 2.
    Al-Qershi OM, Khoo BE. Passive detection of copy-move forgery in digital images: state-of-the-art. Forensic Sci Int. 2013;231(1–3):284295.
  3. 3.
    Walia S, Kumar K. Digital image forgery detection: a systematic scrutiny. Austr J Forensic Sci. 2019;51(5):488526.
  4. 4.
    Scherhag U, Rathgeb C, Busch C. Face morphing attack detection methods. In: Rathgeb C, Tolosana R, Vera-Rodriguez R, Busch C, editors. Handbook of Digital Face Manipulation and Detection. Advances in Computer Vision and Pattern Recognition. Cham: Springer; 2022. doi:10.1007/978-3-030-87664-7_15.
  5. 5.
    Long M, Zhao X, Zhang L-B, Peng F. Detection of face morphing attacks based on patch-level features and lightweight networks. Secur Commun Netw. 2022;2022: 7460330. doi:10.1155/2022/7460330.
  6. 6.
    Fang M, Boutros F, Damer N. Unsupervised face morphing attack detection via self-paced anomaly detection. In: 2022 IEEE International Joint Conference on Biometrics (IJCB). Piscataway, NJ: IEEE; 2022. p. 111. doi:10.1109/IJCB54206.2022.10008003.
  7. 7.
    Jia C-K, Long M, Liu Y-C. Enhanced face morphing attack detection using error-level analysis and efficient selective kernel network. Comput Secur. 2024;137: 103640, doi:10.1016/j.cose.2023.103640.
  8. 8.
    Seibold C, Samek W, Hilsmann A, Eisert P. Accurate and robust neural networks for face morphing attack detection. J Inform Secur Appl. 2020;53: 102526, doi:10.1016/j.jisa.2020.102526.
  9. 9.
    Ramachandra R, Raja K, Busch C. Algorithmic fairness in face morphing attack detection. In: 2022 IEEE/CVF Winter Conference on Applications of Computer Vision Workshops (WACVW). Piscataway, NJ: IEEE; 2022. p. 410418. doi:10.1109/WACVW54805.2022.00047.
  10. 10.
    Venkatesh S, Ramachandra R, Raja K, Busch C. Face morphing attack generation and detection: a comprehensive survey. In: IEEE Transactions on Technology and Society. vol. 2, (3):Piscataway, NJ: IEEE; 2021. p. 128145. doi:10.1109/TTS.2021.3066254.
  11. 11.
    Zanardelli M, Guerrini F, Leonardi R, Image forgery detection: a survey of recent deep-learning approaches. Multimed Tools Appl. 2023;82: 1752117566. doi:10.1007/s11042-022-13797-w.
  12. 12.
    Hassan A, Sharma VK. Texture based image splicing forgery recognition using a passive approach. Int J Integr Eng. 2021;13(4):112121.
  13. 13.
    Srivastava V, Yadav SK. Digital image tampering detection using multilevel local binary pattern texture descriptor. J Appl Secur Res. 2021;17(1):6279.
  14. 14.
    Arun Anoop M, Karthikeyan P, Poonkuntran S. Unsupervised/supervised feature extraction and feature selection for multimedia data (feature extraction with feature selection for image forgery detection). In: Supervised and Unsupervised Data Engineering for Multimedia Data. John Wiley & Sons, Ltd; 2024. p. 2761.
  15. 15.
    Makandar A, Kaman S. Transform domain techniques combined with LBP for tampered image detection using machine learning. In: 2023 IEEE International Carnahan Conference on Security Technology (ICCST). Piscataway, NJ: IEEE; 2023. p. 17.
  16. 16.
    Sulaiman DS, Altaei MS. Image tampering detection using extreme learning machine. In: AIP Conference Proceedings. vol. 2457, (1):AIP Publishing; 2023.
  17. 17.
    Ahmed IT, Hammad BT, Jamil N. Forgery detection algorithm based on texture features. Indones J Electr Eng Comput Sci. 2021;24(1):226235.
  18. 18.
    Aydin Y. Comparison of color features on copy-move forgery detection problem using HSV color space. Austr J Forensic Sci. 2024;56(3):294310.
  19. 19.
    Makandar A, Javeriya SB, Kaman S. Spliced image forgery detection using adaptive over-segmentation combined with AKAZE, ORB, and SIFT feature descriptors. J Trans Syst Eng. 2023;1(3):140147.
  20. 20.
    Tahaoglu G, Ulutas G, Ustubioglu B, Nabiyev VV. Improved copy move forgery detection method via L a b color space and enhanced localization technique. Multimed Tools Appl. 2021;80: 2341923456.
  21. 21.
    Tsai YL, Leou JJ. Image copy-move forgery detection using color features and hierarchical feature point matching. In: Proceedings of the International Conference on Image Processing and Vision Engineering (IMPROVE). SciTePress; 2021. p. 153159. doi:10.5220/0010492301530159.
  22. 22.
    Sharma P, Kumar M, Sharma H. Comprehensive analyses of image forgery detection methods from traditional to deep learning approaches: an evaluation. Multimed Tools Appl. 2023;82(12):1811718150.
  23. 23.
    Heidari A, Jafari Navimipour N, Dag H, Unal M. Deepfake detection using deep learning methods: a systematic and comprehensive review. Wiley Interdiscip Rev: Data Mining Knowl Discov. 2024;14(2):e1520.
  24. 24.
    Khalil HA, Maged SA. Deepfakes creation and detection using deep learning. In: 2021 International Mobile, Intelligent, and Ubiquitous Computing Conference (MIUCC), Cairo, Egypt. Piscataway, NJ: IEEE; 2021. p. 14. doi:10.1109/MIUCC52538.2021.9447642.
  25. 25.
    Suratkar S, Kazi F. Deep fake video detection using transfer learning approach. Arab J Sci Eng. 2023;48: 97279737. doi:10.1007/s13369-022-07321-3.
  26. 26.
    Wodajo D, Atnafu S. Deepfake video detection using convolutional vision transformer [Internet]. arXiv; 2021. Available from: https://arxiv.org/abs/2102.11126.
  27. 27.
    Elhassan A, Al-Fawa’reh M, Jafar MT, Ababneh M, Jafar ST. DFT-MF: enhanced deepfake detection using mouth movement and transfer learning. SoftwareX. 2022;19: 101115.
  28. 28.
    Suratkar S, Kazi F, Sakhalkar M, Abhyankar N, Kshirsagar M. Exposing deepfakes using convolutional neural networks and transfer learning approaches. In: 2020 IEEE 17th India Council International Conference (INDICON). Piscataway, NJ: IEEE; 2020. p. 18.
  29. 29.
    Kong C, Luo A, Bao P, Li H, Wan R, Zheng Z, Rocha A, Kot AC. Open-set deepfake detection: a parameter-efficient adaptation method with forgery style mixture [Internet]. arXiv; 2024. Available from: https://arxiv.org/abs/2408.12791.
  30. 30.
    Lai Y, Yu Z, Yang J, Li B, Kang X, Shen L. Gm-df: generalized multi-scenario deepfake detection [Internet]. arXiv; 2024. Available from: https://arxiv.org/abs/2406.20078.
  31. 31.
    Yan Z, Luo Y, Lyu S, Liu Q, Wu B. Transcending forgery specificity with latent space augmentation for generalizable deepfake detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway, NJ: IEEE; 2024. p. 89848994.
  32. 32.
    Ba Z, Liu Q, Liu Z, Wu S, Lin F, Lu L, Ren K. Exposing the deception: uncovering more forgery clues for deepfake detection. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 38, (2):Washington, DC: AAAI Press; 2024. p. 719728.
  33. 33.
    Zhang D, Chen J, Liao X, Li F, Chen J, Yang G. Face forgery detection via multi-feature fusion and local enhancement. In: IEEE Transactions on Circuits and Systems for Video Technology. vol. 34, (9):Piscataway, NJ: IEEE; 2024. p. 89728977. doi:10.1109/TCSVT.2024.3390945.
  34. 34.
    Roboflow [Internet]; 2024. Available from: https://roboflow.com/.
  35. 35.
    Makandar A, Javeriya SB. Morphed faces in aadhar cards: a resource for understanding biometric security vulnerabilities. Mendeley Data. 2024;V1: doi:10.17632/k7srr5wwhv.1.
  36. 36.
    FaceSwapper [Internet]; 2024. Available from: https://faceswapper.ai/.
  37. 37.
    Dong J, Wang W, Tan T. CASIA image tampering detection evaluation database. In: Proceedings of the 2013 IEEE China Summit and International Conference on Signal and Information Processing. Piscataway, NJ: IEEE; 2013.
  38. 38.
    Kaman S, Makandar A. SDFVD: small-scale deepfake forgery video dataset. Mendeley Data. 2024;V1: doi:10.17632/bcmkfgct2s.1.
  39. 39.
    Pexels [Internet]; 2024. Available from: https://www.pexels.com/.
  40. 40.
    Remaker [Internet]; 2024. Available from: https://remaker.ai/en.
  41. 41.
    Alahmadi AA, Hussain M, Aboalsamh H, Muhammad G, Bebis G. Splicing image forgery detection based on DCT and local binary pattern. In: Global Conference on Signal and Information Processing (GlobalSIP). Piscataway, NJ: IEEE; 2013. p. 253256.
  42. 42.
    Ojala T, Pietikainen M, Maenpaa T. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans Pattern Anal Mach Intell. 2002;24(7):971987.
  43. 43.
    Azam KS, Riya FF, Al Mamun S, Kaiser MS. A novel approach of detecting image forgery using glcm and knn. In: 2021 International Conference on Information and Communication Technology for Sustainable Development (ICICT4SD). Piscataway, NJ: IEEE; 2021. p. 125129.
  44. 44.
    Zhou H, Shen Y, Zhu X, Liu B, Fu Z, Fan N. Digital image modification detection using color information and its histograms. Forensic Sci Int. 2016;266: 379388.
  45. 45.
    Zhang WW, Yang ZH, Niu SZ. Copy-move forgery detection algorithm based on HSV and CLAHE. J Beijing Univ Posts Telecommun. 2016;39(6):2732.
  46. 46.
    Sabeena M, Abraham L, Varghese A. Digital image forgery detection using local binary pattern (LBP) and Harlick transform with classification. In: 2021 IEEE International Power and Renewable Energy Conference (IPRECON). Piscataway, NJ: IEEE; 2021. p. 16. doi:10.1109/IPRECON52453.2021.9640976.
  47. 47.
    Apichardsilkij N. Basic Comparison Between RandomForest, SVM, and XGBoost [Internet]; 2024. Available from: https://medium.com/@ap.nattapoj_st/basic-comparison-between-randomforest-svm-and-xgboost-0e5862871175.
  48. 48.
    Kaya M, Sani KJ, Karakuş S. Copy-move forgery detection in digital forensic images using CNN. In: 2022 7th International Conference on Computer Science and Engineering (UBMK), Diyarbakir, Turkey. Piscataway, NJ: IEEE; 2022. p. 239245. doi:10.1109/UBMK55850.2022.9919560.
  49. 49.
    Mukherjee S. The Annotated ResNet-50 [Internet]; 2022. Available from: https://towardsdatascience.com/the-annotated-resnet-50-a6c536034758.
  50. 50.
    Remya Revi K, Wilscy M. Pretrained convolutional neural networks as feature extractor for image splicing detection. In: 2018 International Conference on Circuits and Systems in Digital Enterprise Technology (ICCSDET). Piscataway, NJ: IEEE; 2018. p. 15. doi:10.1109/ICCSDET.2018.8821242.
  51. 51.
    Dash AK. VGG16 architecture [Internet]; 2024. Available from: https://iq.opengenus.org/vgg16/.
  52. 52.
    Tan M, Quoc L. Efficientnet: rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning. vol. 97, PMLR; 2019. p. 61056114.
  53. 53.
    Ahmed T, Sabab NHN. Classification and understanding of cloud structures via satellite Images with EfficientUNet. SN Comput Sci. 2022;3: 99.
  54. 54.
    Makandar A, Kaman S, Javeriya SB. Enhancing Face recognition through dimensionality reduction techniques and diverse classifiers. LC Int J STEM. 2024;5(1):3644. doi:10.5281/zenodo.11109936.
  55. 55.
    Patel M, Rane K, Jain N, Mhatre P, Jaswal S. Image forgery detection using CNN. In: 2023 3rd International Conference on Intelligent Technologies (CONIT). 2023. p. 14. doi:10.1109/CONIT59222.2023.10205377.
  56. 56.
    Verma M. Binary Classification Using Convolution Neural Network (CNN) Model [Internet]; 2022. Available from: https://medium.com/@mayankverma05032001/binary-classification-using-convolution-neural-network-cnn-model-6e35cdf5bdbb.
  57. 57.
    Doegar A, Dutta M, Gaurav K. CNN based image forgery detection using pretrained AlexNet model [Internet]. Int J Comput Intel & IoT. 2019;2(1): 2019 Mar 19. Available at SSRN: https://ssrn.com/abstract=3355402.
  58. 58.
    Saikia P, Dholaria D, Yadav P, Patel V, Roy M. A hybrid CNN-LSTM model for video deepfake detection by leveraging optical flow features. In: 2022 International Joint Conference on Neural Networks (IJCNN). Piscataway, NJ: IEEE; 2022. p. 17.
  59. 59.
    Pokroy AA, Egorov AD. EfficientNets for deepfake detection: comparison of pretrained models. In: 2021 IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (ElConRus). Piscataway, NJ: IEEE; 2021. p. 598600.
  60. 60.
    Malolan B, Parekh A, Kazi F. Explainable deep-fake detection using visual interpretability methods. In: 3rd International Conference on Information and Computer Technologies (ICICT). Piscataway, NJ: IEEE; 2020. p. 289293.
  61. 61.
    Patel M, Gupta A, Tanwar S, Obaidat M. Trans-df: a transfer learning-based end-to-end deepfake detector. In: 2020 IEEE 5th International Conference on Computing Communication and Automation (ICCCA). Piscataway, NJ: IEEE; 2020. p. 796801.

Written by

Aziz Makandar, Shilpa Kaman and Syeda Bibi Javeriya

Article Type: Research Paper

Date of acceptance: November 2024

Date of publication: November 2024

DOI: 10.5772/acrt.20240042

Copyright: The Author(s), Licensee IntechOpen, License: CC BY 4.0

Download for free

© The Author(s) 2024. Licensee IntechOpen. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited.


Impact of this article

168

Downloads

234

Views

1

Altmetric Score


Share this article

Join us today!

Submit your Article