A Robust Approach for Deepfake Detection Using SWIN Transformer

Read the full article See related articles

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

The widespread use of deepfake technology in recent years has made it extremely difficult to differentiate between real and fake images, usually AI-generated images. Effective detection techniques are desperately needed because one can generate fake images and spread them with ease. In response to the situation, this research paper explores the effectiveness of employing SWIN Transformer, a cutting-edge transformer-based architecture, for deep fake image detection. The foundation of the suggested detection framework is an architecture made up of bottleneck, encoder, and decoder parts which is a type of SWIN transformer. It uses various self-attention mechanisms and advanced features to analyse the images closely whether it is a real image or a deepfake one. It relies on the concept of shifted windows during the processing of the images and is considered more effective than the traditional CNN methods. Our test results show how well the SWIN Transformer-based method performs in precisely recognizing deep fake images. The accuracy is found to be 97.91% for CelebDF dataset and 95.715% for FF++ dataset. The AUC for the newly modelled SWIN transformer is 0.99 and 0.9625 for CelebDF and FF++ datasets respectively. The Log Loss has been calculated to be 0.034 for CelebDF dataset and 0.1573 for FF++ dataset. The proposed methodology not only enhances the accuracy of detecting manipulated images but also offers potential for scalable and efficient deployment in real-world scenarios where the proliferation of deepfakes presents significant challenges to maintaining trust and authenticity in visual media.

Article activity feed