Mniejsze podróbki mogą być większym zagrożeniem
Businesses today are using conversational AI tools such as ChatGPT and Google Gemini to create an evolved form of deepfakes that could potentially rewrite a story within an image. This involves subtle alterations to gestures, props, and backgrounds, which not only misdirect human observers but also deceive AI detectors. This raises new concerns about the authenticity of visual representations shared online.
Deepfakes, despite being widely associated with direct misuse like political propaganda or explicit non-consensual content, have another side to them. History teaches us that the most influential manipulations usually come in the most subtle forms – a notion that rings true here. Think back to Stalin-era USSR when photo edits were used to erase certain individuals from history. Today’s AI-powered image generation echoes an unsettling similarity to that practice – making alterations to reality, one pixel at a time, rather than just swapping faces.
Unlike traditional deepfakes that project brute digital force, these subtle manipulations bear a psychological aspect akin to ‘gaslighting.’ They are more difficult to detect and possess the potential to gradually shift perceptions. Imagine, a missing stethoscope or a changed background. These minor adjustments may seem negligible, but together they pack the capacity to redefine narratives and disrupt identities.
In response to this emerging threat, researchers from Monash University and Curtin University have introduced MultiFakeVerse – a massive dataset constituted of over 845,000 images featuring menial, people-centered manipulations. Instead of focusing on face swaps or noticeable forgeries, this dataset leverages vision-language models like ChatGPT-4o and Gemini-2.0-Flash to produce context-sensitive edits. These alterations don’t switch the person in the picture but manipulate how they appear, by introducing emotional indications, modifying props, or tweaking the narrative tone of the scene.
Building this dataset involved starting with close to 87,000 authentic images from databases like EMOTIC, PISC, PIPA, and PIC 2.0 and generating over 758,000 manipulated versions. Every edit was meticulously guided by AI-generated instructions to subtly alter the perceived emotion or role of the subject. Gemini-2.0-Flash, one of the three tools used to execute the image edits, came out on top with consistent, artifact-free results.
However, the sobering fact remains that despite both humans and leading AI detection systems being subjected to the realism of this dataset, their ability to correctly identify fake images only weighed in at a mere 62% for human observers. AI models, including CnnSpot and SIDA, exhibited poor performance, especially in zero-shot scenarios, frequently misclassifying manipulated images as genuine. Furthermore, even after fine-tuning on MultiFakeVerse, the detection accuracy stayed modest, indicating that current detection frameworks, trained on more overt deepfakes, are ill-equipped to handle these nuanced edits.
The researchers utilized metrics such as PSNR, SSIM, and FID to assess the visual quality of their manipulations, with results indicating that the edits subtly shifted the meaning while preserving the integrity of the images. On procuring captions from ShareGPT-4V and analyzing them using Long-CLIP embeddings, it was confirmed that small changes, like altering a prop or facial expression, could significantly affect how a viewer interpreted the image.
Simultaneously, tests and studies involving participants were conducted to highlight the daunting challenge these manipulated images posed. Participants frequently misclassified images and seemed to struggle when asked to pinpoint manipulated regions. These findings reinforce just how effortlessly subtle deepfakes can evade both human intuition and AI scrutiny.
Leading Anti-fake detection systems were put to task in both zero-shot and fine-tuned modes using the dataset, which was split into training, validation, and test sets. Results gleaned showed that retraining helped when it came to localization of manipulated regions, although it still remained a prominent hurdle to overcome.
In conclusion, MultiFakeVerse uncovers a critical vulnerability in how we perceive and safeguard visual truth. As AI-generated content becomes increasingly integrated into our digital lives, it gets harder to distinguish the real from the fake. Although, subtle manipulations might not incite instantaneous uproar, their cumulative effect can potentially erode trust in visual media. It’s a new breed of deepfakes that don’t roar their presence, but whisper it, therein laying the danger.
First published Thursday, June 5, 2025. You can read the original article at [here](https://www.unite.ai/smaller-deepfakes-may-be-the-bigger-threat/).