The anatomy of a disinformation campaign
A powerful and persistent story has circulated across social media platforms for years: former U.S. President Donald Trump, in a clandestine act of heroism, intervened to save several Iranian women from execution. The posts are often accompanied by poignant, sorrowful images of the women, their faces a mixture of defiance and fear. The narrative is emotionally charged and politically potent. It is also a sophisticated lie, built on a foundation of truth.
While the brutal oppression of women and political dissidents in Iran is a well-documented human rights crisis, the visual “evidence” fueling this specific narrative is largely a product of artificial intelligence. Security researchers and open-source intelligence (OSINT) analysts have confirmed that many of the most-shared images are not photographs of real people but are instead deepfakes, generated by AI to serve a political agenda. This phenomenon represents a troubling convergence of a genuine tragedy and advanced digital forgery, designed to manipulate public perception.
A tale of two realities
To understand this campaign, one must separate fact from fiction. The factual basis is the dire situation in Iran. Organizations like Amnesty International continuously report on the Islamic Republic's use of capital punishment as a tool of political repression, particularly against women, ethnic minorities, and protesters involved in movements like “Woman, Life, Freedom.” Real women are arrested on baseless charges, subjected to sham trials, and face execution for demanding basic human rights. Their stories are harrowing and authentic.
The fiction is layered on top of this reality. The narrative of Trump’s intervention, which lacks any credible evidence from official government or journalistic sources, found fertile ground in pro-Trump and QAnon-affiliated online communities. It fits neatly within a broader mythology of Trump as a global savior battling a shadowy deep state. By co-opting the real suffering of Iranian women and attaching it to AI-generated faces, the campaign creators forge an emotional connection with their audience, making the political propaganda far more effective.
Unmasking the forgery: The technical details
The images used in this campaign are not simply edited photos; many are entirely synthetic, created from scratch by generative AI models. These models, such as Generative Adversarial Networks (GANs) or more recent diffusion models, are trained on vast datasets of real photographs. They learn the patterns, textures, and structures of human faces and can then generate new, unique, and photorealistic images.
While these tools are powerful, they are not yet perfect. Digital forensic analysis reveals several telltale signs of their artificial origin:
- Anatomical Inconsistencies: AI models notoriously struggle with complex anatomy, especially hands. Images often feature people with distorted fingers, an incorrect number of digits, or hands that blend unnaturally into other objects.
- Unnatural Details: Look closely at features like eyes, teeth, and jewelry. You might find that pupils are different shapes, teeth are perfectly uniform and slightly blurry, or earrings are asymmetrical in a way that defies physics.
- Flawed Textures: Skin may appear overly smooth and poreless, like a digital painting, or hair might look like a solid mass rather than individual strands.
- Bizarre Backgrounds: The AI's focus is the subject. Background elements are often a giveaway, appearing as a nonsensical blur of shapes, distorted text, or illogical architectural features.
These forgeries are disseminated through coordinated amplification networks on platforms like X (formerly Twitter), Facebook, and Telegram. A handful of accounts post the content, which is then rapidly shared by a swarm of bot and user accounts to create the illusion of a viral, grassroots movement. The emotionally resonant nature of the content ensures that well-meaning but unsuspecting users also contribute to its spread.
The ripple effect: Assessing the damage
The impact of this disinformation extends far beyond deceiving a few social media users. It inflicts several layers of harm.
First and foremost, it exploits and trivializes the very real struggle of Iranian political prisoners. Their life-or-death fight for freedom is reduced to a prop in a foreign political drama. This can undermine legitimate advocacy efforts by muddying the waters, making it harder for the public to distinguish real appeals for help from fabricated propaganda.
Second, it accelerates the erosion of public trust in visual media. As people become more aware of deepfakes, they may begin to doubt the authenticity of all images, including genuine photographs of atrocities. This phenomenon, known as the “liar’s dividend,” benefits malicious actors, who can dismiss real evidence of their wrongdoing as a “deepfake.”
Finally, such campaigns deepen societal polarization. They are designed to confirm existing biases and reinforce an “us-versus-them” worldview. By feeding a specific political base with what appears to be powerful evidence of their leader’s righteousness, it insulates them from factual reporting and makes constructive political discourse nearly impossible.
How to protect yourself from visual deception
Navigating an information environment polluted with AI-generated content requires a new level of digital literacy. Individuals can take several steps to avoid being manipulated.
- Pause and Question: The most powerful tool is skepticism. If an image or story evokes a strong emotional reaction, pause before sharing. Disinformation is engineered to bypass critical thinking by appealing directly to outrage, sympathy, or fear.
- Examine the Evidence: Look for the technical flaws mentioned above. Zoom in on the hands, the eyes, the background. Are there any details that seem unnatural or illogical?
- Trace the Source: Where did the image originate? Is it from a reputable news agency or a random, anonymous account? Use reverse image search tools like Google Images or TinEye to see where else the image has appeared. Often, this will lead to fact-checking articles that have already debunked it.
- Seek Corroboration: If a major world event, such as a U.S. president saving foreign nationals from execution, had actually occurred, it would be reported by multiple credible, international news organizations. The absence of such reporting is a significant red flag.
- Protect Your Privacy: When investigating sensitive political claims, it is wise to protect your own digital footprint from trackers and observers. Using tools for encryption and anonymity, like a trusted VPN, can help secure your connection and maintain your privacy.
Ultimately, the story of the AI-generated Iranian women is a case study in modern propaganda. It demonstrates how genuine human suffering can be weaponized in information warfare, and it underscores the growing need for a discerning public capable of separating digital truth from sophisticated fiction.




