Section 1: The Generative AI Revolution in Digital Media
1.1 Introduction
The advent of sophisticated generative artificial intelligence (AI) marks a paradigm shift in the creation, consumption, and verification of digital media. Technologies capable of producing hyper-realistic images, videos, and audio—collectively termed synthetic media—have moved from the realm of academic research into the hands of the general public, heralding an era of unprecedented creative potential and profound societal risk. These generative models, powered by deep learning architectures, represent a potent dual-use technology. On one hand, they offer transformative tools for industries ranging from entertainment and healthcare to education, promising to automate complex tasks, personalize user experiences, and unlock new frontiers of artistic expression.1 On the other hand, the same capabilities can be weaponized to generate deceptive content at an unprecedented scale, enabling sophisticated financial fraud, political disinformation campaigns, and egregious violations of personal privacy.4
This report presents a comprehensive investigation into the multifaceted landscape of AI-generated media. It posits that the rapid proliferation of synthetic content creates a series of complex, interconnected challenges that cannot be addressed by any single solution. The central thesis of this analysis is that navigating the era of synthetic media requires a multi-faceted and integrated approach. This approach must combine continued technological innovation in both generation and detection, the development of robust and adaptive legal frameworks, a re-evaluation of platform responsibility, and a foundational commitment to fostering widespread digital literacy. The co-evolution of generative models and the tools designed to detect them has initiated a persistent technological “arms race,” a dynamic that underscores the futility of a purely technological solution and highlights the urgent need for a holistic, societal response.7
1.2 Scope and Structure
This report is structured to provide a systematic and in-depth analysis of AI-generated media. It begins by establishing the technical underpinnings of the technology before exploring its real-world implications and the societal responses it has engendered.
Section 2: The Technological Foundations of Synthetic Media provides a detailed technical examination of the core generative models. It deconstructs the architectures of Generative Adversarial Networks (GANs), diffusion models, the autoencoder-based systems used for deepfake video, and the neural networks enabling voice synthesis.
Section 3: The Dual-Use Dilemma: Applications of Generative AI explores the dichotomy of these technologies. It first examines their benevolent implementations in fields such as entertainment, healthcare, and education, before detailing their malicious weaponization for financial fraud, political disinformation, and the creation of non-consensual explicit material.
Section 4: Ethical and Societal Fault Lines moves beyond specific applications to analyze the deeper, systemic ethical challenges. This section investigates issues of algorithmic bias, the erosion of epistemic trust and shared reality, unresolved intellectual property disputes, and the profound psychological harm inflicted upon victims of deepfake abuse.
Section 5: The Counter-Offensive: Detecting AI-Generated Content details the technological and strategic responses designed to identify synthetic media. It covers both passive detection methods, which search for digital artifacts, and proactive approaches, such as digital watermarking and the C2PA standard, which embed provenance at the point of creation. This section also analyzes the adversarial “cat-and-mouse” game between content generators and detectors.
Section 6: Navigating the New Reality: Legal Frameworks and Future Directions concludes the report by examining the emerging landscape of regulation and policy. It provides a comparative analysis of global legislative efforts, discusses the role of platform policies, and offers a set of integrated recommendations for a path forward, emphasizing the critical role of public education as the ultimate defense against deception.
Section 2: The Technological Foundations of Synthetic Media
The capacity to generate convincing synthetic media is rooted in a series of breakthroughs in deep learning. This section provides a technical analysis of the primary model architectures that power the creation of AI-generated images, videos, and voice, forming the foundation for understanding both their capabilities and their limitations.
2.1 Image Generation I: Generative Adversarial Networks (GANs)
Generative Adversarial Networks (GANs) were a foundational breakthrough in generative AI, introducing a novel training paradigm that pits two neural networks against each other in a competitive game.11 This adversarial process enables the generation of highly realistic data samples, particularly images.
The core mechanism of a GAN involves two distinct networks:
- The Generator: This network’s objective is to create synthetic data. It takes a random noise vector as input and, through a series of learned transformations, attempts to produce an output (e.g., an image) that is indistinguishable from real data from the training set. The generator’s goal is to effectively “fool” the second network.11
- The Discriminator: This network acts as a classifier. It is trained on a dataset of real examples and is tasked with evaluating inputs to determine whether they are authentic (from the real dataset) or synthetic (from the generator). It outputs a probability score, typically between 0 (fake) and 1 (real).12
The training process is an iterative, zero-sum game. The generator and discriminator are trained simultaneously. The generator’s loss function is designed to maximize the discriminator’s error, while the discriminator’s loss function is designed to minimize its own error. Through backpropagation, the feedback from the discriminator’s evaluation is used to update the generator’s parameters, allowing it to improve its ability to create convincing fakes. Concurrently, the discriminator learns from its mistakes, becoming better at identifying the generator’s outputs. This cycle continues until an equilibrium is reached, a point at which the generator’s outputs are so realistic that the discriminator’s classifications are no better than random chance.11
Several types of GANs have been developed for specific applications. Vanilla GANs represent the basic architecture, while Conditional GANs (cGANs) introduce additional information (such as class labels or text descriptions) to both the generator and discriminator, allowing for more controlled and targeted data generation.11
StyleGANs are designed for producing extremely high-resolution, photorealistic images by controlling different levels of detail at various layers of the generator network.12
CycleGANs are used for image-to-image translation without paired training data, such as converting a photograph into the style of a famous painter.12
2.2 Image Generation II: Diffusion Models
While GANs were revolutionary, they are often difficult to train and can suffer from instability. In recent years, diffusion models have emerged as a dominant and more stable alternative, powering many state-of-the-art text-to-image systems like Stable Diffusion, DALL-E 2, and Midjourney.7 Inspired by principles from non-equilibrium thermodynamics, these models generate high-quality data by learning to reverse a process of gradual noising.14
The mechanism of a diffusion model consists of two primary phases:
- Forward Diffusion Process (Noising): This is a fixed process, formulated as a Markov chain, where a small amount of Gaussian noise is incrementally added to a clean image over a series of discrete timesteps (t=1,2,…,T). At each step, the image becomes slightly noisier, until, after a sufficient number of steps (T), the image is transformed into pure, unstructured isotropic Gaussian noise. This process does not involve machine learning; it is a predefined procedure for data degradation.14
- Reverse Diffusion Process (Denoising): This is the learned, generative part of the model. A neural network, typically a U-Net architecture, is trained to reverse the forward process. It takes a noisy image at a given timestep t as input and is trained to predict the noise that was added to the image at that step. By subtracting this predicted noise, the model can produce a slightly cleaner image corresponding to timestep t−1. This process is repeated iteratively, starting from a sample of pure random noise (xT), until a clean, coherent image (x0) is generated.14
The technical process is governed by a variance schedule, denoted by βt, which controls the amount of noise added at each step of the forward process. The model’s training objective is to minimize the difference—typically the mean-squared error—between the noise it predicts and the actual noise that was added at each timestep. By learning to accurately predict the noise at every level of degradation, the model implicitly learns the underlying structure and patterns of the original data distribution.14 This shift from the unstable adversarial training of GANs to the more predictable, step-wise denoising of diffusion models represents a critical inflection point. It has made the generation of high-fidelity synthetic media more reliable and scalable, democratizing access to powerful creative tools and, consequently, lowering the barrier to entry for both benevolent and malicious actors.
2.3 Video Generation: The Architecture of Deepfakes
Deepfake video generation, particularly face-swapping, primarily relies on a type of neural network known as an autoencoder. An autoencoder is composed of two parts: an encoder, which compresses an input image into a low-dimensional latent representation that captures its core features (like facial expression and orientation), and a decoder, which reconstructs the original image from this latent code.16
To perform a face swap, two autoencoders are trained. One is trained on images of the source person (Person A), and the other on images of the target person (Person B). Crucially, both autoencoders share the same encoder but have separate decoders. The shared encoder learns to extract universal facial features that are independent of identity. After training, video frames of Person A are fed into the shared encoder. The resulting latent code, which captures Person A’s expressions and pose, is then passed to the decoder trained on Person B. This decoder reconstructs the face using the identity of Person B but with the expressions and movements of Person A, resulting in a face-swapped video.16
To improve the realism and overcome common artifacts, this process is often enhanced with a GAN architecture. In this setup, the decoder acts as the generator, and a separate discriminator network is trained to distinguish between the generated face-swapped images and real images of the target person. This adversarial training compels the decoder to produce more convincing outputs, reducing visual inconsistencies and making the final deepfake more difficult to detect.13
2.4 Voice Synthesis and Cloning
AI voice synthesis, or voice cloning, creates a synthetic replica of a person’s voice capable of articulating new speech from text input. The process typically involves three stages:
- Data Collection: A sample of the target individual’s voice is recorded.
- Model Training: A deep learning model is trained on this audio data. The model analyzes the unique acoustic characteristics of the voice, including its pitch, tone, cadence, accent, and emotional inflections.17
- Synthesis: Once trained, the model can take text as input and generate new audio that mimics the learned vocal characteristics, effectively speaking the text in the target’s voice.17
A critical technical detail that has profound societal implications is the minimal amount of data required for this process. Research and real-world incidents have demonstrated that as little as three seconds of audio can be sufficient for an AI tool to produce a convincing voice clone.20 This remarkably low data requirement is the single most important technical factor enabling the widespread proliferation of voice-based fraud. It means that virtually anyone with a public-facing role, a social media presence, or even a recorded voicemail message has provided enough raw material to be impersonated. This transforms voice cloning from a niche technological capability into a practical and highly scalable tool for social engineering, directly enabling the types of sophisticated financial scams detailed later in this report.
| Table 1: Comparison of Generative Models (GANs vs. Diffusion Models) | |
| Attribute | Generative Adversarial Networks (GANs) |
| Core Mechanism | An adversarial “game” between a Generator (creates data) and a Discriminator (evaluates data).11 |
| Training Stability | Often unstable and difficult to train, prone to issues like mode collapse where the generator produces limited variety.12 |
| Output Quality | Can produce very high-quality, sharp images but may struggle with overall diversity and coherence.12 |
| Computational Cost | Training can be computationally expensive due to the dual-network architecture. Inference (generation) is typically fast.11 |
| Key Applications | High-resolution face generation (StyleGAN), image-to-image translation (CycleGAN), data augmentation.11 |
| Prominent Examples | StyleGAN, CycleGAN, BigGAN |
Section 3: The Dual-Use Dilemma: Applications of Generative AI
Generative AI technologies are fundamentally dual-use, possessing an immense capacity for both societal benefit and malicious harm. Their application is not inherently benevolent or malevolent; rather, the context and intent of the user determine the outcome. This section explores this dichotomy, first by examining the transformative and positive implementations across various sectors, and second by detailing the weaponization of these same technologies for deception, fraud, and abuse.
3.1 Benevolent Implementations: Augmenting Human Potential
In numerous fields, generative AI is being deployed as a powerful tool to augment human creativity, accelerate research, and improve accessibility.
Transforming Media and Entertainment:
The creative industries have been among the earliest and most enthusiastic adopters of generative AI. The technology is automating tedious and labor-intensive tasks, reducing production costs, and opening new avenues for artistic expression.
- Visual Effects (VFX) and Post-Production: AI is revolutionizing VFX workflows. Machine learning models have been used to de-age actors with remarkable realism, as seen with Harrison Ford in Indiana Jones and the Dial of Destiny.21 In the Oscar-winning film
Everything Everywhere All At Once, AI tools were used for complex background removal, reducing weeks of manual rotoscoping work to mere hours.21 Furthermore, AI can upscale old or low-resolution archival footage to modern high-definition standards, preserving cultural heritage and making it accessible to new audiences. - Audio Production: In music, AI has enabled remarkable feats of audio restoration. The 2023 release of The Beatles’ song “Now and Then” was made possible by an AI model that isolated John Lennon’s vocals from a decades-old, low-quality cassette demo, allowing the surviving band members to complete the track.21 AI-powered tools also provide advanced noise reduction and audio enhancement, cleaning up dialogue tracks and saving productions from costly reshoots.
- Content Creation and Personalization: Generative models are used for rapid prototyping in pre-production, generating concept art, storyboards, and character designs from simple text prompts.1 Streaming services and media companies also leverage AI to analyze vast datasets of viewer preferences, enabling them to generate personalized content recommendations and even inform decisions about which new projects to greenlight.23
Advancing Healthcare and Scientific Research:
One of the most promising applications of generative AI is in the creation of synthetic data, particularly in healthcare. This addresses a fundamental challenge in medical research: the need for large, diverse datasets is often at odds with strict patient privacy regulations like HIPAA and GDPR.
- Privacy-Preserving Data: Generative models can be trained on real patient data to learn its statistical properties. They can then generate entirely new, artificial datasets that mimic the characteristics of the real data without containing any personally identifiable information.3 This synthetic data acts as a high-fidelity, privacy-preserving proxy.
- Accelerating Research: This approach allows researchers to train and validate AI models for tasks like rare disease detection, where real-world data is scarce. It also enables the simulation of clinical trials, the reduction of inherent biases in existing datasets by generating more balanced data, and the facilitation of secure, collaborative research across different institutions without the risk of exposing sensitive patient records.3
Innovating Education and Accessibility:
Generative AI is being used to create more personalized, engaging, and inclusive learning environments.
- Personalized Learning: AI can function as a personal tutor, generating customized lesson plans, interactive simulations, and unlimited practice problems that adapt to an individual student’s pace and learning style.2
- Assistive Technologies: For individuals with disabilities, AI-powered tools are a gateway to greater accessibility. These include advanced speech-to-text services that provide real-time transcriptions for the hearing-impaired, sophisticated text-to-speech readers that assist those with visual impairments or reading disabilities, and generative tools that help individuals with executive functioning challenges by breaking down complex tasks into manageable steps.2
This analysis reveals a profound paradox inherent in generative AI. The same technological principles that enable the creation of synthetic health data to protect patient privacy are also used to generate non-consensual deepfake pornography, one of the most severe violations of personal privacy imaginable. The technology itself is ethically neutral; its application within a specific context determines whether it serves as a shield for privacy or a weapon against it. This complicates any attempt at broad-stroke regulation, suggesting that policy must be highly nuanced and application-specific.
3.2 Malicious Weaponization: The Architecture of Deception
The same attributes that make generative AI a powerful creative tool—its accessibility, scalability, and realism—also make it a formidable weapon for malicious actors.
Financial Fraud and Social Engineering:
AI voice cloning has emerged as a particularly potent tool for financial crime. By replicating a person’s voice with high fidelity, scammers can bypass the natural skepticism of their targets, exploiting psychological principles of authority and urgency.27
- Case Studies: A series of high-profile incidents have demonstrated the devastating potential of this technique. In 2019, criminals used a cloned voice of a UK energy firm’s CEO to trick a director into transferring $243,000.28 In 2020, a similar scam involving a cloned director’s voice resulted in a $35 million loss.29 In 2024, a multi-faceted attack in Hong Kong used a deepfaked CFO in a video conference, leading to a fraudulent transfer of $25 million.28
- Prevalence and Impact: These are not isolated incidents. Surveys indicate a dramatic rise in deepfake-related fraud. One study found that one in four people had experienced or knew someone who had experienced an AI voice scam, with 77% of victims reporting a financial loss.20 The ease of access to voice cloning tools and the minimal data required to create a clone have made this a scalable and effective form of attack.30
Political Disinformation and Propaganda:
Generative AI enables the creation and dissemination of highly convincing disinformation designed to manipulate public opinion, sow social discord, and interfere in democratic processes.
- Tactics: Malicious actors have used generative AI to create fake audio of political candidates appearing to discuss election rigging, deployed AI-cloned voices in robocalls to discourage voting, as seen in the 2024 New Hampshire primary, and fabricated videos of world leaders to spread false narratives during geopolitical conflicts.5
- Scale and Believability: AI significantly lowers the resource and skill threshold for producing sophisticated propaganda. It allows foreign adversaries to overcome language and cultural barriers that previously made their influence operations easier to detect, enabling them to create more persuasive and targeted content at scale.5
The Weaponization of Intimacy: Non-Consensual Deepfake Pornography:
Perhaps the most widespread and unequivocally harmful application of generative AI is the creation and distribution of non-consensual deepfake pornography.
- Statistics: Multiple analyses have concluded that an overwhelming majority—estimated between 90% and 98%—of all deepfake videos online are non-consensual pornography, and the victims are almost exclusively women.36
- Nature of the Harm: This practice constitutes a severe form of image-based sexual abuse and digital violence. It inflicts profound and lasting psychological trauma on victims, including anxiety, depression, and a shattered sense of safety and identity. It is used as a tool for harassment, extortion, and reputational ruin, exacerbating existing gender inequalities and making digital spaces hostile and unsafe for women.38 While many states and countries are moving to criminalize this activity, legal frameworks and enforcement mechanisms are struggling to keep pace with the technology’s proliferation.6
The applications of generative AI reveal an asymmetry of harm. While benevolent uses primarily create economic and social value—such as increased efficiency in film production or new avenues for medical research—malicious applications primarily destroy foundational societal goods, including personal safety, financial security, democratic integrity, and epistemic trust. This imbalance suggests that the negative externalities of misuse may far outweigh the positive externalities of benevolent use, presenting a formidable challenge for policymakers attempting to foster innovation while mitigating catastrophic risk.
| Table 2: Case Studies in AI-Driven Financial Fraud | ||||
| Case / Year | Technology Used | Method of Deception | Financial Loss (USD) | Source(s) |
| Hong Kong Multinational, 2024 | Deepfake Video & Voice | Impersonation of CFO and other employees in a multi-person video conference to authorize transfers. | $25 Million | 28 |
| Unnamed Company, 2020 | AI Voice Cloning | Impersonation of a company director’s voice over the phone to confirm fraudulent transfers. | $35 Million | 29 |
| UK Energy Firm, 2019 | AI Voice Cloning | Impersonation of the parent company’s CEO voice to demand an urgent fund transfer. | $243,000 | 28 |
Section 4: Ethical and Societal Fault Lines
The proliferation of generative AI extends beyond its direct applications to expose and exacerbate deep-seated ethical and societal challenges. These issues are not merely side effects but are fundamental consequences of deploying powerful, data-driven systems into complex human societies. This section analyzes the systemic fault lines of algorithmic bias, the erosion of shared reality, unresolved intellectual property conflicts, and the profound human cost of AI-enabled abuse.
4.1 Algorithmic Bias and Representation
Generative AI models, despite their sophistication, are not objective. They are products of the data on which they are trained, and they inherit, reflect, and often amplify the biases present in that data.
- Sources of Bias: Bias is introduced at multiple stages of the AI development pipeline. It begins with data collection, where training datasets may not be representative of the real-world population, often over-representing dominant demographic groups. It continues during data labeling, where human annotators may embed their own subjective or cultural biases into the labels. Finally, bias can be encoded during model training, where the algorithm learns and reinforces historical prejudices present in the data.42
- Manifestations of Bias: The consequences of this bias are evident across all modalities of generative AI. Facial recognition systems have been shown to be less accurate for women and individuals with darker skin tones.44 AI-driven hiring tools have been found to favor male candidates for technical roles based on historical hiring patterns.45 Text-to-image models, when prompted with neutral terms like “doctor” or “CEO,” disproportionately generate images of white men, while prompts for “nurse” or “homemaker” yield images of women, thereby reinforcing harmful gender and racial stereotypes.42
- The Amplification Feedback Loop: A particularly pernicious aspect of algorithmic bias is the creation of a societal feedback loop. When a biased AI system generates stereotyped content, it is consumed by users. This exposure can reinforce their own pre-existing biases, which in turn influences the future data they create and share online. This new, biased data is then scraped and used to train the next generation of AI models, creating a cycle where societal biases and algorithmic biases mutually reinforce and amplify each other.45
4.2 The Epistemic Crisis: Erosion of Trust and Shared Reality
The ability of generative AI to create convincing, fabricated content at scale poses a fundamental threat to our collective ability to distinguish truth from fiction, creating an epistemic crisis.
- Undermining Trust in Media: As the public becomes increasingly aware that any image, video, or audio clip could be a sophisticated fabrication, a general skepticism toward all digital media takes root. This erodes trust not only in individual pieces of content but in the institutions of journalism and public information as a whole. Studies have shown that even the mere disclosure of AI’s involvement in news production, regardless of its specific role, can lower readers’ perception of credibility.35
- The Liar’s Dividend: The erosion of trust produces a dangerous second-order effect known as the “liar’s dividend.” The primary, or first-order, threat of deepfakes is that people will believe fake content is real. The liar’s dividend is the inverse and perhaps more insidious threat: that people will dismiss real content as fake. As public awareness of deepfake technology grows, it becomes a plausible defense for any malicious actor caught in a genuinely incriminating audio or video recording to simply claim the evidence is an AI-generated fabrication. This tactic undermines the very concept of verifiable evidence, which is a cornerstone of democratic accountability, journalism, and the legal system.35
- Impact on Democracy: A healthy democracy depends on a shared factual basis for public discourse and debate. By flooding the information ecosystem with synthetic content and providing a pretext to deny objective reality, generative AI pollutes this shared space. It exacerbates political polarization, as individuals retreat into partisan information bubbles, and corrodes the social trust necessary for democratic governance to function.35
4.3 Intellectual Property in the Age of AI
The development and deployment of generative AI have created a legal and ethical quagmire around intellectual property (IP), challenging long-standing principles of copyright law.
- Training Data and Fair Use: The dominant paradigm for training large-scale generative models involves scraping and ingesting massive datasets from the public internet, a process that inevitably includes vast quantities of copyrighted material. AI developers typically argue that this constitutes “fair use” under U.S. copyright law, as the purpose is transformative (training a model rather than reproducing the work). Copyright holders, however, contend that this is mass-scale, uncompensated infringement. Recent court rulings on this matter have been conflicting, creating a profound legal uncertainty that hangs over the entire industry.48 This unresolved legal status of training data creates a foundational instability for the generative AI ecosystem. If legal precedent ultimately rules against fair use, it could retroactively invalidate the training processes of most major models, exposing developers to enormous liability and potentially forcing a fundamental re-architecture of the industry.
- Authorship and Ownership of Outputs: A core tenet of U.S. copyright law is the requirement of a human author. The U.S. Copyright Office has consistently reinforced this position, denying copyright protection to works generated “autonomously” by AI systems. It argues that for a work to be copyrightable, a human must exercise sufficient creative control over its expressive elements. Simply providing a text prompt to an AI model is generally considered insufficient to meet this standard.48 This raises complex questions about the copyrightability of works created with significant AI assistance and where the line of “creative control” is drawn.
- Confidentiality and Trade Secrets: The use of public-facing generative AI tools poses a significant risk to confidential information. When users include proprietary data or trade secrets in their prompts, that information may be ingested by the AI provider, used for future model training, and potentially surface in the outputs generated for other users, leading to an inadvertent loss of confidentiality.49
4.4 The Human Cost: Psychological Impact of Deepfake Abuse
Beyond the systemic challenges, the misuse of generative AI inflicts direct, severe, and lasting harm on individuals, particularly through the creation and dissemination of non-consensual deepfake pornography.
- Victim Trauma: This form of image-based sexual abuse causes profound psychological trauma. Victims report experiencing humiliation, shame, anxiety, powerlessness, and emotional distress comparable to that of victims of physical sexual assault. The harm is compounded by the viral nature of digital content, as the trauma is re-inflicted each time the material is viewed or shared.37
- A Tool of Gendered Violence: The overwhelming majority of deepfake pornography victims are women. This is not a coincidence; it reflects the weaponization of this technology as a tool of misogyny, harassment, and control. It is used to silence women, damage their reputations, and reinforce patriarchal power dynamics, contributing to an online environment that is hostile and unsafe for women and girls.37
- Barriers to Help-Seeking: Victims, especially minors, often face significant barriers to reporting the abuse. These include intense feelings of shame and self-blame, as well as a legitimate fear of not being believed by parents, peers, or authorities. The perception that the content is “fake” can lead others to downplay the severity of the harm, further isolating the victim and discouraging them from seeking help.38
Section 5: The Counter-Offensive: Detecting AI-Generated Content
In response to the threats posed by malicious synthetic media, a field of research and development has emerged focused on detection and verification. These efforts can be broadly categorized into two approaches: passive detection, which analyzes content for tell-tale signs of artificiality, and proactive detection, which embeds verifiable information into content at its source. These approaches are locked in a continuous adversarial arms race with the generative models they seek to identify.
5.1 Passive Detection: Unmasking the Artifacts
Passive detection methods operate on the finished media file, seeking intrinsic artifacts and inconsistencies that betray its synthetic origin. These techniques require no prior information or embedded signals and function like digital forensics, examining the evidence left behind by the generation process.51
- Visual Inconsistencies: Early deepfakes were often riddled with obvious visual flaws, and while generative models have improved dramatically, subtle inconsistencies can still be found through careful analysis.
- Anatomical and Physical Flaws: AI models can struggle with the complex physics and biology of the real world. This can manifest as unnatural or inconsistent blinking patterns, stiff facial expressions that lack micro-expressions, and flawed rendering of complex details like hair strands or the anatomical structure of hands.54 The physics of light can also be a giveaway, with models producing inconsistent shadows, impossible reflections, or lighting on a subject that does not match its environment.54
- Geometric and Perspective Anomalies: AI models often assemble scenes from learned patterns without a true understanding of three-dimensional space. This can lead to violations of perspective, such as parallel lines on a single building converging to multiple different vanishing points, a physical impossibility.57
- Auditory Inconsistencies: AI-generated voice, while convincing, can lack the subtle biometric markers of authentic human speech. Detection systems analyze these acoustic properties to identify fakes.
- Biometric Voice Analysis: These systems scrutinize the nuances of speech, such as tone, pitch, rhythm, and vocal tract characteristics. Synthetic voices may exhibit unnatural pitch variations, a lack of “liveness” (the subtle background noise and imperfections of a live recording), or time-based anomalies that deviate from human speech patterns.59 Robotic inflection or a lack of natural breathing and hesitation can also be indicators.57
- Statistical and Digital Fingerprints: Beyond what is visible or audible, synthetic media often contains underlying statistical irregularities. Detection models can be trained to identify these digital fingerprints, which can include unnatural pixel correlations, unique frequency domain artifacts, or compression patterns that are characteristic of a specific generative model rather than a physical camera sensor.55
5.2 Proactive Detection: Embedding Provenance
In contrast to passive analysis, proactive methods aim to build a verifiable chain of custody for digital media from the moment of its creation.
- Digital Watermarking (SynthID): This approach, exemplified by Google’s SynthID, involves embedding a digital watermark directly into the content’s data during the generation process. For an image, this means altering pixel values in a way that is imperceptible to the human eye but can be algorithmically detected by a corresponding tool. The presence of this watermark serves as a definitive indicator that the content was generated by a specific AI system.63
- The C2PA Standard and Content Credentials: A more comprehensive proactive approach is championed by the Coalition for Content Provenance and Authenticity (C2PA). The C2PA has developed an open technical standard for attaching secure, tamper-evident metadata to media files, known as Content Credentials. This system functions like a “nutrition label” for digital content, cryptographically signing a manifest of information about the asset’s origin (e.g., the camera model or AI tool used), creator, and subsequent edit history. This creates a verifiable chain of provenance that allows consumers to inspect the history of a piece of media and see if it has been altered. Major technology companies and camera manufacturers are beginning to adopt this standard.64
5.3 The Adversarial Arms Race
The relationship between generative models and detection systems is not static; it is a dynamic and continuous “cat-and-mouse” game.7
- Co-evolution: As detection models become proficient at identifying specific artifacts (e.g., unnatural blinking), developers of generative models train new versions that explicitly learn to avoid creating those artifacts. This co-evolutionary cycle means that passive detection methods are in a constant race to keep up with the ever-improving realism of generative AI.8
- Adversarial Attacks: A more direct threat to detection systems comes from adversarial attacks. In this scenario, a malicious actor intentionally adds small, carefully crafted, and often imperceptible perturbations to a deepfake. These perturbations are not random; they are specifically optimized to exploit vulnerabilities in a detection model’s architecture, causing it to misclassify a fake piece of content as authentic. The existence of such attacks demonstrates that even highly accurate detectors can be deliberately deceived, undermining their reliability.71
This adversarial dynamic reveals an inherent asymmetry that favors the attacker. A creator of malicious content only needs their deepfake to succeed once—to fool a single detection system or a single influential individual—for it to spread widely and cause harm. In contrast, defenders—such as social media platforms and detection tool providers—must succeed consistently to be effective. Given that generative models are constantly evolving to eliminate the very artifacts that passive detectors rely on, and that adversarial attacks can actively break detection models, it becomes clear that relying solely on a technological “fix” for detection is an unsustainable long-term strategy. The solution space must therefore expand beyond technology to encompass the legal, educational, and social frameworks discussed in the final section of this report.
| Table 3: Typology of Passive Detection Artifacts Across Modalities | ||
| Modality | Category of Artifact | Specific Example(s) |
| Image / Video | Physical / Anatomical | Unnatural or lack of blinking; Stiff facial expressions; Flawed rendering of hair, teeth, or hands; Airbrushed skin lacking pores or texture.54 |
| Geometric / Physics-Based | Inconsistent lighting and shadows that violate the physics of a single light source; Impossible reflections; Inconsistent vanishing points in architecture.54 | |
| Behavioral | Unnatural crowd uniformity (everyone looks the same or in the same direction); Facial expressions that do not match the context of the event.57 | |
| Digital Fingerprints | Unnatural pixel patterns or noise; Compression artifacts inconsistent with camera capture; Resolution inconsistencies between different parts of an image.55 | |
| Audio | Biometric / Acoustic | Unnatural pitch, tone, or rhythm; Lack of “liveness” (e.g., absence of subtle background noise or breath sounds); Robotic or monotonic inflection.57 |
| Linguistic | Flawless pronunciation without natural hesitations; Use of uncharacteristic phrases or terminology; Unnatural pacing or cadence.57 |
Section 6: Navigating the New Reality: Legal Frameworks and Future Directions
The rapid integration of generative AI into the digital ecosystem has prompted a global response from policymakers, technology companies, and civil society. The challenges posed by synthetic media are not merely technical; they are deeply intertwined with legal principles, platform governance, and public trust. This final section examines the emerging regulatory landscape, the role of platform policies, and proposes a holistic strategy for navigating this new reality.
6.1 Global Regulatory Responses
Governments worldwide are beginning to grapple with the need to regulate AI and deepfake technology, though their approaches vary significantly, reflecting different legal traditions and political priorities.
- A Comparative Analysis of Regulatory Models:
- The European Union: A Risk-Based Framework. The EU has taken a comprehensive approach with its AI Act, which classifies AI systems based on their potential risk to society. Under this framework, generative AI systems are subject to specific transparency obligations. Crucially, the act mandates that AI-generated content, such as deepfakes, must be clearly labeled as such, empowering users to know when they are interacting with synthetic media.75
- The United States: A Harm-Specific Approach. The U.S. has pursued a more targeted, sector-specific legislative strategy. A prominent example is the TAKE IT DOWN Act, which focuses directly on the harm caused by non-consensual intimate imagery. This bipartisan law makes it illegal to create or share such content, including AI-generated deepfakes, and imposes a 48-hour takedown requirement on online platforms that receive a report from a victim. This approach prioritizes addressing specific, demonstrable harms over broad, preemptive regulation of the technology itself.6
- China: A State-Control Model. China’s regulatory approach is characterized by a focus on maintaining state control over the information ecosystem. Its regulations require that all AI-generated content be conspicuously labeled and traceable to its source. The rules also explicitly prohibit the use of generative AI to create and disseminate “fake news” or content that undermines national security and social stability, reflecting a top-down approach to managing the technology’s societal impact.75
- Emerging Regulatory Themes: Despite these different models, a set of common themes is emerging in the global regulatory discourse. These include a strong emphasis on transparency (through labeling and disclosure), the importance of consent (particularly regarding the use of an individual’s likeness), and the principle of platform accountability for harmful content distributed on their services.75
6.2 Platform Policies and Content Moderation
In parallel with government regulation, major technology and social media platforms are developing their own internal policies to govern the use of generative AI.
- Industry Self-Regulation: Platforms like Meta, TikTok, and Google have begun implementing policies that require users to label realistic AI-generated content. They are also developing their own automated tools to detect and flag synthetic media that violates their terms of service, which often prohibit deceptive or harmful content like spam, hate speech, or non-consensual intimate imagery.79
- The Challenge of Scale: The primary challenge for platforms is the sheer volume of content uploaded every second. Manual moderation is impossible at this scale, forcing a reliance on automated detection systems. However, as discussed in Section 5, these automated tools are imperfect. They can fail to detect sophisticated fakes while also incorrectly flagging legitimate content (false positives), which can lead to accusations of censorship and the suppression of protected speech.6 This creates a difficult balancing act between mitigating harm and protecting freedom of expression.
6.3 Recommendations and Concluding Remarks
The analysis presented in this report demonstrates that the challenges posed by AI-generated media are complex, multifaceted, and dynamic. No single solution—whether technological, legal, or social—will be sufficient to address them. A sustainable and effective path forward requires a multi-layered, defense-in-depth strategy that integrates efforts across society.
- Synthesis of Findings: Generative AI is a powerful dual-use technology whose technical foundations are rapidly evolving. Its benevolent applications in fields like medicine and entertainment are transformative, yet its malicious weaponization for fraud, disinformation, and abuse poses a systemic threat to individual safety, economic stability, and democratic integrity. The ethical dilemmas it raises—from algorithmic bias and the erosion of truth to unresolved IP disputes and profound psychological harm—are deep and complex. While detection technologies offer a line of defense, they are locked in an asymmetric arms race with generative models, making them an incomplete solution.
- A Holistic Path Forward: A resilient societal response must be built on four pillars:
- Continued Technological R&D: Investment must continue in both proactive detection methods like the C2PA standard, which builds trust from the ground up, and in more robust passive detection models. However, this must be done with a clear-eyed understanding of their inherent limitations in the face of an adversarial dynamic.
- Nuanced and Adaptive Regulation: Policymakers should pursue a “smart regulation” approach that is both technology-neutral and harm-specific. International collaboration is needed to harmonize regulations where possible, particularly regarding cross-border issues like disinformation and fraud, while allowing for legal frameworks that can adapt to the technology’s rapid evolution.
- Meaningful Platform Responsibility: Platforms must be held accountable not just for removing illegal content but for the role their algorithms play in amplifying harmful synthetic media. This requires greater transparency into their content moderation and recommendation systems and a shift in incentives away from engagement at any cost.
- Widespread Public Digital Literacy: The ultimate line of defense is a critical and informed citizenry. A massive, sustained investment in public education is required to equip individuals of all ages with the skills to critically evaluate digital media, recognize the signs of manipulation, and understand the psychological tactics used in disinformation and social engineering.
The generative AI revolution is not merely a technological event; it is a profound societal one. The challenges it presents are, in many ways, a reflection of our own societal vulnerabilities, biases, and values. Successfully navigating this new, synthetic reality will depend less on our ability to control the technology itself and more on our collective will to strengthen the human, ethical, and democratic systems that surround it.
| Table 4: Comparative Overview of International Deepfake Regulations | |||
| Jurisdiction | Key Legislation / Initiative | Core Approach | Key Provisions |
| European Union | EU AI Act | Comprehensive, Risk-Based: Classifies AI systems by risk level and applies obligations accordingly.76 | Mandatory, clear labeling of AI-generated content (deepfakes). Transparency requirements for training data. High fines for non-compliance.75 |
| United States | TAKE IT DOWN Act, NO FAKES Act (proposed) | Targeted, Harm-Specific: Focuses on specific harms like non-consensual intimate imagery and unauthorized use of likeness.77 | Makes sharing non-consensual deepfake pornography illegal. Imposes 48-hour takedown obligations on platforms. Creates civil right of action for victims.6 |
| China | Regulations on Deep Synthesis | State-Centric Control: Aims to ensure state oversight and control over the information environment.79 | Mandatory labeling of all AI-generated content (both visible and in metadata). Requires user consent and provides a mechanism for recourse. Prohibits use for spreading “fake news”.75 |
| United Kingdom | Online Safety Act | Platform Accountability: Places broad duties on platforms to protect users from illegal and harmful content.75 | Requires platforms to remove illegal content, including deepfake pornography, upon notification. Focuses on platform systems and processes rather than regulating the technology directly.75 |
Works cited
- Generative AI in Media and Entertainment- Benefits and Use Cases – BigOhTech, accessed September 3, 2025, https://bigohtech.com/generative-ai-in-media-and-entertainment
- AI in Education: 39 Examples, accessed September 3, 2025, https://onlinedegrees.sandiego.edu/artificial-intelligence-education/
- Synthetic data generation: a privacy-preserving approach to …, accessed September 3, 2025, https://pmc.ncbi.nlm.nih.gov/articles/PMC11958975/
- Deepfake threats to companies – KPMG International, accessed September 3, 2025, https://kpmg.com/xx/en/our-insights/risk-and-regulation/deepfake-threats.html
- AI-pocalypse Now? Disinformation, AI, and the Super Election Year – Munich Security Conference – Münchner Sicherheitskonferenz, accessed September 3, 2025, https://securityconference.org/en/publications/analyses/ai-pocalypse-disinformation-super-election-year/
- Take It Down Act, addressing nonconsensual deepfakes and …, accessed September 3, 2025, https://www.klobuchar.senate.gov/public/index.cfm/2025/4/take-it-down-act-addressing-nonconsensual-deepfakes-and-revenge-porn-passes-what-is-it
- Generative artificial intelligence – Wikipedia, accessed September 3, 2025, https://en.wikipedia.org/wiki/Generative_artificial_intelligence
- Generative Artificial Intelligence and the Evolving Challenge of …, accessed September 3, 2025, https://www.mdpi.com/2224-2708/14/1/17
- AI’s Catastrophic Crossroads: Why the Arms Race Threatens Society, Jobs, and the Planet, accessed September 3, 2025, https://completeaitraining.com/news/ais-catastrophic-crossroads-why-the-arms-race-threatens/
- A new arms race: cybersecurity and AI – The World Economic Forum, accessed September 3, 2025, https://www.weforum.org/stories/2024/01/arms-race-cybersecurity-ai/
- What is a GAN? – Generative Adversarial Networks Explained – AWS, accessed September 3, 2025, https://aws.amazon.com/what-is/gan/
- What are Generative Adversarial Networks (GANs)? | IBM, accessed September 3, 2025, https://www.ibm.com/think/topics/generative-adversarial-networks
- Deepfake: How the Technology Works & How to Prevent Fraud, accessed September 3, 2025, https://www.unit21.ai/fraud-aml-dictionary/deepfake
- What are Diffusion Models? | IBM, accessed September 3, 2025, https://www.ibm.com/think/topics/diffusion-models
- Introduction to Diffusion Models for Machine Learning | SuperAnnotate, accessed September 3, 2025, https://www.superannotate.com/blog/diffusion-models
- Deepfake – Wikipedia, accessed September 3, 2025, https://en.wikipedia.org/wiki/Deepfake
- What’s Voice Cloning? How It Works and How To Do It — Captions, accessed September 3, 2025, https://www.captions.ai/blog-post/what-is-voice-cloning
- http://www.forasoft.com, accessed September 3, 2025, https://www.forasoft.com/blog/article/voice-cloning-synthesis#:~:text=The%20voice%20cloning%20process%20typically,tools%20and%20machine%20learning%20algorithms.
- Voice Cloning and Synthesis: Ultimate Guide – Fora Soft, accessed September 3, 2025, https://www.forasoft.com/blog/article/voice-cloning-synthesis
- Scammers use AI voice cloning tools to fuel new scams | McAfee AI …, accessed September 3, 2025, https://www.mcafee.com/ai/news/ai-voice-scam/
- AI in Media and Entertainment: Applications, Case Studies, and …, accessed September 3, 2025, https://playboxtechnology.com/ai-in-media-and-entertainment-applications-case-studies-and-impacts/
- 7 Use Cases for Generative AI in Media and Entertainment, accessed September 3, 2025, https://www.missioncloud.com/blog/7-use-cases-for-generative-ai-in-media-and-entertainment
- 5 AI Case Studies in Entertainment | VKTR, accessed September 3, 2025, https://www.vktr.com/ai-disruption/5-ai-case-studies-in-entertainment/
- How Quality Synthetic Data Transforms the Healthcare Industry …, accessed September 3, 2025, https://www.tonic.ai/guides/how-synthetic-healthcare-data-transforms-healthcare-industry
- Teach with Generative AI – Generative AI @ Harvard, accessed September 3, 2025, https://www.harvard.edu/ai/teaching-resources/
- How AI in Assistive Technology Supports Students and Educators …, accessed September 3, 2025, https://www.everylearnereverywhere.org/blog/how-ai-in-assistive-technology-supports-students-and-educators-with-disabilities/
- The Psychology of Deepfakes in Social Engineering – Reality Defender, accessed September 3, 2025, https://www.realitydefender.com/insights/the-psychology-of-deepfakes-in-social-engineering
- http://www.wa.gov.au, accessed September 3, 2025, https://www.wa.gov.au/system/files/2024-10/case.study_.deepfakes.docx
- Three Examples of How Fraudsters Used AI Successfully for Payment Fraud – Part 1: Deepfake Audio – IFOL, Institute of Financial Operations and Leadership, accessed September 3, 2025, https://acarp-edu.org/three-examples-of-how-fraudsters-used-ai-successfully-for-payment-fraud-part-1-deepfake-audio/
- 2024 Deepfakes Guide and Statistics | Security.org, accessed September 3, 2025, https://www.security.org/resources/deepfake-statistics/
- How can we combat the worrying rise in deepfake content? | World …, accessed September 3, 2025, https://www.weforum.org/stories/2023/05/how-can-we-combat-the-worrying-rise-in-deepfake-content/
- The Malicious Exploitation of Deepfake Technology: Political Manipulation, Disinformation, and Privacy Violations in Taiwan, accessed September 3, 2025, https://globaltaiwan.org/2025/05/the-malicious-exploitation-of-deepfake-technology/
- Elections in the Age of AI | Bridging Barriers – University of Texas at Austin, accessed September 3, 2025, https://bridgingbarriers.utexas.edu/news/elections-age-ai
- We Looked at 78 Election Deepfakes. Political Misinformation Is Not …, accessed September 3, 2025, https://knightcolumbia.org/blog/we-looked-at-78-election-deepfakes-political-misinformation-is-not-an-ai-problem
- How AI Threatens Democracy | Journal of Democracy, accessed September 3, 2025, https://www.journalofdemocracy.org/articles/how-ai-threatens-democracy/
- What are the Major Ethical Concerns in Using Generative AI?, accessed September 3, 2025, https://research.aimultiple.com/generative-ai-ethics/
- How Deepfake Pornography Violates Human Rights and Requires …, accessed September 3, 2025, https://www.humanrightscentre.org/blog/how-deepfake-pornography-violates-human-rights-and-requires-criminalization
- The Impact of Deepfakes, Synthetic Pornography, & Virtual Child …, accessed September 3, 2025, https://www.aap.org/en/patient-care/media-and-children/center-of-excellence-on-social-media-and-youth-mental-health/qa-portal/qa-portal-library/qa-portal-library-questions/the-impact-of-deepfakes-synthetic-pornography–virtual-child-sexual-abuse-material/
- Deepfake nudes and young people – Thorn Research – Thorn.org, accessed September 3, 2025, https://www.thorn.org/research/library/deepfake-nudes-and-young-people/
- Unveiling the Threat- AI and Deepfakes’ Impact on … – Eagle Scholar, accessed September 3, 2025, https://scholar.umw.edu/cgi/viewcontent.cgi?article=1627&context=student_research
- State Laws Criminalizing AI-generated or Computer-Edited CSAM – Enough Abuse, accessed September 3, 2025, https://enoughabuse.org/get-vocal/laws-by-state/state-laws-criminalizing-ai-generated-or-computer-edited-child-sexual-abuse-material-csam/
- Bias in AI | Chapman University, accessed September 3, 2025, https://www.chapman.edu/ai/bias-in-ai.aspx
- What Is Algorithmic Bias? – IBM, accessed September 3, 2025, https://www.ibm.com/think/topics/algorithmic-bias
- research.aimultiple.com, accessed September 3, 2025, https://research.aimultiple.com/ai-bias/#:~:text=Facial%20recognition%20software%20misidentifies%20certain,to%20non%2Ddiverse%20training%20datasets.
- Bias in AI: Examples and 6 Ways to Fix it – Research AIMultiple, accessed September 3, 2025, https://research.aimultiple.com/ai-bias/
- Deepfakes and the Future of AI Legislation: Ethical and Legal …, accessed September 3, 2025, https://gdprlocal.com/deepfakes-and-the-future-of-ai-legislation-overcoming-the-ethical-and-legal-challenges/
- Study finds readers trust news less when AI is involved, even when …, accessed September 3, 2025, https://news.ku.edu/news/article/study-finds-readers-trust-news-less-when-ai-is-involved-even-when-they-dont-understand-to-what-extent
- Generative Artificial Intelligence and Copyright Law | Congress.gov …, accessed September 3, 2025, https://www.congress.gov/crs-product/LSB10922
- Generative AI: Navigating Intellectual Property – WIPO, accessed September 3, 2025, https://www.wipo.int/documents/d/frontier-technologies/docs-en-pdf-generative-ai-factsheet.pdf
- Generative Artificial Intelligence in Hollywood: The Turbulent Future …, accessed September 3, 2025, https://researchrepository.wvu.edu/cgi/viewcontent.cgi?article=6457&context=wvlr
- AI-generated Image Detection: Passive or Watermark? – arXiv, accessed September 3, 2025, https://arxiv.org/html/2411.13553v1
- Passive Deepfake Detection: A Comprehensive Survey across Multi-modalities – arXiv, accessed September 3, 2025, https://arxiv.org/html/2411.17911v2
- [2411.17911] Passive Deepfake Detection Across Multi-modalities: A Comprehensive Survey – arXiv, accessed September 3, 2025, https://arxiv.org/abs/2411.17911
- How To Spot A Deepfake Video Or Photo – HyperVerge, accessed September 3, 2025, https://hyperverge.co/blog/how-to-spot-a-deepfake/
- yuezunli/CVPRW2019_Face_Artifacts: Exposing DeepFake Videos By Detecting Face Warping Artifacts – GitHub, accessed September 3, 2025, https://github.com/yuezunli/CVPRW2019_Face_Artifacts
- Don’t Be Duped: How to Spot Deepfakes | Magazine | Northwestern Engineering, accessed September 3, 2025, https://www.mccormick.northwestern.edu/magazine/spring-2025/dont-be-duped-how-to-spot-deepfakes/
- Reporter’s Guide to Detecting AI-Generated Content – Global …, accessed September 3, 2025, https://gijn.org/resource/guide-detecting-ai-generated-content/
- Defending Deepfake via Texture Feature Perturbation – arXiv, accessed September 3, 2025, https://arxiv.org/html/2508.17315v1
- How voice biometrics are evolving to stay ahead of AI threats? – Auraya Systems, accessed September 3, 2025, https://aurayasystems.com/blog-post/voice-biometrics-and-ai-threats-auraya/
- Leveraging GenAI for Biometric Voice Print Authentication – SMU Scholar, accessed September 3, 2025, https://scholar.smu.edu/cgi/viewcontent.cgi?article=1295&context=datasciencereview
- Traditional Biometrics Are Vulnerable to Deepfakes – Reality Defender, accessed September 3, 2025, https://www.realitydefender.com/insights/traditional-biometrics-are-vulnerable-to-deepfakes
- Challenges in voice biometrics: Vulnerabilities in the age of deepfakes, accessed September 3, 2025, https://bankingjournal.aba.com/2024/02/challenges-in-voice-biometrics-vulnerabilities-in-the-age-of-deepfakes/
- SynthID – Google DeepMind, accessed September 3, 2025, https://deepmind.google/science/synthid/
- C2PA in ChatGPT Images – OpenAI Help Center, accessed September 3, 2025, https://help.openai.com/en/articles/8912793-c2pa-in-chatgpt-images
- C2PA | Verifying Media Content Sources, accessed September 3, 2025, https://c2pa.org/
- How it works – Content Authenticity Initiative, accessed September 3, 2025, https://contentauthenticity.org/how-it-works
- Guiding Principles – C2PA, accessed September 3, 2025, https://c2pa.org/principles/
- C2PA Explainer :: C2PA Specifications, accessed September 3, 2025, https://spec.c2pa.org/specifications/specifications/1.2/explainer/Explainer.html
- Cat-and-Mouse: Adversarial Teaming for Improving Generation and Detection Capabilities of Deepfakes – Institute for Creative Technologies, accessed September 3, 2025, https://ict.usc.edu/research/projects/cat-and-mouse-deepfakes/
- (PDF) Generative Artificial Intelligence and the Evolving Challenge of Deepfake Detection: A Systematic Analysis – ResearchGate, accessed September 3, 2025, https://www.researchgate.net/publication/388760523_Generative_Artificial_Intelligence_and_the_Evolving_Challenge_of_Deepfake_Detection_A_Systematic_Analysis
- Adversarially Robust Deepfake Detection via Adversarial Feature Similarity Learning – arXiv, accessed September 3, 2025, https://arxiv.org/html/2403.08806v1
- Adversarial Attacks on Deepfake Detectors: A Practical Analysis – ResearchGate, accessed September 3, 2025, https://www.researchgate.net/publication/359226182_Adversarial_Attacks_on_Deepfake_Detectors_A_Practical_Analysis
- Deepfake Face Detection and Adversarial Attack Defense Method Based on Multi-Feature Decision Fusion – MDPI, accessed September 3, 2025, https://www.mdpi.com/2076-3417/15/12/6588
- 2D-Malafide: Adversarial Attacks Against Face Deepfake Detection Systems – Eurecom, accessed September 3, 2025, https://www.eurecom.fr/publication/7876/download/sec-publi-7876.pdf
- The State of Deepfake Regulations in 2025: What Businesses Need to Know – Reality Defender, accessed September 3, 2025, https://www.realitydefender.com/insights/the-state-of-deepfake-regulations-in-2025-what-businesses-need-to-know
- EU AI Act: first regulation on artificial intelligence | Topics – European Parliament, accessed September 3, 2025, https://www.europarl.europa.eu/topics/en/article/20230601STO93804/eu-ai-act-first-regulation-on-artificial-intelligence
- Navigating the Deepfake Dilemma: Legal Challenges and Global Responses – Rouse, accessed September 3, 2025, https://rouse.com/insights/news/2025/navigating-the-deepfake-dilemma-legal-challenges-and-global-responses
- AI and Deepfake Laws of 2025 – Regula, accessed September 3, 2025, https://regulaforensics.com/blog/deepfake-regulations/
- China’s top social media platforms take steps to comply with new AI content labeling rules, accessed September 3, 2025, https://siliconangle.com/2025/09/01/chinas-top-social-media-platforms-take-steps-comply-new-ai-content-labeling-rules/
- AI Product Terms – Canva, accessed September 3, 2025, https://www.canva.com/policies/ai-product-terms/
- The Rise of AI-Generated Content on Social Media: A Second Viewpoint | Pfeiffer Law, accessed September 3, 2025, https://www.pfeifferlaw.com/entertainment-law-blog/the-rise-of-ai-generated-content-on-social-media-legal-and-ethical-concerns-a-second-view
- AI-generated Social Media Policy – TalentHR, accessed September 3, 2025, https://www.talenthr.io/resources/hr-generators/hr-policy-generator/data-protection-and-privacy/social-media-policy/









