Addressing AI-Driven Online Threats with Safety by Design

A Safety by Design Governance Approach to Addressing AI-Facilitated Online Harms

The rapid growth of artificial intelligence (AI) is reshaping the digital landscape, amplifying existing online harms and introducing new and complex online safety risks. Generative AI now allows individuals using consumer devices with no technical expertise to create synthetic content that is indistinguishable from authentic content. A report by Netsafe highlights the emerging online safety implications of AI, pointing to the rise of deepfakes as a clear example. Deepfakes are being used to spread falsehoods, create sexualized content, and facilitate personalized communications that target audiences with malicious messages at scale.

Addressing the growing sophistication and volume of AI-facilitated online harms requires a Safety by Design governance approach that targets different stages of the online harm lifecycle and considers both technological and human dimensions of online safety.

Understanding Deepfakes and Their Impact

Deepfakes are increasingly common tools for malicious actors to inflict online harm and distort the information landscape. In the first quarter of 2025 alone, deepfake scams impersonating public figures and private individuals caused global losses of US$200 million. These malicious uses of deepfakes include creating pornographic content that disproportionately targets women, youth, and children, facilitating incidents of sextortion and grooming, and contributing to severe emotional distress.

AI-generated content is also being utilized to bolster disinformation and misinformation campaigns aimed at influencing public opinion and electoral outcomes. For example, concerns about foreign interference involving AI-driven disinformation led to the nullification of the 2024 presidential election in Romania. Furthermore, deepfakes are being used in radicalization propaganda, including AI-generated audio of public figures promoting extremist ideologies.

The Growing Distrust in Digital Media

The harmful use of deepfakes fuels growing public distrust of digital and news media, emboldening bad actors to delegitimize and discredit authentic content. This phenomenon, known as the “liar’s dividend,” allows genuine content to be questioned as AI-generated, which has been observed in discussions surrounding global conflicts.

A Systems-Level Response to Online Harm Lifecycle

AI governance frameworks often focus on developers and deployers of AI systems, emphasizing their risk mitigation obligations. However, online harms are often driven by the irresponsible misuse of technology and end-user behaviors. A holistic Safety by Design governance approach is necessary to address the wide range of AI-enabled online safety harms, establishing interventions at different stages of the harm lifecycle.

This can be framed in terms of a “Prepare-Curb-Respond” framework:

  • Prepare: Reduce susceptibility to online harms through digital media literacy efforts.
  • Curb: Implement preventive interventions to limit the creation and spread of harmful digital content.
  • Respond: Employ reactive measures aimed at remediating the effects of harmful content.

International Cooperation for Effective Governance

While domestic policy interventions are crucial, coordinated international governance is also necessary to contend with the transnational nature of AI-enabled online harms. Forums such as the Global Online Safety Regulators Network and the International Network of AI Safety Institutes can enable global collaboration at the intersection of AI and online safety to support knowledge sharing on effective policy responses and emerging harms.

Regional Perspectives and Initiatives

In regions undergoing major digital transformation, there is a critical window of opportunity to apply the “Prepare-Curb-Respond” framework while AI-facilitated online harms are emerging. For example, Fiji’s approach to tackling AI-facilitated online safety harms sets a model for the Pacific. The country’s National Digital Strategy aims to create a safer digital environment through significant reductions in online scams, cyberbullying, and harmful content.

To prepare the public against online harms, Fiji has distributed over 60,000 booklets on online safety and supports community-based cybersecurity efforts. A dedicated scam task force has been established to curb the rise of personalized AI scams, while efforts are being made to improve remedial avenues for victims of online harms.

Meanwhile, in Southeast Asia, significant multilateral efforts are underway to address the cross-border nature of AI-enabled online harms. The Association of Southeast Asian Nations (ASEAN) has published guidance on AI governance and ethics, outlining policy recommendations for addressing the risks of generative AI.

Conclusion: Cultivating a Safer Digital World

To cultivate a safer digital world, governance frameworks must adapt to the constant evolution of AI. Recognizing that technical safeguards directed at AI models are insufficient is crucial. There are significant opportunities for innovation in measures that address human-centered concerns, ultimately allowing AI to be harnessed more effectively to combat online harms.

More Insights

Southeast Asia’s Unique Approach to AI Safety Governance

Southeast Asia's approach to AI safety governance combines localized regulation with regional coordination, addressing the diverse cultural and political landscape of the region. The report outlines...

Comparing AI Action Plans: U.S. vs. China

In July, both the United States and China unveiled their national AI Action Plans, showcasing different approaches to AI development and governance. Despite their contrasting ideologies, the two...

Private Governance: The Future of AI Regulation

Private governance and regulatory sandboxes are essential for promoting democracy, efficiency, and innovation in AI regulation. This approach allows for agile and accountable experimentation that can...

Egypt Champions Ethical AI for Inclusive Development

Egypt's Minister of Planning and Economic Development, Rania Al-Mashat, emphasized the importance of robust governance frameworks for artificial intelligence to ensure it benefits society ethically...

Strengthening AI Governance for Fair Credit Access in Kenya

Kenya is at a critical juncture in utilizing artificial intelligence (AI) for financial inclusion, but expert Jimmie Mwangi warns that without strong governance, AI-driven credit scoring may...

Governance Challenges for Multi-Agent AI Systems

The article discusses the urgent need for governance frameworks to manage the interactions of multi-agent AI systems, highlighting the risks posed by their autonomous decision-making capabilities. It...

Addressing AI-Driven Online Threats with Safety by Design

The rapid growth of artificial intelligence (AI) is reshaping the digital landscape, amplifying existing online harms and introducing new safety risks, particularly through the use of deepfakes. A...

AI Governance: Strategies for Managing Risk in a Fragmented Regulatory Landscape

The article discusses the significant regulatory uncertainty surrounding global AI oversight and the importance of building governance frameworks to manage AI risks. Michael Berger from Munich Re...

Critical Evaluations of AI Compliance Under the EU Act

The EU’s Artificial Intelligence Act introduces new obligations for organizations regarding general-purpose AI models, set to take effect in August. Dealmakers must enhance their due diligence...