UN Launches New Bodies to Govern AI for Global Inclusivity

The UN’s New AI Governance Bodies Explained

The United Nations has taken a significant step towards addressing the challenges posed by artificial intelligence (AI) by launching two new governance bodies: The Global Dialogue on AI Governance and The Independent International Scientific Panel on AI.

A New Framework for AI Governance

This new governance architecture aims to establish a more inclusive form of international governance for AI. According to the World Economic Forum, trustworthy AI ecosystems will be critical for allowing AI to scale safely, sustainably, and inclusively.

Artificial intelligence is not merely a concept from science fiction; it has become an integral part of modern life. During a recent meeting of the UN General Assembly in New York, the risks and opportunities that AI presents were emphasized, with calls for governments to ensure the technology benefits all people.

As the capabilities and deployment of AI continue to accelerate, a concerning statistic emerged: 118 countries are not party to any significant international AI governance initiative. This highlights a significant gap in the establishment of effective, internationally agreed rules for governing AI technology.

The High-Level Meeting on AI Governance

To tackle these issues, the UN General Assembly hosted a high-level meeting on AI governance that brought together diplomats, scientists, members of the tech community, private sector representatives, and civil society. This marked the first time all 193 UN Member States could contribute to the development of international AI governance.

Bridging the Gap in Responsible AI Adoption

AI is described as “the fastest-moving technology in human history,” and it is already transforming various sectors, from healthcare to manufacturing and climate monitoring. However, the widespread adoption of responsible AI—defined as the practice of building and managing AI systems to maximize benefits while minimizing risks—is vital for realizing its full potential. Unfortunately, less than 1% of organizations have fully operationalized responsible AI practices.

Fragmented regulatory approaches pose significant challenges for businesses aiming to implement responsible AI. If left unaddressed, these challenges could undermine public trust, compliance, and investment in AI technology.

Introducing the New UN Bodies

The newly launched UN bodies are designed to address these governance gaps. The Global Dialogue on AI Governance serves as a forum for governments, industry, civil society, and scientists to share best practices and establish common approaches to AI governance. Its goal is to promote interoperability among various governance strands and encourage open innovation.

The Independent International Scientific Panel on AI, likened to an “IPCC for AI,” will consist of 40 expert members who will provide evidence-based insights into the opportunities, risks, and impacts of AI. Its findings will inform the Global Dialogue, positioning it as the world’s early-warning system and evidence engine, helping to differentiate between genuine signals and noise in the AI landscape.

Facilitating Safe Innovation

Experts have praised these new UN initiatives as “the world’s most globally inclusive approach to governing AI.” The need for urgent collaboration on AI governance was echoed by numerous heads of state and corporate leaders during the General Assembly, despite some criticisms regarding the role of international bodies in this domain.

Government initiatives are essential for creating trustworthy AI ecosystems. The Hiroshima AI Process International Guiding Principles, developed by G7 leaders, emphasize the importance of responsible governance frameworks to harness AI’s benefits while managing its associated risks. The World Economic Forum’s report on advancing responsible AI highlights that responsible governance is not a constraint on innovation; rather, it enables AI to scale safely and sustainably throughout society.

To operationalize responsible AI principles, organizations must align corporate strategy with responsible AI innovation, increase organizational capacity, and oversee the entire lifecycle of AI development, acquisition, and use. A strong, collaborative ecosystem that fosters public-private partnerships and international cooperation will be pivotal.

In conclusion, as the Prime Minister of Spain remarked at the UN General Assembly: “The rise of AI is unstoppable, but it cannot be ungovernable.” Ensuring that everyone has a voice in how AI technology is governed will be crucial for its successful integration into society.

More Insights

Revolutionizing Drone Regulations: The EU AI Act Explained

The EU AI Act represents a significant regulatory framework that aims to address the challenges posed by artificial intelligence technologies in various sectors, including the burgeoning field of...

Revolutionizing Drone Regulations: The EU AI Act Explained

The EU AI Act represents a significant regulatory framework that aims to address the challenges posed by artificial intelligence technologies in various sectors, including the burgeoning field of...

Embracing Responsible AI to Mitigate Legal Risks

Businesses must prioritize responsible AI as a frontline defense against legal, financial, and reputational risks, particularly in understanding data lineage. Ignoring these responsibilities could...

AI Governance: Addressing the Shadow IT Challenge

AI tools are rapidly transforming workplace operations, but much of their adoption is happening without proper oversight, leading to the rise of shadow AI as a security concern. Organizations need to...

EU Delays AI Act Implementation to 2027 Amid Industry Pressure

The EU plans to delay the enforcement of high-risk duties in the AI Act until late 2027, allowing companies more time to comply with the regulations. However, this move has drawn criticism from rights...

White House Challenges GAIN AI Act Amid Nvidia Export Controversy

The White House is pushing back against the bipartisan GAIN AI Act, which aims to prioritize U.S. companies in acquiring advanced AI chips. This resistance reflects a strategic decision to maintain...

Experts Warn of EU AI Act’s Impact on Medtech Innovation

Experts at the 2025 European Digital Technology and Software conference expressed concerns that the EU AI Act could hinder the launch of new medtech products in the European market. They emphasized...

Ethical AI: Transforming Compliance into Innovation

Enterprises are racing to innovate with artificial intelligence, often without the proper compliance measures in place. By embedding privacy and ethics into the development lifecycle, organizations...

AI Hiring Compliance Risks Uncovered

Artificial intelligence is reshaping recruitment, with the percentage of HR leaders using generative AI increasing from 19% to 61% between 2023 and 2025. However, this efficiency comes with legal...