AI Governance: From Theory to Practical Challenges

Key Takeaways from the AI Governance Roundtable at Loeb’s AI Summit

The AI Governance roundtable at Loeb’s AI Summit provided valuable insights into the varied governance structures that organizations are employing as they navigate the complexities of artificial intelligence.

AI Governance Structures

Participants highlighted that organizations are operating under a wide range of AI governance structures. Some have adopted centralized models, where oversight is concentrated in a core AI governance team. In contrast, others prefer federated structures, which allow individual business units to maintain responsibility through designated AI leads.

Regardless of the structure, a common theme emerged: speed is a primary friction point. Lengthy review and approval cycles are perceived as significant barriers to innovation and business adoption.

Evolution of Governance Frameworks

Most organizations have advanced beyond standalone AI policies, now maintaining more developed governance frameworks and operational processes. AI governance teams are generally tasked with:

  • Reviewing and approving AI tools and use cases
  • Implementing guardrails and assessing risk prior to launch

However, several participants noted that these frameworks are unevenly socialized. In some cases, tools or use cases slip through without adequate review.

Training and Literacy Gaps

Training and literacy emerged as major gaps in the current landscape. Companies reported insufficient education regarding:

  • AI risk
  • Governance obligations
  • Practical tool usage

Even where enterprise AI tools are licensed, underutilization remains a challenge if teams lack the knowledge or confidence to deploy them effectively.

Lifecycle Governance Challenges

Participants emphasized that lifecycle governance presents additional challenges. Oversight often focuses on pre-launch review, but fewer organizations have mature processes for:

  • Post-deployment monitoring
  • Version updates
  • Ongoing risk reassessment

This gap is expected to widen with the rise of agentic AI, which may significantly alter governance models and complicate centralized visibility.

Integration with Privacy Governance

Companies with mature privacy governance programs appear better positioned to integrate AI governance into existing risk management structures. Where strong data governance foundations exist, AI oversight can be layered onto established processes. Conversely, organizations lacking these foundations are still in the process of building baseline structures while simultaneously addressing AI-specific risks.

Conclusion

Overall, the discussion at the roundtable reflected a shift from theoretical AI governance to operational execution challenges. Key issues identified include speed, visibility, lifecycle oversight, and workforce enablement. As organizations continue to navigate these complexities, the insights gained from this roundtable will be invaluable in shaping future AI governance strategies.

More Insights

Revolutionizing Drone Regulations: The EU AI Act Explained

The EU AI Act represents a significant regulatory framework that aims to address the challenges posed by artificial intelligence technologies in various sectors, including the burgeoning field of...

Revolutionizing Drone Regulations: The EU AI Act Explained

The EU AI Act represents a significant regulatory framework that aims to address the challenges posed by artificial intelligence technologies in various sectors, including the burgeoning field of...

Embracing Responsible AI to Mitigate Legal Risks

Businesses must prioritize responsible AI as a frontline defense against legal, financial, and reputational risks, particularly in understanding data lineage. Ignoring these responsibilities could...

AI Governance: Addressing the Shadow IT Challenge

AI tools are rapidly transforming workplace operations, but much of their adoption is happening without proper oversight, leading to the rise of shadow AI as a security concern. Organizations need to...

EU Delays AI Act Implementation to 2027 Amid Industry Pressure

The EU plans to delay the enforcement of high-risk duties in the AI Act until late 2027, allowing companies more time to comply with the regulations. However, this move has drawn criticism from rights...

White House Challenges GAIN AI Act Amid Nvidia Export Controversy

The White House is pushing back against the bipartisan GAIN AI Act, which aims to prioritize U.S. companies in acquiring advanced AI chips. This resistance reflects a strategic decision to maintain...

Experts Warn of EU AI Act’s Impact on Medtech Innovation

Experts at the 2025 European Digital Technology and Software conference expressed concerns that the EU AI Act could hinder the launch of new medtech products in the European market. They emphasized...

Ethical AI: Transforming Compliance into Innovation

Enterprises are racing to innovate with artificial intelligence, often without the proper compliance measures in place. By embedding privacy and ethics into the development lifecycle, organizations...

AI Hiring Compliance Risks Uncovered

Artificial intelligence is reshaping recruitment, with the percentage of HR leaders using generative AI increasing from 19% to 61% between 2023 and 2025. However, this efficiency comes with legal...