Switzerland’s Bold Move Towards AI Innovation

Can Switzerland Steer a Safe Course to AI Innovation?

Switzerland’s long-awaited strategy for artificial intelligence (AI) focuses on promoting business while postponing regulations aimed at shielding the public from potential risks associated with the technology.

This strategy typifies Switzerland’s light-touch regulatory approach, similar to its practices in other sectors like commodities trading. The government has committed to a broad set of principles drawn up by the Council of Europe, yet it has not opted for the stringent regulations enacted by the European Union last year.

The Shift in Global Sentiment

This announcement from Switzerland has been enthusiastically received by business associations but has raised concerns among civil society groups regarding issues of privacy, sustainability, and the increasing power of corporations. The recent trend prioritizing safety, exemplified by the 2024 European Union AI Act, is being overshadowed by a global scramble for AI dominance, primarily driven by the United States.

Late to the Game

Switzerland’s official AI strategy has been released later than many other advanced economies, as it seeks to balance the conflicting views from the EU and US. The government aims to regulate AI in a manner that leverages its potential to enhance Switzerland’s business and innovation landscape while minimizing societal risks.

Legal Foundations and Measures

The Council of Europe AI Convention seeks to defend democracy, the rule of law, and human rights against abuses of AI technology. This convention is more targeted towards public sector projects and offers signatories significant latitude for legal implementation. Proposed law changes will be presented to the Swiss parliament by the end of 2026, with additional time required for amending existing laws, including data protection legislation.

In tandem with these legal frameworks, the Swiss government plans to implement “non-legally binding measures” for private companies, which may include self-disclosure agreements or industry-specific solutions.

Risk Levels and Self-Regulation

AI has evolved from merely analyzing large datasets to making independent conclusions that can both fascinate and alarm society. The technology has far-reaching implications across various sectors, including healthcare, law enforcement, and automated transport.

In contrast to the EU’s structured approach to AI risks, the US has adopted a more hands-off policy since the administration of Donald Trump. This sentiment was echoed by US Vice President JD Vance, who emphasized the need for a regulatory regime that encourages the growth of AI technology rather than stifling it.

Concerns from Civil Society

While some in the Swiss AI sector welcome this balanced approach, civil society groups like AlgorithmWatch consider the strategy to be “a step in the right direction” but lacking in foresight. They urge the government to act promptly and decisively to address sustainability issues and protect individual rights in the face of growing corporate dominance in the AI sector.

Conclusion

The Swiss government has positioned itself to outline broad strokes for AI policy now, with plans to fill in the details later. By ratifying the Council of Europe AI Convention, Switzerland aims to avoid being sidelined in terms of trustworthiness in the international arena. However, the effectiveness of this approach relies on how closely these measures align with the EU AI Act and whether they provide robust protections for society while fostering innovation.

More Insights

US Rejects UN’s Call for Global AI Governance Framework

U.S. officials rejected the establishment of a global AI governance framework at the United Nations General Assembly, despite broad support from many nations, including China. Michael Kratsios of the...

Agentic AI: Managing the Risks of Autonomous Systems

As companies increasingly adopt agentic AI systems for autonomous decision-making, they face the emerging challenge of agentic AI sprawl, which can lead to security vulnerabilities and operational...

AI as a New Opinion Gatekeeper: Addressing Hidden Biases

As large language models (LLMs) become increasingly integrated into sectors like healthcare and finance, a new study highlights the potential for subtle biases in AI systems to distort public...

AI Accountability: A New Era of Regulation and Compliance

The burgeoning world of Artificial Intelligence (AI) is at a critical juncture as regulatory actions signal a new era of accountability and ethical deployment. Recent events highlight the shift...

Choosing Effective AI Governance Tools for Safer Adoption

As generative AI continues to evolve, so do the associated risks, making AI governance tools essential for managing these challenges. This initiative, in collaboration with Tokio Marine Group, aims to...

UN Initiatives for Trustworthy AI Governance

The United Nations is working to influence global policy on artificial intelligence by establishing an expert panel to develop standards for "safe, secure and trustworthy" AI. This initiative aims to...

Data-Driven Governance: Shaping AI Regulation in Singapore

The conversation between Thomas Roehm from SAS and Frankie Phua from United Overseas Bank at the SAS Innovate On Tour in Singapore explores how data-driven regulation can effectively govern rapidly...

Preparing SMEs for EU AI Compliance Challenges

Small and medium-sized enterprises (SMEs) must navigate the complexities of the EU AI Act, which categorizes many AI applications as "high-risk" and imposes strict compliance requirements. To adapt...

Draft Guidance on Reporting Serious Incidents Under the EU AI Act

On September 26, 2025, the European Commission published draft guidance on serious incident reporting requirements for high-risk AI systems under the EU AI Act. Organizations developing or deploying...