“Why Some Experts Believe We Should Ban Artificial Intelligence: Insights from the EU AI Act”

Introduction to the EU AI Act

The European Union’s Artificial Intelligence Act (EU AI Act) marks a significant milestone in AI regulation, aiming to protect citizens’ rights while fostering innovation. Adopted on June 13, 2024, the Act began its phased implementation on February 2, 2025. It is designed to provide a comprehensive framework for the development and deployment of artificial intelligence across Europe. The Act’s importance is underscored by its potential to set global standards in AI governance, following in the footsteps of the GDPR and the Digital Markets Act.

Banned AI Applications

Biometric Categorization and Facial Image Scraping

One of the most controversial aspects of the EU AI Act is its ban on certain AI applications, which has led some experts to argue for a broader movement to ban artificial intelligence. The Act prohibits biometric categorization systems that classify individuals based on sensitive characteristics such as race, gender, or sexual orientation. Additionally, the untargeted scraping of facial images from the internet or CCTV to create databases is strictly forbidden. These measures aim to protect citizens’ privacy and prevent potential abuse by AI systems.

Emotion Recognition and Social Scoring

Further restrictions under the Act include bans on emotion recognition technologies in workplaces and schools, unless used for medical or safety reasons. Social scoring systems, which could lead to discriminatory practices, are also prohibited for both public and private purposes. These prohibitions highlight the EU’s commitment to preventing AI from manipulating human behavior or exploiting vulnerabilities.

Predictive Policing and AI in Law Enforcement

The Act also addresses the use of AI in law enforcement. Predictive policing based solely on profiling is banned, reflecting concerns over bias and fairness in AI-driven law enforcement. However, there are specific exemptions and conditions for AI use by law enforcement agencies, underscoring the need for a balanced approach to AI regulation.

Regulatory Framework

Risk-Based Approach

The EU AI Act introduces a risk-based approach to AI regulation, categorizing systems into four levels: unacceptable risk, high risk, limited risk, and minimal risk. This approach ensures that high-risk AI applications are subject to stricter regulations, while allowing innovation in lower-risk areas. This structured framework is crucial for organizations developing and deploying AI technologies, as it provides clear guidelines on compliance requirements.

Obligations for Providers and Deployers

Under the Act, AI providers and deployers are required to adhere to several obligations, including technical documentation, data quality maintenance, human oversight, and transparency. These requirements aim to ensure that AI systems are safe, reliable, and free from bias. Providers must also ensure that their staff possess sufficient AI literacy, underscoring the importance of education and training in the AI sector.

Enforcement and Penalties

Enforcement of the EU AI Act is managed by national authorities, with penalties for non-compliance including fines of up to EUR 15 million or 3% of worldwide turnover for general-purpose AI model providers. This robust enforcement mechanism is designed to ensure adherence to the Act’s provisions and promote accountability among AI developers and users.

Real-World Examples and Case Studies

Biometric Surveillance

Biometric surveillance has raised significant privacy concerns, leading to its prohibition under the EU AI Act. Case studies have shown how such practices can infringe on individuals’ rights and lead to misuse. The Act’s ban on these technologies is a response to these challenges, emphasizing the need for ethical AI deployment.

AI in Healthcare

High-risk AI systems in healthcare, such as those used for diagnosis and treatment recommendations, are subject to stringent regulations under the Act. These systems must demonstrate compliance with safety and transparency standards, ensuring that they provide accurate and unbiased results.

AI in Education

The use of AI in educational settings is another area of focus. The Act mandates compliance with strict guidelines to protect students’ privacy and prevent discrimination. Educational institutions must ensure that their AI systems adhere to these standards to provide a safe learning environment.

Actionable Insights and Best Practices

Establishing Risk Management Systems

To comply with the EU AI Act, businesses should establish comprehensive risk management systems to assess and mitigate AI risks. This involves identifying potential risks associated with AI applications and implementing strategies to address them effectively.

Ensuring Data Quality and Transparency

  • Maintain high-quality data inputs for AI systems.
  • Ensure transparency in AI operations to build trust and accountability.
  • Regularly review and update data quality practices to comply with evolving regulations.

Implementing Human Oversight

Integrating human oversight into AI decision-making processes is crucial for ensuring ethical AI deployment. This involves assigning human operators to monitor AI systems and intervene when necessary to prevent errors or biases.

Challenges & Solutions

Compliance Challenges

Navigating the complexities of the EU AI Act can be challenging for organizations, particularly in balancing AI innovation with data privacy requirements. The regulatory landscape is constantly evolving, requiring businesses to stay informed and adapt to new developments.

Solutions

  • Engage with national supervising authorities for guidance on compliance.
  • Develop internal AI ethics frameworks to align with regulatory standards.
  • Invest in AI governance tools and platforms to manage compliance effectively.

Latest Trends & Future Outlook

Emerging AI Technologies

The rapid advancement of generative AI and other emerging technologies presents new challenges and opportunities for AI regulation. The EU AI Act is expected to influence global regulatory frameworks, setting a precedent for future developments.

Global AI Governance

The EU AI Act’s impact on global AI governance is significant, as it serves as a model for other regions developing their regulatory frameworks. This influence is likely to grow as more countries adopt similar approaches to AI regulation.

Future Developments

As the EU AI Act continues to evolve, businesses must prepare for potential updates and changes to the regulatory landscape. Staying informed and proactive in compliance efforts will be essential for navigating future developments.

Conclusion

In conclusion, while the EU AI Act represents a balanced approach to AI regulation, some experts believe that certain applications warrant a broader movement to ban artificial intelligence. The Act’s provisions aim to protect citizens’ rights and promote ethical AI deployment, setting a global standard for AI governance. As AI technologies continue to evolve, it is crucial for businesses and policymakers to stay informed and engaged in the ongoing dialogue surrounding AI regulation.

More Insights

AI Regulations: Comparing the EU’s AI Act with Australia’s Approach

Global companies need to navigate the differing AI regulations in the European Union and Australia, with the EU's AI Act setting stringent requirements based on risk levels, while Australia adopts a...

Quebec’s New AI Guidelines for Higher Education

Quebec has released its AI policy for universities and Cégeps, outlining guidelines for the responsible use of generative AI in higher education. The policy aims to address ethical considerations and...

AI Literacy: The Compliance Imperative for Businesses

As AI adoption accelerates, regulatory expectations are rising, particularly with the EU's AI Act, which mandates that all staff must be AI literate. This article emphasizes the importance of...

Germany’s Approach to Implementing the AI Act

Germany is moving forward with the implementation of the EU AI Act, designating the Federal Network Agency (BNetzA) as the central authority for monitoring compliance and promoting innovation. The...

Global Call for AI Safety Standards by 2026

World leaders and AI pioneers are calling on the United Nations to implement binding global safeguards for artificial intelligence by 2026. This initiative aims to address the growing concerns...

Governance in the Era of AI and Zero Trust

In 2025, AI has transitioned from mere buzz to practical application across various industries, highlighting the urgent need for a robust governance framework aligned with the zero trust economy...

AI Governance Shift: From Regulation to Technical Secretariat

The upcoming governance framework on artificial intelligence in India may introduce a "technical secretariat" to coordinate AI policies across government departments, moving away from the previous...

AI Safety as a Catalyst for Innovation in Global Majority Nations

The commentary discusses the tension between regulating AI for safety and promoting innovation, emphasizing that investments in AI safety and security can foster sustainable development in Global...

ASEAN’s AI Governance: Charting a Distinct Path

ASEAN's approach to AI governance is characterized by a consensus-driven, voluntary, and principles-based framework that allows member states to navigate their unique challenges and capacities...