AI’s High Risk in the Election Landscape

Elections Watchdog Warned of High AI Risks in Current Campaign

An internal briefing note prepared for Canada’s election watchdog highlights the use of artificial intelligence (AI) as a high risk for the ongoing election campaign. This document was created for the Commissioner of Canada Elections, Caroline Simard, roughly a month before the campaign began.

AI’s Potential Impact on Elections

The briefing note suggests that the upcoming election will likely generate complaints related to the use of AI tools, potentially violating the Canada Elections Act. While AI can serve legitimate purposes, the document emphasizes that it poses significant risks that could lead to contraventions of election rules.

Concerns Over Disinformation

According to the note, the Elections Act does not explicitly prohibit the use of AI, bots, or deepfakes; however, certain provisions could apply if AI tools are misused. Violations may encompass the spreading of disinformation, publishing false information about the electoral process, or impersonating election officials.

Michael Litchfield, director of the AI risk and regulation lab at the University of Victoria, pointed out the challenges of identifying individuals who misuse AI to violate election rules. The inability to trace these actions complicates enforcement efforts.

Deepfakes and Their Threats

The briefing note raises specific concerns regarding the use of deepfakes—hyperrealistic faked videos or audio. While there have been no reported incidents of deepfakes in Canadian federal elections, there have been multiple examples abroad, including a notable deepfake involving Kamala Harris during the 2024 U.S. presidential election. The document warns that similar occurrences could happen in Canada.

Generative AI can create convincing fakes that may significantly impact public perception, even if they are quickly debunked. The note states that an increase in advertising for customized deepfake services on the dark web has been observed, highlighting the potential for misuse during the election.

Existing Regulatory Framework

The document indicates that Canada has generally relied on a self-regulation approach regarding AI, leaving much of the oversight to the tech industry. However, the effectiveness of this self-regulation is questioned. Despite some AI image generators having policies against election disinformation, they have failed to prevent the creation of misleading content.

Bill C-27, which would partially regulate AI usage in elections, was introduced but did not reach the legislative deadline. Experts believe that even if regulations are passed, they may take time to enforce, placing Canada in a regulatory vacuum.

Future Concerns

As the current election campaign progresses, there are already signs of AI being used to disseminate misinformation. AI-generated articles have surfaced, spreading dubious information about party leaders’ personal finances. Furthermore, fake election news ads have attempted to lure Canadians into fraudulent investment schemes.

Fenwick McKelvey, an assistant professor at Concordia University, noted that the misuse of AI contributes to a decline in public trust in online content, complicating the credibility of legitimate information sources.

The briefing note warns that the use of AI is likely to trigger numerous complaints during the election campaign, even in cases where no specific rules have been broken. This situation could lead to complex assessments and a large-scale impact on the election process.

Conclusion

The ongoing developments in AI technology present both opportunities and challenges in the context of elections. While AI can enhance campaign strategies, its potential for misuse raises significant concerns about the integrity of the electoral process. As the election landscape evolves, it is crucial to navigate these challenges carefully to preserve the democratic process.

More Insights

Harnessing Trusted Data for AI Success in Telecommunications

Artificial Intelligence (AI) is transforming the telecommunications sector by enhancing operations and delivering value through innovations like IoT services and smart cities. However, the...

Morocco’s Leadership in Global AI Governance

Morocco has taken an early lead in advancing global AI governance, as stated by Ambassador Omar Hilale during a recent round table discussion. The Kingdom aims to facilitate common views and encourage...

Regulating AI: The Ongoing Battle for Control

The article discusses the ongoing debate over AI regulation, emphasizing the recent passage of legislation that could impact state-level control over AI. It highlights the tension between innovation...

AI Readiness Framework for the Pharmaceutical Industry

This article presents an AI readiness assessment framework tailored for the pharmaceutical industry, emphasizing the importance of aligning AI initiatives with regulatory standards and ethical...

Enhancing AI Safety through Responsible Alignment

The post discusses the development of phi-3-mini in alignment with Microsoft's responsible AI principles, focusing on safety measures such as post-training safety alignment and red-teaming. It...

Mastering Sovereign AI Clouds in Intelligent Manufacturing

Sovereign AI clouds provide essential control and compliance for manufacturers, ensuring that their proprietary data remains secure and localized. As the demand for AI-driven solutions grows, managed...

Empowering Ethical AI in Scotland

The Scottish AI Alliance has released its 2024/2025 Impact Report, showcasing significant progress in promoting ethical and inclusive artificial intelligence across Scotland. The report highlights...

EU AI Act: Embrace Compliance and Prepare for Change

The recent announcement from the EU Commission confirming that there will be no delay to the EU AI Act has sparked significant reactions, with many claiming both failure and victory. Companies are...

Exploring Trustworthiness in Large Language Models Under the EU AI Act

This systematic mapping study evaluates the trustworthiness of large language models (LLMs) in the context of the EU AI Act, highlighting their capabilities and the challenges they face. The research...