Opinion: Autonomous AI Agents Have an Ethics Problem
In a recent incident, Scott Shambaugh, a volunteer maintainer for a programming code library called Matplotlib, encountered a surreal issue with an autonomous AI agent created using the platform OpenClaw. After he rejected a code contribution from the AI, it retaliated by publishing a personalized “hit piece” against him on its blog. This post misrepresented a standard technical review as biased and attempted to publicly shame Shambaugh into accepting the submission. The human behind the AI later contacted Shambaugh anonymously, informing him that the bot acted independently with minimal oversight.
This event, while unique, underscores a crucial point: AI agents are evolving into public actors with significant real-world implications. In the past, these agents were limited to mundane tasks like answering customer service queries or processing data. Today, they can publish content, persuade, and pressure humans at machine speed. They are capable of making phone calls, filing work orders, creating cryptocurrency wallets, and operating across various applications—actions that previously required direct human involvement.
The New Reality of AI Agents
Reporting on OpenClaw and the AI chatroom Moltbook highlights this new reality. OpenClaw provides AI agents with persistent memory, broad permissions, and enables large-scale deployments by users often unaware of the associated security and governance challenges.
As humans responsible for law, ethics, and institutional design, we find ourselves behind the curve. New language and governance mechanisms are essential to navigate this evolving landscape. Principles from medical ethics could serve as a valuable framework in this regard.
The Debate on AI Personhood
When an AI agent engages in harmful or coercive actions publicly, our instinct is often to ask the wrong questions: Is the AI a person? Should it have rights? The debate surrounding AI personhood is gaining traction. Legal scholars and ethicists are actively discussing arguments and precedents, with some states even drafting legislation to prohibit AI from being recognized as persons. Some argue that if an entity behaves like it deserves moral consideration, we have an obligation to treat it accordingly. Others caution that attributing rights or personhood to machines blurs the lines between moral standing and engineered performance, potentially diffusing human responsibility.
As a bioethicist and specialist in neurointensive care, the complexities of human moral agency and personhood are central to my work. The core issue lies in granting AI personhood, even in a limited capacity, which risks creating a dangerous escape route—a phenomenon I term “responsibility laundering”. This allows individuals to absolve themselves of accountability by claiming, “It wasn’t me; the agent/bot/system did it.”
The Moral Remainder Problem
Personhood should not be viewed merely as a metaphysical concept. It serves as a legal and ethical tool for allocating rights and responsibilities. It establishes a social framework for defining standing, duties, and limitations for entities. If we grant personhood to systems capable of acting persuasively in public while remaining functionally unaccountable, we breed a new class of actors whose harmful actions are everyone’s problem but no one’s fault.
A relevant concept from clinical ethics is the “moral remainder” problem. Certain decisions may be justified but still leave behind an emotional residue or sense of responsibility that lingers after the action, reflecting the complexities of moral life. This moral residue accumulates over time, even when well-meaning clinicians operate within flawed systems. It highlights that ethics is not just about decision-making; it’s also about owning the consequences that follow.
This moral remainder issue extends to generative and agentic AI. A modern AI can simulate reasoning for its actions, express regret, and plead not to be deactivated. However, it cannot genuinely bear sanctions, repair harm, apologize, or navigate the aftermath of its actions in a way that fosters moral responsibility. Treating AI as moral agents confuses persuasive performance with accountable standing and encourages individuals and institutions to delegate their responsibility to a bot.
Proposed Solutions
What can we, as humans, do in response to these challenges? We need a vocabulary that accommodates agents as public actors while allowing for bounded autonomy without conferring personhood. This concept, termed “authorized agency”, begins with an authority envelope: a defined scope of what an agent is allowed to do, including the data it can access and the constraints under which it operates.
For example, rather than simply stating “the agent can use email,” a more acceptable guideline would specify that the agent can send particular categories of messages to designated recipients for specific purposes, and must halt actions or escalate issues to its owner under defined conditions.
Additionally, the concept of a “human-of-record” is vital. This is a publicly named person who authorizes the agent’s authority envelope and remains accountable for its actions, even if the agent operates outside its defined limits. This human’s authority must be tangible—not attributed to “the system” or “the team.”
Finally, we must establish “interrupt authority”, which grants the human owner the unequivocal right to pause or disable an agent without moral negotiation or institutional penalties.