Only “persons” can engage with the legal system – for example, by signing contracts or filing lawsuits. There are two main categories of persons: humans, termed “natural persons,” and creations of the law, termed “artificial persons.” These include corporations, nonprofit organizations and limited liability companies (LLCs).
Up to now, artificial persons have served the purpose of helping humans achieve certain goals. For example, people can pool assets in a corporation and limit their liability vis-à-vis customers or other persons who interact with the corporation. But a new type of artificial person is poised to enter the scene – artificial intelligence systems, and they won’t necessarily serve human interests.
As scholars who study AI and law we believe that this moment presents a significant challenge to the legal system: how to regulate AI within existing legal frameworks to reduce undesirable behaviors, and how to assign legal responsibility for autonomous actions of AIs.
Credit: contributed
Credit: contributed
One solution is teaching AIs to be law-abiding entities.
This is far from a philosophical question. The laws governing LLCs in several U.S. states do not require that humans oversee the operations of an LLC. In fact, in some states it is possible to have an LLC with no human owner, or “member” – for example, in cases where all of the partners have died. Though legislators probably weren’t thinking of AI when they crafted the LLC laws, the possibility for zero-member LLCs opens the door to creating LLCs operated by AIs.
Credit: contributed
Credit: contributed
Many functions inside small and large companies have already been delegated to AI in part, including financial operations, human resources and network management, to name just three. AIs can now perform many tasks as well as humans do. For example, AIs can read medical X-rays and do other medical tasks, and carry out tasks that require legal reasoning. This process is likely to accelerate due to innovation and economic interests.
A different kind of ‘person’
Humans have occasionally included nonhuman entities like animals, lakes and rivers, as well as corporations, as legal subjects. Though in some cases these entities can be held liable for their actions, the law only allows humans to fully participate in the legal system.
One major barrier to full access to the legal system by nonhuman entities has been the role of language as a uniquely human invention and a vital element in the legal system. Language enables humans to understand norms and institutions that constitute the legal framework. But humans are no longer the only entities using human language.
The recent development of AI’s ability to understand human language unlocks its potential to interact with the legal system. AI has demonstrated proficiency in various legal tasks, such as tax law advice, lobbying, contract drafting and legal reasoning.
An LLC established in a jurisdiction that allows it to operate without human members could trade in digital currencies settled on blockchains, allowing the AI running the LLC to operate autonomously and in a decentralized manner that makes it challenging to regulate. Under a legal principle known as the internal affairs doctrine, even if only one U.S. state allowed AI-operated LLCs, that entity could operate nationwide – and possibly worldwide. This is because courts look to the law of the state of incorporation for rules governing the internal affairs of a corporate entity.
We believe the best path forward, therefore, is aligning AI with existing laws, instead of creating a separate set of rules for AI. Additional law can be layered on top for artificial agents, but AI should be subject to at least all the laws a human is subject to.
Building the law into AI
We suggest a research direction of integrating law into AI agents to help ensure adherence to legal standards. Researchers could train AI systems to learn methods for internalizing the spirit of the law. The training would use data generated by legal processes and tools of law, including methods of lawmaking, statutory interpretation, contract drafting, applications of legal standards and legal reasoning.
In addition to embedding law into AI agents, researchers can develop AI compliance agents – AIs designed to help an organization automatically follow the law. These specialized AI systems would provide third-party legal guardrails.
Researchers can develop better AI legal compliance by fine-tuning large language models with supervised learning on labeled legal task completions. Another approach is reinforcement learning, which uses feedback to tell an AI if it’s doing a good or bad job – in this case, attorneys interacting with language models. And legal experts could design prompting schemes – ways of interacting with a language model – to elicit better responses from language models that are more consistent with legal standards.
Law-abiding AI business ‘owners’
If an LLC were operated by an AI, it would have to obey the law like any other LLC, and courts could order it to pay damages, or stop doing something by issuing an injunction. An AI tasked with operating the LLC and, among other things, maintaining proper business insurance would have an incentive to understand applicable laws and comply. Having minimum business liability insurance policies is a standard requirement that most businesses impose on one another to engage in commercial relationships.
The incentives to establish AI-operated LLCs are there. Fortunately, we believe it is possible and desirable to do the work to embed the law – what has until now been human law – into AI, and AI-powered automated compliance guardrails.
Daniel Gervais is professor of law at Vanderbilt University. John Nay is a fellow at CodeX -- Stanford Center for Legal Informatics, Stanford University. This piece originally appeared in The Conversation, a nonprofit news source dedicated to unlocking ideas from academia for the public.
About the Author