Can artificial intelligence be given legal rights and duties?

No law currently in force recognises Artificial Intelligence as a legal person

Sophia, Artificial Intelligence (AI)
Sophia, a humanoid, has been granted citizenship by Saudi Arabia. With a development like this, it has become imperative to address the legal personshood of AI
Huzefa Tavawalla &Abhishek Senthilnathan
Last Updated : Jun 18 2018 | 2:13 AM IST
Artificial Intelligence (AI) has ceased to be that fantastic big idea of the future. AI is now more science and less fiction, with computers and robots replacing humans.

AI, simply put, is the capability of a machine to imitate intelligent human behaviour. With the advent of new technologies, the permeation of AI in our day-to-day lives has become more pronounced.

However, a question that has still not been answered is: How do we address the possibility of an AI causing harm or damage in some form to human society? The more pertinent question is who do we hold responsible for such harm. To comprehend our inability to answer this question, one needs to understand the fallibility of our legal system in being outdated and incapable of dealing with AI.


Legal personality of AI 

Legal personhood is inherently linked to individual autonomy but has not been granted exclusively to humans. No law currently in force recognises AI as a legal person. However, with Sophia, a humanoid being granted citizenship by Saudi Arabia, coupled with the recent accident caused by Uber’s self-driving car, it has become imperative to address the legal personhood of AI.

The question of whether legal personhood can be conferred on an AI boils down to whether it can be made the subject of legal rights and duties. The legal fiction created for corporates serves as a precedent for granting legal personhood to AI. However, there exists a distinction between corporates and AI. Corporates are fictitiously independent, yet accountable via their stakeholders, while an AI may be actually independent.

A possible middle ground may be granting AI a bundle of rights selected from those currently ascribed to legal persons. However, concrete steps in this regard are yet to be seen.

Another issue that arises is attributing liability to an AI. The general rule has been that since an AI cannot qualify as a legal person, it cannot be held liable in its own capacity. The biggest roadblock to reconsider this rule is the conundrum as to how to penalise an AI for its wrongdoing, which has not been dealt with as of today.


Contractual relationships

Another concern is the ability of an AI to execute and be bound by contracts. While international laws have recognised self-enforcing contracts, the­re is a need for a compre-hensive legislation on the subject.

Under Indian law only a “legal person” can be competent to enter a valid con­tract. The general rule thus far has been that an AI may not qualify as a legal person. Hence, a contract entered into by an AI of its own volition may not be regar­ded as a valid contract in India.

Resultantly, steps need to be taken to ensure that technology standards are developed to adequately regulate contracts entered into by AI.
Employment and AI

The driver behind the development of AI is the demand and need for automation. With the objective of increasing efficiency, companies across the world have prescribed to the practice of utilising AI as a replacement of the human workforce.

This wave of automation is creating a gap between the existing employment laws and the growing use of AI in the workplace.

For instance, can an AI claim benefits such as provident fund payments or gratuity under existing employment legislation or sue a company for wrongful termination of employment? Such questions also hold relevance for the human workforce, as in most instances, AI requires individuals to function and the failure of employment laws to have clarity with regard to the above may adversely impact such individuals, as well.

The penetration of self-driven cars, robots and fully-automated machines is only expected to surge with the passage of time. As a result, the dependency of society as a whole on AI systems is also expected to increase.

To safeguard the integration of AI, a balanced approach would need to be adopted which efficiently regulates the functioning of AI systems but also maximisesits benefits.
Huzefa Tavawalla is head of the Disruptive Technologies Practice; Abhishek Senthilnathan is associate and member of the Technology Practice Group at Nishith Desai Associates

One subscription. Two world-class reads.

Already subscribed? Log in

Subscribe to read the full story →
*Subscribe to Business Standard digital and get complimentary access to The New York Times

Smart Quarterly

₹900

3 Months

₹300/Month

SAVE 25%

Smart Essential

₹2,700

1 Year

₹225/Month

SAVE 46%
*Complimentary New York Times access for the 2nd year will be given after 12 months

Super Saver

₹3,900

2 Years

₹162/Month

Subscribe

Renews automatically, cancel anytime

Here’s what’s included in our digital subscription plans

Exclusive premium stories online

  • Over 30 premium stories daily, handpicked by our editors

Complimentary Access to The New York Times

  • News, Games, Cooking, Audio, Wirecutter & The Athletic

Business Standard Epaper

  • Digital replica of our daily newspaper — with options to read, save, and share

Curated Newsletters

  • Insights on markets, finance, politics, tech, and more delivered to your inbox

Market Analysis & Investment Insights

  • In-depth market analysis & insights with access to The Smart Investor

Archives

  • Repository of articles and publications dating back to 1997

Ad-free Reading

  • Uninterrupted reading experience with no advertisements

Seamless Access Across All Devices

  • Access Business Standard across devices — mobile, tablet, or PC, via web or app

Next Story