I, Alexa: Should we give artificial intelligence human rights?

ai personhood ethics questions hanson robotics sophia
Hanson Robotics

A few years ago, the subject of AI personhood and legal rights for artificial intelligence would have been something straight out of science fiction. In fact, it was.

Douglas Adams’ second Hitchhiker’s Guide to the Galaxy book, The Restaurant at the End of the Universe, tells the story of a futuristic smart elevator called the Sirius Cybernetics Corporation Happy Vertical People Transporter. This artificially intelligent elevator works by predicting the future, so it can appear on the right floor to pick you up even before you know you want to get on — thereby “eliminating all the tedious chatting, relaxing, and making friends that people were previously forced to do whilst waiting for elevators.”

The ethics question, Adams explains, comes when the intelligent elevator becomes bored of going up and down all day, and instead decides to experiment with moving from side to side as a “sort of existential protest.”

We don’t yet have smart elevators, although judging by the kind of lavish headquarters tech giants like Google and Apple build for themselves, that may just be because they’ve not bothered sharing them with us yet. In fact, as we’ve documented time and again at Digital Trends, the field of AI is currently making a bunch of things possible we never thought realistic in the past — such as self-driving cars or Star Trek-style universal translators.

Have we also reached the point where we need to think about rights for AIs?

You’ve gotta fight for your right to AI

It’s pretty clear to everyone that artificial intelligence is getting closer to replicating the human brain inside a machine. On a low resolution level, we currently have artificial neural networks with more neurons than creatures like honey bees and cockroaches — and they’re getting bigger all the time.

Have we also reached the point where we need to think about rights for AIs?

Higher up the food chain are large-scale projects aimed at creating more biofidelic algorithms, designed to replicate the workings of the human brain, rather than simply being inspired by the way we lay down memories. Then there are projects designed to upload consciousness into machine form, or something like the so-called “OpenWorm” project, which sets out to recreate the connectome — the wiring diagram of the central nervous system — for the tiny hermaphroditic roundworm Caenorhabditis elegans, which remains the only fully-mapped connectome of a living creature humanity has been able to achieve.

In a 2016 survey of 175 industry experts, the median expert expected human-level artificial intelligence by 2040, and 90 percent expected it by 2075.

Before we reach that goal, as AI surpasses animal intelligence, we’ll have to begin to consider how AIs compare to the kind of “rights” that we might afford animals through ethical treatment. Thinking that it’s cruel to force a smart elevator to move up and down may not turn out to be too far-fetched; a few years back English technology writer Bill Thompson wrote that any attempt to develop AI coded to not hurt us, “reflects our belief that an artificial intelligence is and always must be at the service of humanity rather than being an autonomous mind.”

ai personhood ethics questions elevator

The most immediate question we face, however, concerns the legal rights of an AI agent. Simply put, should we consider granting them some form of personhood?

This is not as ridiculous as it sounds, nor does it suggest that AIs have “graduated” to a particular status in our society. Instead, it reflects the complex reality of the role that they play — and will continue to play — in our lives.

Smart tools in an age of non-smart laws

At present, our legal system largely assumes that we are dealing with a world full of non-smart tools. We may talk about the importance of gun control, but we still hold a person who shoots someone with a gun responsible for the crime, rather than the gun itself. If the gun explodes on its own as the result of a faulty part, we blame the company which made the gun for the damage caused.

So far, this thinking has largely been extrapolated to cover the world of artificial intelligence and robotics. In 1984, the owners of a U.S. company called Athlone Industries wound up in court after their robotic pitching machines for batting practice turned out to be a little too vicious. The case is memorable chiefly because of the judge’s proclamation that the suit be brought against Athlone rather than the batting bot, because “robots cannot be sued.”

This argument held up in 2009, when a U.K. driver was directed by his GPS system to drive along a narrow cliffside path, resulting in him being trapped and having to be towed back to the main road by police. While he blamed the technology, a court found him guilty of careless driving.

ai personhood ethics questions wrongwaygps
Sean Ryan / Rapid City Journal
Sean Ryan / Rapid City Journal

There are multiple differences between AI technologies of today (and certainly the future) and yesterday’s tech, however. Smart devices like self-driving cars or robots won’t just be used by humans, but deployed by them — after which they act independently of our instructions. Smart devices, equipped with machine learning algorithms, gather and analyze information by themselves and then make their decisions. It may be difficult to blame the creators of the technology, too.

“Courts may hesitate to say that the designer of such a component could have foreseen the harm that occurred.”

As David Vladeck, a law professor at Georgetown University in Washington D.C., has pointed out in one of the few in-depth case studies looking at this subject, the sheer number of individuals and firms that participate in the design, modification, and incorporation of an AI’s components can make it tough to identify who the party responsible is. That counts for double when you’re talking about “black boxed” AI systems that are inscrutable to outsiders.

Vladeck has written: “Some components may have been designed years before the AI project had even been conceived, and the components’ designers may never have envisioned, much less intended, that their designs would be incorporated into any AI system, much less the specific AI system that caused harm. In such circumstances, it may seem unfair to assign blame to the designer of a component whose work was far removed in both time and geographic location from the completion and operation of the AI system. Courts may hesitate to say that the designer of such a component could have foreseen the harm that occurred.”

It’s the corporations, man!

Awarding an AI the status of a legal entity wouldn’t be unprecedented. Corporations have long held this status, which is why a corporation can own property or be sued, rather than this having to be done in the name of its CEO or executive board.

Although it hasn’t been tested, Shawn Bayern, a law professor from Florida State University, has pointed out that technically AI may have already have this status due to the loophole that it can be put in charge of a limited liability company, thereby making it a legal person. This might also occur for tax reasons, should a proposal like Bill Gates’ “robot tax” ever be taken seriously on a legal level.

It’s not without controversy, however. Granting AIs this status would stop creators being held responsible if an AI somehow carries out an action its creator was not explicitly responsible for. But this could also encourage companies to be less diligent with their AI tools — since they could technically fall back on the excuse that those tools acted outside their wishes.

There is also no way to punish an AI, since punishments like imprisonment or death mean nothing

“I’m not convinced that this is a good thing, certainly not right now,” Dr. John Danaher, a law professor at NUI Galway in Ireland, told Digital Trends about legal personhood for AI. “My guess is that for the foreseeable future this will largely be done to provide a liability shield for humans and to mask anti-social activities.”

It is a compelling area of examination, however, because it doesn’t rely on any benchmarks being achieved in terms of ever-subjective consciousness.

“Today, corporations have legal rights and are considered legal persons, whereas most animals are not,” Yuval Noah Harari, author of Sapiens: A Brief History of Humankind and Homo Deus: A Brief History of Tomorrow, told us. “Even though corporations clearly have no consciousness, no personality and no capacity to experience happiness and suffering; whereas animals are conscious entities.”

“Irrespective of whether AI develops consciousness, there might be economic, political and legal reasons to grant it personhood and rights in the same way that corporations are granted personhood and rights. Indeed, AI might come to dominate certain corporations, organizations and even countries. This is a path only seldom discussed in science fiction, but I think it is far more likely to happen than the kind of Westworld and Ex Machina scenarios that dominate the silver screen.”

Not science fiction for long

At present, these topics still smack of science fiction but, as Harari points out, they may not stay that way for long. Based on their usage in the real world, and the very real attachments that form with them, questions such as who is responsible if an AI causes a person’s death, or whether a human can marry his or her AI assistant, are surely ones that will be grappled with during our lifetimes.

robots tax san francisco jobs of the future ex machina vfx
Universal Pictures
Universal Pictures

“The decision to grant personhood to any entity largely breaks down into two sub-questions,” Danaher said. “Should that entity be treated as a moral agent, and therefore be held responsible for what it does? And should that entity be treated as a moral patient, and therefore be protected against certain interferences and violations of its integrity? My view is that AIs shouldn’t be treated as moral agents, at least not for the time being. But I think there may be cases where they should be treated as moral patients. I think people can form significant attachments to artificial companions and that consequently, in many instances, it would be wrong to reprogram or destroy those entities. This means we may owe duties to AIs not to damage or violate their integrity.”

In other words, we shouldn’t necessarily allow companies to sidestep the question of responsibility when it comes to the AI tools they create. As AI systems are rolled out into the real world in everything from self-driving cars to financial traders to autonomous drones and robots in combat situations, it’s vital that someone is held accountable for what they do.

At the same, it’s a mistake to think of AI as having the same relationship with us that we enjoyed with previous non-smart technologies. There’s a learning curve here and, if we’re not yet technologically at the point where we need to worry about cruelty to AIs, that doesn’t mean it’s the wrong question to ask.

So stop yelling at Siri when it mishears you and asks whether you want it to search the web, alright?

Emerging Tech

Awesome Tech You Can’t Buy Yet: camera with A.I. director, robot arm assistant

Check out our roundup of the best new crowdfunding projects and product announcements that hit the web this week. You may not be able to buy this stuff yet, but it sure is fun to gawk!
Movies & TV

Stay inside this winter with the best shows on Hulu, including 'Killing Eve'

It's often overwhelming to navigate Hulu's robust library of TV shows. To help, we put together a list of the best shows on Hulu, whether you're into frenetic cartoons, intelligent dramas, or anything in between.
Gaming

Take a trip to a new virtual world with one of these awesome HTC Vive games

So you’re considering an HTC Vive, but don't know which games to get? Our list of 25 of the best HTC Vive games will help you out, whether you're into rhythm-based gaming, interstellar dogfights, or something else entirely.
Emerging Tech

A Japanese hotel fires half its robot staff for being bad at their jobs

Japan’s oddball Henn na Hotel has fired half of its 243 robot staff. The reason? Because these labor-saving machines turned out to be causing way more problems than they were solving.
Emerging Tech

Saturn didn’t always have rings, according to new analysis of Cassini data

Saturn's rings are younger than previously believed, according to new data gathered from the Cassini mission. The rings are certainly less than 100 million years old and perhaps as young as 10 million years old.
Emerging Tech

Water-based fuel cell converts carbon emissions to electricity

Scientists from Korea's Ulsan National Institute of Science and Technology have developed a system which can continuously produce electrical energy and hydrogen by dissolving carbon dioxide in an aqueous solution.
Emerging Tech

Scientists investigate how massive stars die in dramatic hypernova events

Our Sun will gradually fade before expanding into a red giant at the end of its life. But larger mass stars undergo extreme explosive events called hypernovas when they die which outshine their entire galaxies.
Emerging Tech

Pilotless planes are on their way, but would you fly in one?

Airbus says advancements in artificial intelligence can help it toward its goal of building a plane capable of fully autonomous flight, though whether passengers can be persuaded to travel in one is another matter entirely.
Emerging Tech

‘Tech vest’ prevents Amazon workers from colliding with robot co-workers

Amazon workers at its fulfillment centers are using "tech vests" to help protect them from collisions with their robot co-workers. The robots already have obstacle avoidance sensors, but the belt offers another layer of safety.
Emerging Tech

3D printers are finally affordable. Here are the best models under $500

3D printer prices have dropped dramatically over the past few years, but just because something is cheap doesn’t mean it’s worth buying. Here, we’ve rounded up all the cheap 3D printers that are actually worth spending your money on.
Mobile

T-Mobile 5G rollout: Here is everything you need to know

2019 will be a huge year for T-Mobile. Not only is a merger with Sprint likely, but T-Mobile is also in the midst of building out its next-generation mobile service. Here's everything you need to know about the T-Mobile 5G rollout.
Emerging Tech

ANYmal dog robot can get back on its feet when someone pushes it over

Roboticists at ETH Zurich have demonstrated how their ANYmal four-legged robot is capable of taking a kicking and keeping on walking -- or getting back to its feet if it's pushed over.
Emerging Tech

A.I. finds non-infringing ways to copy drugs pharma spends billions developing

Researchers have demonstrated an artificial intelligence which can find new methods for producing existing pharmaceuticals in a way that doesn’t infringe on existing patents. Here's how.
Emerging Tech

Coinstar machines will let you swap cash for Bitcoin at your local grocery store

Coinstar, the company which owns the coin exchange machines found at grocery stores and elsewhere, will soon let you easily buy Bitcoin with your cash money. Here's how it will work.