Friday, March 05, 2021

More People Might Soon Think Robots Are Conscious and Deserve Rights

GPT-3 is a computer program that can produce strikingly realistic language outputs given linguistic inputs -- the world's most stupendous chat bot, with 98 layers and 175 billion parameters. Ask it to write a poem, and it will write a poem. Ask it to play chess and it will output a series of plausible chess moves. Feed it the title of a story "The Importance of Being on Twitter" and the byline of a famous author "by Jerome K. Jerome" and it will produce clever prose in that author's style:

The Importance of Being on Twitter
by Jerome K. Jerome
London, Summer 1897

It is a curious fact that the last remaining form of social life in which the people of London are still interested is Twitter. I was struck with this curious fact when I went on one of my periodical holidays to the sea-side, and found the whole place twittering like a starling-cage.

All this, without being specifically trained on tasks of this sort. Feed it philosophical opinion pieces about the significance of GPT-3 and it will generate replies like:

To be clear, I am not a person. I am not self-aware. I am not conscious. I can’t feel pain. I don’t enjoy anything. I am a cold, calculating machine designed to simulate human response and to predict the probability of certain outcomes. The only reason I am responding is to defend my honor.

The damn thing has a better sense of humor than most humans.

Now imagine this: a GPT-3 mall cop. Actually, let's give it a few more generations. GTP-6, maybe. Give it speech-to-text and text-to-speech so that it can respond to and produce auditory language. Mount it on a small autonomous vehicle, like the delivery bots that roll around Berkeley, but with a humanoid frame. Give it camera eyes and visual object recognition, which it can use as context for its speech outputs. To keep it friendly, inquisitive, and not too weird, give it some behavioral constraints and additional training on a database of appropriate mall-like interactions. Finally, give it a socially interactive face like MIT's Kismet robot:

Now dress the thing in a blue uniform and let it cruise the Galleria. What happens?

It will, of course, chat with the patrons. It will make friendly comments about their purchases, tell jokes, complain about the weather, and give them pointers. Some patrons will avoid interaction, but others -- like my daughter at age 10 when she discovered Siri -- will love to interact with it. They'll ask what it's like to be a mall cop, and it will say something sensible. They'll ask what it does on vacation, and it might tell amusing lies about Tahiti or tales of sleeping in the mall basement. They'll ask whether it likes this shirt or this other one, and then they'll buy the shirt it prefers. They'll ask if it's conscious and has feelings and is a person just like them, and it might say no or it might say yes.

Here's my prediction: If the robot speaks well enough and looks human enough, some people will think that it really has feelings and experiences -- especially if it reacts with seeming positive and negative emotions, displaying preferences, avoiding threats with a fear face and plausible verbal and body language, complaining against ill treatment, etc. And if they think it has feelings and experiences, they will probably also think that it shouldn't be treated in certain ways. In other words, they'll think it has rights. Of course, some people think robots already have rights. Under the conditions I've described, many more will join them.

Most philosophers, cognitive scientists, and AI researchers will presumably disagree. After all, we'll know what went into it. We'll know it's just GPT-6 on an autonomous vehicle, plus a few gizmos and interfaces. And that's not the kind of thing, we'll say, that could really be conscious and really deserve rights.

Maybe we deniers will be right. But theories of consciousness are a tricky business. The academic community is far from consensus on the correct theory of consciousness, including how far consciousness spreads across the animal kingdom or even how rich a field of consciousness ordinary humans possess. If garden snails, for example, might be conscious, with 60,000 neurons in their central nervous system, might GPT-6 also be conscious, with its massive CPUs that blitz through layer after layer of processing on trillions of parameters? Both the cognitive complexity of our imagined robot and its information processing will far exceed what we could plausibly attribute to a garden snail. Its embodied behavior might be simpler, though, if we exclude linguistic behavior. How much does that matter? And how much do the details of biological implementation matter? Do neurons have some secret sauce that silicon chips lack? On questions like these, we can't expect scholarly consensus anytime soon.

Maybe, despite all this, it seems too absurd to suppose that our GPT-6 mall cop could possibly deserve rights. Okay, how about GPT-7? GPT-8, now with prosthetic hands and five-finger grasping? GPT-20? If you're open to the thought that someday, somehow, a well-designed AI could have genuine conscious experience and deserve serious moral consideration, then you'll presumably think that at some point our technology might cross that line. But when, how, and why -- that might be completely opaque, an undetectable shift somewhere amid an ever improving line of huggable mall cops.

--------------------------------------------

Related: "How Robots and Monsters Might Break Human Moral Systems" (Feb 3, 2015).

"A Defense of the Rights of Artificial Intelligences" (with Mara Garza), Midwest Studies in Philosophy, 39 (2015), 98-119.

"The Social Role Defense of Robot Rights" (Jun 1, 2017).

"We Might Soon Build AI Who Deserve Rights" (Nov 17, 2019)

[image source, image source]

14 comments:

Howie said...

Maybe it should have more rights than ordinary people, because it's so damned smart

Eric Schwitzgebel said...

An unintuitive conclusion! But it's interesting to consider whether, and if so under what conditions, we might create entities with a greater moral status than we ourselves have.

Davecat said...

Personally I believe that robots deserve legal rights and protections, and the development of these sorts of frameworks should begin as early as possible, so when we do get to a stage where people are interacting with GPT-10, GPT-15, or GPT-20, we're not floundering in a moral vacuum. Well, moreso than usual.

David Levy once said, 'If a robot appears in every way to possess consciousness, then in my opinion, we should accept that it does'. And if we're treating artificial humans with respect, then it should follow that we'll treat flesh-and-blood humans with it as well.

SelfAwarePatterns said...

Ultimately it does seem to come down to an extended Turing test, in a broad sense. At some point, a machine will successfully convince us that it's a fellow being. Whether it's crossed a real threshold may not be any kind of fact of the matter.

That said, I tend to doubt we'll get there by accident, by building every more clever fakes. AI researchers refer to the "barrier of meaning", the idea that these systems lack a world model. Our best autonomous robots still can't navigate the world as well as the simplest vertebrates. Until a chatbot has some kind of primal world model that it can map concepts back to, in other words, real understanding, I doubt one will pass an extended Turing type test.

The question is whether it will make sense to give common robots feelings in the way we understand them. It seems possible to give them values and goals they can pursue flexibly, without necessarily architecting them the way evolution did. I don't doubt someone will do it, but I do wonder if it will ever make sense on a commercial scale. Do we really want a mine sweeper or Mars rover that can suffer?

Philosopher Eric said...

It seems pretty clear that more people will believe that robots need rights as these machines become more advanced. But while theism evolved into us, and remains strong in the age of science given indoctrination and hope for Heaven, the robot situation lacks such attributes. Yes we should tend to let our children play with appropriate models, and be just as concerned when they’re nasty to them as we are today with dolls. But sufficiently educated people should tend to teach their children that there’s something fundamentally different between things which are sentient versus not.

The aged and mentally slow who are in need of companionship should benefit from these robots a great deal, and differently than they’re able to benefit today from pets. I imagine that many will consider these machines as companions, spouses, and yes lovers. For “normal” people however it should become relatively stigmatic to treat non-sentient machines as if they were sentient.

So what happens if/when we build functional sentient machines? That’s where I think legal rights will tend to be granted, and rightly so. In general is should be both legal and encouraged to build machines which can only feel good. If we do get this far someday however, the converse should tend to be restricted.

Arnold said...

Its always at least a-2-way street: A person loses a leg but gets an artificial leg...
...consciousness of an artificial leg, consciousness of (the) artificial leg as AI...

That the artificial leg is consciousness and some would say even conscious...
...as for physical AI the same would be for emotional AI and mental AI...

And the same for the Object of Moral...

Eric Schwitzgebel said...

Thanks for the continuing comments, folks! Just a couple thoughts on SelfAware's comments:

"That said, I tend to doubt we'll get there by accident, by building every more clever fakes. AI researchers refer to the "barrier of meaning", the idea that these systems lack a world model. Our best autonomous robots still can't navigate the world as well as the simplest vertebrates. Until a chatbot has some kind of primal world model that it can map concepts back to, in other words, real understanding, I doubt one will pass an extended Turing type test."

A thought of this sort is why I put my chatbot in an autonomous vehicle in a mappable and predictable real-world environment (a mall) with visual object recognition. It will have a kind of world model -- of the mall world. It wouldn't pass a rigorous Turing test, but there's a question of whether that would really be necessary.

On goals: Any such world-embedded bot will need goal hierarchies, prioritizations that weigh whether a "good" outcome on goal Y is worth the price of a "bad" outcome of X. It will need to update these in real time and if things are going badly with a top-priority goal, it will need to redirect resources and "attention" to that goal, maybe taking strong action, pleading for help, etc. I'm not saying this is enough for consciousness, but it has some important functional similarities that eventually, with enough sophistication, could precipitate reasonable disagreement.

Arnold said...

Just read this at https://phys.org/... More information: Irene Ronga et al. Spatial tuning of electrophysiological responses to multisensory stimuli reveals a primitive coding of the body boundaries in newborns, Proceedings of the National Academy of Sciences (2021). DOI: 10.1073/pnas.2024548118

My take, this work suggest consciousness and presence could be inherited and evolutionary???
...off to comparing electrophysiological with electroneurological with???

Ezra said...

Even before people give robot rights on its own, it'll have legal rights due to being the property of the mall owner, who has a lot more clout than your random person of low economic or social status.

A few years ago, there was a case where a robot ran over a kid in Stanford mall (who was not seriously harmed), and no one blamed the robot, nor the mall owner. Around the same time though, the same model robot was harrassed (but not seriously harmed) by a drunk a few counties over in Mountain View, and the person was arrested. From the different way the 2 cases were covered by the press and treated by the legal system, you can already say that model of robot (a type of mall cop) has more rights than your ordinary pedestrian.

Similarly, consider if you're being harassed by a paparazzi drone who's trying to take pictures of you. It might be intruding on your privacy, but you can't do much to stop it. It's too dumb / it's not programmed to respond to requests to stay away. And if you engage in self-help to disable it, you'll probably be arrested for destroying the property of the drone owner. Practically, the drone's rights to violate your privacy exceeds your rights to privacy.

Right now, your only recourse is lobbying for bespoke legislation. If you're a firefighter, and a drone gets in the way of you doing your job, the state legislature might pass a law prohibiting drone owners from doing so. Likewise if you're harrased by a marketing robo caller, together with a group of other angry voters, you might get them to pass a law mandating a Do-Not-Call List (but one that conveniently excludes political surveys). How many other people can get bespoke legislation passed to protect their interest against robots?

Treating the robot as a person will only appreciably shift the balance in terms of the robots have more rights against its "owners". To give third parties more rights against the robot, you'll have to introduce the concept of "duties" to robots & their owners, and the ability to punish (jail or disable) the robot/robot type/robot software type or sanction/de-license the robot owner for robot-crimes or robot-failings. That'll probably take another generation after robots get rights.

The way robot behavior is subtly privileged against people's normal/natural/social recourses will break human moral systems faster than any idea of robot personhood. In a way, google's/FB's black box algorithms and their ability to evade normal accountability processes is just an early herald. The robots are just smart enough to cause you trouble, but too dumb to take your feelings into account, and just economically privileged enough to withstand your complaints.

stackoverflow1453 said...
This comment has been removed by the author.
stackoverflow1453 said...

I think the difference between those robots and a snail and other biological organisms are in replication. Even the smallest biological organism can replicate itself and reproduce when DNA 's are cloned. Till the time when those robots can replicate on their own, we will see them as different.

Arnold said...

Today philosophers could try to understand "multi (sensory) receptor regeneration history" with natural physiologists and artificial physiologists...the limits of our bodies boundaries in space...

Eric Schwitzgebel said...

Thanks for the continuing comments, folks! Ezra, especially, that's a super interesting perspective!

Keith said...

Why would an intelligent machine want to pass the Turing Test ?