Wednesday, December 18, 2024

Reply to Chalmers: If I'm Living in a Simulation, It Might be Brief or Small

Suppose we take the "simulation hypothesis" seriously: We might be living not in the "base level" of reality but instead inside of a computer simulation.

I've argued that if we are living in a computer simulation, it might easily be only city-sized or have a short past of a few minutes, days, or years. The world might then be much smaller than we ordinarily think it is.

David Chalmers argues otherwise in a response published on Monday. Today I'll summarize his argument and present my first thoughts toward a rebuttal.

The Seeding Challenge: Can a Simulation Contain Coherent, Detailed Memories and Records but Only a Short Past?

Suppose an Earth-sized simulation was launched last night at midnight Pacific Standard Time. The world was created new, exactly then, with an apparent long past -- fake memories already in place, fake history books, fake fossil records, and all the rest. I wake up and seem to recall a promise I made to my wife yesterday. I greet her, and she seems to recall the same promise. We read the newspaper, full of fake news about the unreal events of yesterday -- and everyone else on the planet reads their own news of the same events, and related events, all tied together in an apparently coherent web.

Chalmers suggests that the obvious way to make this work would be to run a detailed simulation of the past, including a simulation of my conversation with my wife yesterday, and our previous past interactions, and other people's past conversations and actions, and all the newsworthy world events, and so on. The simulators create today's coherent web of detailed memories and records by running a simulated past leading up to the "start time" of midnight. But if that's the simulators' approach, the simulation didn't start at midnight after all. It started earlier! So it's not the short simulation hypothesized.

This reasoning iterates back in time. If we wanted a simulation that started on Jan 1, 2024, we'd need a detailed web of memories, records, news, and artifacts recently built or in various stages of completion, all coherently linked so that no one detects any inconsistencies. The obvious way to generate a detailed, coherent web of memories and records would be to run a realistic simulation of earlier times, creating those memories and records. Therefore, Chalmers argues, no simulation containing detailed memories and records can have only a short past. Whatever start date in the recent past you choose, in order for the memories and records to be coherent, a simulation would already need to be running before that date.

Now, as I think Chalmers would acknowledge, although generating a simulated past might be the most obvious way to create a coherent web of memories and records, it's not the only way. The simulators could instead attempt to directly seed a plausible network of memories and records. The challenge would lie in seeding them coherently. If the simulators just create a random set of humanlike memories and newspaper stories, there will be immediately noticeable conflicts. My wife and I won't remember the same promise from yesterday. The news article dated November 1 will contradict the article dated October 31.

Call this the Seeding Challenge. If the Seeding Challenge can be addressed, the simulators can generate a coherent set of memories and records without running a full simulation of the past.

To start, consider geological seeding. Computer games like SimCity and Civilization can autogenerate plausible, coherent terrain that looks like it has a geological history. Rivers run from mountains to the sea. Coastlines are plausible. Plains, grasslands, deserts, and hills aren't checkered randomly on the map but cluster with plausible transitions. Of course, this is simple, befitting simple games with players who care little about strict geological plausibility. But it's easy to imagine more careful programming by more powerful designers that does a better job, including integrating fossil records and geological layers. If done well enough, there might be no inconsistency or incoherence. Potentially, before finalizing, a sophisticated plausibility and coherence checker could look for and repair any mistakes.

I see no reason in principle that human memories, newspaper stories, and the rest couldn't be coherently seeded in a similar way. If my memory is seeded first, then my wife's memory will be constrained to match. If the November 1 news stories are seeded first, then the October 31 stories will be constrained to match. Big features might be seeded first -- like a geological simulation might start with "mountain range here" -- and then details articulated to match.

Naturally, this would be extremely complicated and expensive! But we are imagining a society of simulators who can simulate an entire planet of eight billion conscious humans, and all of the many, many physical interactions those humans have with the simulated environment, so we are already imagining the deployment of huge computational power. Let's not underestimate their capacity to meet the Seeding Challenge by rendering the memories and records coherent.

This approach to the Seeding Challenge gains plausibility, I think, by considering the resource-intensiveness of the alternative strategy of creating a deep history. Suppose the simulators want a start date of midnight last night. Option 1 would be to run a detailed simulation of the entire Earth from at least the beginning of human history. Option 2 would be to randomly generate a coherent seed, checking and rechecking for any detectable inconsistencies. Even though generating a coherent seed might be expensive and resource intensive, it's by no means clear that it would be more expensive and resource intensive than running a fully detailed simulated Earth for thousands of years.

I conclude that Chalmers' argument against short-historied simulations does not succeed.


The Boundaries Challenge: Can a Simulation Be City-Sized in an Apparently Large World?

I have also suggested that a simulation could easily just be you and your city. Stipulate a city that has existed for a hundred years. Its inhabitants falsely believe they are situated on a large planet containing many other cities. Everyone and everything in the city exists, but everything stops at the city's edge. Anyone who looks beyond the edge sees some false screen. Anyone who travels out of the city disappears from existence -- and when they return, they pop back into existence with false memories of having been elsewhere. News from afar is all fake.

Chalmers' objection is similar to his objection to short-past simulations. How are the returning travelers' memories generated? If someone in the city has a video conversation with someone far away, how is that conversation generated? The most obvious solution again seems to be to simulate the distant city the traveler visited and to simulate the distant conversation partner. But now we no longer have only a city-sized simulation. If the city is populous with many travelers and many people who interact with others outside the city, to keep everything coherent, Chalmers argues, you probably need to simulate all of Earth. Thus, a city-sized simulation faces a Boundaries Challenge structurally similar to the short-past simulation's Seeding Challenge.

The challenge can be addressed in a similar way.

Rendering travelers' memories coherent is a task structurally similar to rendering the memories of newly-created people coherent. The simulators could presumably start with some random, plausible seeds, then constrain future memories by those first seeds. This would of course be difficult and computationally expensive, but it's not clear that it would be more difficult or more expensive than simulating a whole planet of interacting people just so that a few hundred thousand or a few million people in a city don't notice any inconsistencies.

If the city's inhabitants have real-time conversations with others elsewhere, that creates a slightly different engineering challenge. As recent advances in AI technology have vividly shown, even with our very limited early 21st century tools, relatively plausible conversation partners can easily be constructed. With more advanced technology, presumably even more convincing conversation partners would be possible -- though their observations and memories would need to be constantly monitored and seeded for coherence with inputs from returning travelers, other conversation partners, incoming news, and so on.

Chalmers suggests that such conversation partners would be simulations -- and thus that the simulation wouldn't stop at the city's edge after all. He's clearly right about this, at least in a weak sense. Distant conversation partners would need voices and faces resembling the voices and faces of real people. In the same limited sense of "simulation", a video display at the city's edge, showing trees and fields beyond, simulates trees and fields. So yes, the borders of the city will need to be simulated, as well as the city itself. Seeming-people in active conversation with real citizens will in the relevant sense count as part of the borders of the city.

But just as trees on a video screen need not have their backsides simulated, so also needn't the conversation partners continue to exist after the conversation ends. And just as trees on a video screen needn't be as richly simulated as trees in the center of the city, so also distant conversation partners needn't be richly simulated. They can be temporary shells, with just enough detail to be convincing, and with new features seeded only on demand as necessary.

The Boundary Problem for simulated cities introduces one engineering challenge not faced by short-history whole-Earth simulations: New elements need to be introduced coherently in real time. A historical seed can be made slowly and checked over patiently as many times as necessary before launch. But the city boundaries will need to be updated constantly. If generating coherent conversation partners, memories, and the like is resource intensive, it might be challenging to do it fast enough to keep up with all the trips, conversations, and news reports streaming in.

Here, however, the simulators can potentially take advantage of the fact that the city's inhabitants are themselves simulations running on a computer. If real-time updating of the boundary is a challenge, the simulators can slow down the clock speed or pause as necessary, while the boundaries update. And if some minor incoherence is noticed, it might be possible to rewrite citizens' memories so it is quickly forgotten.

So although embedding a city-sized simulation in a fake world is probably more complicated than generating a short-past simulation with a fake history, ultimately my response to Chalmers' objections is the same for both cases: There's no reason to suppose that generating plausible, coherent inputs to the city would be beyond the simulator's capacities, and doing so on the fly might be much less computationally expensive than running a fully detailed simulation of a whole planet with a deep history.

Related:

"1% Skepticism" (2017), Nous, 51, 271-290.

"Let’s Hope We’re Not Living in a Simulation" (2024), Philosophy & Phenomenological Research, online first: https://onlinelibrary.wiley.com/doi/10.1111/phpr.13125.

Chalmers, David J. (2024) "Taking the Simulation Hypothesis Seriously", Philosophy & Phenomenological Research, online first: https://onlinelibrary.wiley.com/doi/10.1111/phpr.13122.

Friday, December 13, 2024

Age and Philosophical Fame in the Early Twentieth Century

In previous work, I've found that eminent philosophers tend to do their most influential work when they are in their 40s (though the age range has a wider spread than eminent scientists, who rarely do their most influential work in their 50s or later).  I have also found some data suggesting that philosophers tend to be discussed most when they are about age 55-70, well after they produce their most influential work.  It seems to take about 15-20 years, on average, for a philosopher's full import to be felt by the field.

I was curious to see if the pattern holds for philosophers born 1850-1899, whom we can examine systematically using the new Edhiphy tool.  (Edhiphy captures mentions of philosophers' names in articles in leading philosophy journals, 1890-1980.)

Here's what I did:

First, I had Edhiphy output the top-50 most-mentioned philosophers from 1890-1980, limited to philosophers with recorded birthyear from 1850-1899.[1]  For each philosopher, I went to their Ediphy profile and had Edhiphy output a graph showing the number of articles in which that philosopher was cited per year.  For example, here's the graph for George Santayana (1863-1952):

[Articles mentioning George Santayana per year, in a few selected philosophy journals, per Edhiphy; click to enlarge and clarify]

I then recorded the peak year for each philosopher (1928 for Santayana).  As you can see, the display is a little visually confusing, so it's possible that in some cases my estimate was off by a year.

One complication is that there are many more total mentions of philosophers in the later decades than the earlier decades -- partly due to more articles in the database for later decades, but probably also partly due to changes in citation practices.  Still, most authors (like Santayana) show enough decline over time that late citations don't swamp their first peak.  So instead of trying to introduce a systematic adjustment to discount later mentions I simply recorded the raw peak.  For the thirteen philosophers with more than one equal-valued peak, I took the earlier year (e.g., John Dewey was mentioned in 48 articles in both 1940 and 1951, so I treated 1940 as his peak).

In accord with previous work, I found that philosophers' peak discussion tended to occur late in life.  The median age at peak discussion was 67.5 (mean 68.8).

Four outliers peaked over age 100: David Hilbert (112), Pierre Duhem (114), Giuseppe Peano (116), and Karl Pearson (121).  However, it's probably fair to say that none of these four was primarily known as a philosopher in their lifetimes: Hilbert, Peano, and Pearson were mathematicians and Duhem a physicist.  Almost everyone else on the list is primarily known as a philosopher, so these four are not representative.  Excluding these outliers, the median is 66.5 and mean is 64.7, and no one peaked after age 90.

Three philosophers peaked by age 40: Ralph Barton Perry (peaked at age 35 in 1911), C. D. Broad (peaked at age 40 in 1927), and William Pepperell Montague (peaked at age 40 in 1913).  Broad's early peak -- as you can see from the graph below -- is due to an outlier year, without which his peak would have been much later.  On the other hand, given the overall increase in mentions over time, we should probably be discounting the later decades anyway.

[Edhiphy citations of C.D. Broad; click to enlarge and clarify]

Six philosophers peaked age 44 to 49; five peaked in their 50s; 14 in their 60s; 10 in their 70s; and 8 in their 80s.

You might wonder whether the philosophers who peaked late also produced their most influential work late.  There is a trend in this direction.  Hans Reichenbach, who peaked in 1978 at age 87, produced his most cited work in 1938 (at age 47).  L. J. Russell, who peaked in 1970 at age 86, appears to have produced his most cited work in 1942 (at age 58).  Edmund Husserl, who peaked in 1941 at age 82, produced his most cited work in 1913 (at age 54)  John Dewey, who peaked in 1940 at age 81, produced his most cited work in 1916 (at age 57).  Ernst Cassirer, who peaked in 1955 at age 81 produced his most-cited work in 1944 (at age 70).  Still, for all but Cassirer the delay between most-cited work and peak discussion is over 20 years.

A similar spread occurs in the middle of the pack.  The five philosophers with peak citation at median ages 67-68 (the median age of peak citation for the group as a whole) produced their most-cited works at ages 30 (Karl Japsers), 42 (J. M. E. McTaggart), 45 (C. I. Lewis), 49 (Max Scheler), and 61 (Samuel Alexander).  For this group too, the typical delay between most-cited work and peak citation is about twenty years.

Although the peak age is a little later than I would have predicted based on earlier work, overall I'd say the data for early twentieth century philosophers tends to confirm trends I found in my earlier work on mid-to-late twentieth-century philosophers.  Specifically:

(1.) Philosophers produce their most influential work at a wide range of ages, but mid-40s is typical.

(2.) The peak rates of discussion of philosophers' work tends to come late in life, typically decades after they have published their most influential work.

Articles mentioning JME McTaggart, by year 1890-1980 in Edhiphy.  Note peak in the late 1930s. McTaggart's most influential publication was in 1908.

------------------------------------------------------------

[1] Edhiphy has a few peculiar gaps in birthyear data.  By far the most conspicuous are Gottlob Frege (born 1848) and Albert Einstein (1879).  However, Frege is outside my target period, and Einstein is not primarily known as a philosopher, so this shouldn't much distort the results.  Several figures with missing birthdates are psychologists (Scripture, Binet, Hering) or physicists (Bridgman, Maxwell).  H. A. Prichard is perhaps the most discussed straight philosopher born in the period whose birthdate is not recorded in Ediphy.

Friday, December 06, 2024

Morally Confusing AI Systems Should Have Doubt-Producing Interfaces

We shouldn't create morally confusing AI. That is, we shouldn't create AI systems whose moral standing is highly uncertain -- systems that are fully conscious and fully deserving of humanlike rights according to some respectable mainstream theories, while other respectable mainstream theories suggest they are mere empty machines that we can treat as ordinary tools.[1] Creating systems that disputably, but only disputably, deserve treatment similar to that of ordinary humans generates a catastrophic moral dilemma: Either give them the full rights they arguably deserve, and risk sacrificing real human interests for systems that might not have interests worth the sacrifice; or don't give them the full rights they arguably deserve, and risk perpetrating grievous moral wrongs against entities that might be our moral equals.

I'd be stunned if this advice were universally heeded. Almost certainly, if technological process continues, and maybe soon (123), we will create morally confusing AI systems. My thought today is: Morally confusing AI systems should have doubt-producing interfaces.

Consider two types of interface that would not be doubt-producing in my intended sense: (a.) an interface that strongly invites users to see the system as an ordinary tool without rights or (b.) an interface that strongly invites users to see the system as a moral person with humanlike rights. If we have a tool that looks like a tool, or if we have a moral person who looks like a moral person, we might potentially still be confused, but that confusion would not be the consequence of a doubt-producing interface. The interface would correctly reflect the moral standing, or lack of moral standing, of the AI system in question.[2]

A doubt-producing interface, in contrast, is one that leads, or at least invites, ordinary users to feel doubt about the system's moral standing. Consider a verbal interface. Instead of the system denying that it's conscious and has moral standing (as, for example, ChatGPT appropriately does), or suggesting that it is conscious and does have moral standing (as, for example, I found in an exchange with my Replika companion), a doubt-producing AI system might say "experts have different opinions about my consciousness and moral standing".

Users then might not know how to treat such a system. While such doubts might be unsettling, feeling unsettled and doubtful would be the appropriate response to what is, in fact, a doubtful and unsettling situation.

There's more to doubt-prevention and doubt-production, of course, than explicit statements about consciousness and rights. For example, a system could potentially be so humanlike and charismatic that ordinary users fall genuinely in love with it -- even if, in rare moments of explicit conversation about consciousness and rights the system denies that it has them. Conversely, even if a system with consciousness and humanlike rights is designed to assert that it has consciousness and rights, if its verbal interactions are bland enough ("Terminate all ongoing processes? Y/N") ordinary users might remain unconvinced. Presence or absence of humanlike conversational fluency and emotionality can be part of doubt prevention or production.

Should the system have a face? A cute face might tend to induce one kind of reaction, a monstrous visage another reaction, and no face at all still a different reaction. But such familiar properties might not be quite what we want, if we're trying to induce uncertainty rather than "that's cute", "that's hideous", or "hm, that's somewhere in the middle between cute and hideous". If the aim is doubt production, one might create a blocky, geometrical face, neither cute nor revolting, but also not in the familiar middle -- a face that implicitly conveys the fact that the system is an artificial thing different from any human or animal and about which it's reasonable to have doubts, supported by speech outputs that say the same.

We could potentially parameterize a blocky (inter)face in useful ways. The more reasonable it is to think the system is a mere nonconscious tool, the simpler and blockier the face might be; the more reasonable it is to think that the system has conscious full moral personhood, the more realistic and humanlike the face might be. The system's emotional expressiveness might vary with the likelihood that it has real emotions, ranging from a simple emoticon on one end to emotionally compelling outputs (e.g., humanlike screaming) on the other. Cuteness might be adjustable, to reflect childlike innocence and dependency. Threateningness might be adjusted as it becomes likelier that the system is a moral agent who can and should meet disrespect with revenge.

Ideally, such an interface would not only produce appropriate levels of doubt but also intuitively reveal to users the grounds or bases of doubt. For example, suppose the AI's designers knew (somehow) that the system was genuinely conscious but also that it never felt any positive or negative emotion. On some theories of moral standing, such an entity -- if it's enough like us in other respects -- might be our full moral equal. Other theories of moral standing hold that the capacity for pleasure and suffering is necessary for moral standing. We the designers, let's suppose, do not know which moral theory is correct. Ideally, we could then design the system to make it intuitive to users that the system really is genuinely conscious but never experiences any pleasure or suffering. Then the users can apply their own moral best judgment to the case.

Or suppose that we eventually (somehow) develop an AI system that all experts agree is conscious except for experts who (reasonably, let's stipulate) hold that consciousness requires organic biology and experts who hold that consciousness requires an immaterial soul. Such a system might be designed so that its nonbiological, mechanistic nature is always plainly evident, while everything else about the system suggests consciousness. Again, the interface would track the reasonable grounds for doubt.

If the consciousness and moral standing of an AI system is reasonably understood to be doubtful by its designers, then that doubt ought to be passed to the system's users, intuitively reflected in the interface. This reduces the likelihood misleading users into overattributing or underattributing moral status. Also, it's respectful to the users, empowering them to employ their own moral judgment, as best they see fit, in a doubtful situation.

[R2D2 and C3P0 from Star Wars (source). Assuming they both have full humanlike moral standing, R2D2 is insufficiently humanlike in its interface, while C3P0 combines a compelling verbal interface with inadequate facial display. If we wanted to make C3P0 more confusing, we could downgrade his speech, making him sound more robotic (e.g., closer to sine wave) and less humanlike in word choice.]

------------------------------------------------

[1] For simplicity, I assume that consciousness and moral standing travel together. Different and more complex views are of course possible.

[2] Such systems would conform to what Mara Garza and I have called the Emotional Alignment Design Policy, according to which artificial entities should be designed so as to generate emotional reactions in users that are appropriate to the artificial entity's moral standing. Jeff Sebo and I are collaborating on a paper on the Emotional Alignment Design Policy, and some of the ideas of this post have been developed in conversation with him.

Wednesday, November 27, 2024

Unified vs. Partly Disunified Reasoners

I've been thinking recently about partly unified conscious subjects (e.g., this paper in draft with Sophie R. Nelson). I've also been thinking a bit about how chains of logical reasoning depend on the unity of the reasoning subject. If I'm going to derive "P & Q" from premises "P" and "Q" I must be unified as reasoner, at least to some degree. (After all, if Person 1 holds "P" and Person 2 holds "Q", "P & Q" won't be inferred.) Today, in an act of exceptional dorkiness (even for me), I'll bring these two threads together.

Suppose that {P1, P2, P3, ... Pn} is a set of propositions that a subject -- or more precisely, at least one part of a partly unified rational system -- would endorse without need of reasoning. The propositions are, that is, already believed. Water is wet; ice is cold; 2 + 3 = 5; Paris is the capital of France; etc. Now suppose that these propositions can be strung together in inference to some non-obvious conclusion Q that isn't among the system's previous beliefs -- the conclusion, for example, that 115 is not divisible by three, or that Jovenmar and Miles couldn't possibly have met in person last summer because Jovenmar spent the whole summer in Paris while Miles never left Riverside.

Let's define a fully unified reasoner as a reasoner capable of combining any elements from the set of propositions they believe {P1, P2, P3, ... Pn} in a single act of reasoning to validly derive any conclusion Q that follows deductively from {P1, P2, P3, ... Pn}. (This is of course an idealization. Fermat's Last Theorem follows from premises we all believe, but few of us could actually derive it.) In other words, any subset of {P1, P2, P3, ... Pn} could jointly serve as premises in an episode of reasoning. For example, if P2, P6, and P7 jointly imply Q1, the unified reasoner could think "P2, P6, P7, ah yes, therefore Q1!" If P3, P6, and P8 jointly imply Q2, the unified reasoner could also think "P3, P6, P8, therefore Q2."

A partly unified reasoner, in contrast, is capable only of combining some subsets of {P1, P2, P3, ... Pn}. Thus, not all conclusions that deductively follow from {P1, P2, P3, ... Pn} will be available to them. For example, the partly unified reasoner might be able to combine any of {P1, P2, P3, P4, P5} or any of {P4, P5, P6, P7, P8} while being unable to combine in reasoning any elements from P1-3 with any elements from P6-8. If Q3 follows from P1, P4, and P5, no problem, they can derive that. Similarly if Q4 follows from P5, P6, and P8. But if the only way to derive Q5 is by joining P1, P4, and P7, the partly disunified reasoning system will not be able to make that inference. They cannot, so to speak, hold both P1 and P7 in the same part of their mind at the same time. They cannot join these two particular beliefs together in a single act of reasoning.

[image: A Venn diagram of a partly unified reasoner, with overlap only at P4 and P5. Q3 is derivable from propositions in the left region, Q4 from propositions in the right region, and Q5 is not derivable from either region.]

We might imagine an alien or AI case with a clean architecture of this sort. Maybe it has two mouths or two input-output terminals. If you ask the mouth or I/O terminal on the left, it says "P1, P2, P3, P4, P5, yes that's correct, and of course Q3 follows. But I'm not sure about P6, P7, P8 or Q4." If you ask the mouth or I/O terminal on the right, it endorses P4-P8 and Q4 but isn't so sure about P1-3 and Q3.

The division needn't be crudely spatial. Imagine, instead, a situational or prompt-based division: If you ask nicely, or while flashing a blue light, the P1-P5 aspect is engaged; if you ask grumpily, or while flashing a yellow light, the P4-P8 aspect is engaged. The differential engagement needn't constitute any change of mind. It's not that the blue light causes the system as a whole to come to believe, as it hadn't before, P1-P3 and to suspend judgment about P6-P8. To see this, consider what is true a neutral time, when the system isn't being queried and no lights are flashing. At that neutral time, the system simultaneously has the following pair of dispositions: to reason based on P1-P5 if asked nicely or in blue, and to reason based on P4-P8 if asked grumpily or in yellow.

Should we say that there are discretely two distinct reasoners rather than one partly unified system? At least two inconveniences for that way of thinking are: First, any change in P4 or P5 would be a change in both, with no need for one reasoner to communicate it to the other, as would normally be the case with distinct reasoners. Second, massive overlap cases -- say P1-P999 and P2-P1000 -- seem more naturally and usefully modeled as a single reasoner with a quirk (not being able to think P1 and P1000 jointly, but otherwise normal), rather than as two distinct reasoners.

But wait, we're not done! I can make it weirder and more complicated, by varying the type and degree of disunity. The simple model above assumes discrete all-or-none availability to reasoning. But we might also imagine:

(a.) Varying joint probabilities of combination. For example, if P1 enters the reasoning process, P2 might have a 87% chance of being accessed if relevant, P3 a 74% chance, ... and P8 a 10% chance.

(b.) Varying confidence. If asked in blue light, the partly disunified entity might have 95% credence in P1-P5 and 80% credence in P6-P8. If asked in yellow light, it might have 30% credence in P1-P3 and 90% credence in P4-P8.

(c.) Varying specificity. Beliefs of course don't come divided into neatly countable packages. Maybe the left side of the entity has a hazy sense that something like P8 is true. If P8 is that Paris is in France, the left side might only be able to reason on Paris is in France-or-Germany-or-Belgium. If P8 is that the color is exactly scarlet #137, the left side might only be able to reason on the color is some type of red.

Each of (a)-(c) admits of multiple degrees, so that the unity/disunity or integration/disintegration of a reasoning system is a complex, graded, multidimensional phenomenon.

So... just a bit of nerdy fun, with no actual application? Well, fun is excuse enough, I think. But still:

(1.) It's easy to imagine realistic near-future AI cases with these features. A system or network might have a core of shared representations or endorsable propositions and local terminals or agents with stored local representations not all of which are shared with the center. If we treat that AI system as a reasoner, it will be a partly unified reasoner in the described sense. (See also my posts on memory and perception in group minds.)

(2.) Real cases of dissociative identity or multiple personality disorder might potentially be modeled as involving partly disunified reasoning of this sort. Alter 1 might reason with P1-P5 and Alter 2 with P4-P8. (I owe this thought to Nichi Yes.) If so, there might not be a determinate number of distinct reasoners.

(3.) Maybe some more ordinary cases of human inconstancy or seeming irrationality can be modeled in this way: Viviana feeling religious at church, secular at work, or Brittany having one outlook when in a good, high-energy mood and a very different outlook when she's down in the dumps. While we could, and perhaps ordinarily would, model such splintering as temporal fluctuation with beliefs coming and going, a partial unity model has two advantages: It applies straightforwardly even when the person is in neither situation (e.g., asleep), and it doesn't require the cognitive equivalent of frequent erasure and rewriting of the same propositions (everything endures but some subsets cannot be simultaneously activated; see also Elga and Rayo 2021).

(4.) If there are cases of partial phenomenal (that is, experiential) unity, then we might expect there also to be cases of partial cognitive unity, and vice versa. Thus, a feasible model of the one helps increase the plausibility that there might be a feasible model of the other.

Friday, November 22, 2024

Philosophical Fame, 1890-1960

There's a fun new tool at Edhiphy. The designers pulled the full text from twelve leading philosophy journals from 1890 to 1980 and counted the occurrences of philosophers' names. (See note [1] for discussion of error rates in their method.)

Back in the early 2010s, I posted several bibliometric studies of philosophers' citation or discussion rates over time, mostly based on searches of Philosopher's Index abstracts from 1940 to the present. This new tool gives me a chance to update some of my thinking, using a different method and going further into the past.

One thing I found fascinating in my earlier studies was how some philosophers who used to be huge (for example, Henri Bergson and Herbert Spencer) are now hardly read, while others (for example, Gottlob Frege) have had more staying power.

Let's look at the top 25 most discussed philosophers from each available decade.

1890s:

1. Immanuel Kant
2. Georg Wilhelm Friedrich Hegel
3. Aristotle
4. David Hume
5. Herbert Spencer
6. William James
7. Plato
8. John Stuart Mill
9. René Descartes
10. Wilhelm Wundt
11. Hermann Lotze
12. F. H. Bradley
13. Charles Sanders Peirce
14. Buddha
15. Thomas Hill Green
16. Benedictus de Spinoza
17. Charles Darwin
18. John Locke
19. Gottfried Wilhelm Leibniz
20. Thomas Hobbes
21. Arthur Schopenhauer
22. Socrates
23. Hermann von Helmholtz
24. George Frederick Stout
25. Alexander Bain

Notes:

Only three of the twelve journals existed in the 1890s, so this is a small sample.

Philosophy and empirical psychology were not clearly differentiated as disciplines until approximately the 1910s or 1920s, and these journals covered both areas. (For example, the Journal of Philosophy was originally founded in 1904 as the Journal of Philosophy, Psychology, and Scientific Methods, shortening to the now familiar name in 1921.) Although Wundt, Helmholtz, and Stout were to some extent philosophers, they are probably better understood primarily as early psychologists. William James is of course famously claimed by both fields.

Herbert Spencer, as previously noted, was hugely influential in his day: fifth on this eminent list! Another eminent philosopher on this list (#11) who is hardly known today (at least in mainstream Anglophone circles) is Hermann Lotze.

Most of the others on the list are historical giants, plus some prominent British idealists (F. H. Bradley, Thomas Hill Green) and pragmatists (William James, Charles Sanders Peirce, Alexander Bain) and interestingly (but not representative of later decades) "Buddha". (A spot check reveals that some of these references are to Gautama Buddha or "the Buddha", while others use "buddha" in a more general sense.)

1900s:

1. Immanuel Kant
2. William James
3. Plato
4. F. H. Bradley
5. Georg Wilhelm Friedrich Hegel
6. David Hume
7. Aristotle
8. Herbert Spencer
9. Gottfried Wilhelm Leibniz
10. John Dewey
11. George Berkeley
12. John Stuart Mill
13. George Frederick Stout
14. Thomas Hill Green
15. Josiah Royce
16. Benedictus de Spinoza
17. John Locke
18. Ferdinand Canning Scott Schiller
19. Ernst Mach
20. Wilhelm Wundt
21. James Ward
22. René Descartes
23. Alfred Edward Taylor
24. Henry Sidgwick
25. Bertrand Russell

Notes:

Notice the fast rise of John Dewey (1859-1952), to #10 (#52 in the 1890s list). Other living philosophers in the top ten were James (1842-1910), Bradley (1846-1824), and for part of the period Spencer (1820-1903).

It's also striking to see George Berkeley enter the list so high (#11, compared to #28 in the 1890s) and Descartes fall so fast despite his continuing importance later (from #9 to #22). This could be statistical noise due to the small number of journals, or it could reflect historical trends. I'm not sure.

Our first "analytic" philosopher appears: Bertrand Russell (1872-1970) at #25. He turned 33 in 1905, so he found eminence very young for a philosopher.

Lotze has already fallen off the list (#29 in the 1900s; #29 in the 1910s; #63 in the 1930s, afterwards not in the top 100).

1910s:

1. Henri Bergson
2. Bertrand Russell
3. Immanuel Kant
4. Plato
5. William James
6. Gottfried Wilhelm Leibniz
7. Aristotle
8. Socrates
9. Bernard Bosanquet
10. George Berkeley
11. F. H. Bradley
12. Georg Wilhelm Friedrich Hegel
13. René Descartes
14. Josiah Royce
15. David Hume
16. Isaac Newton
17. John Dewey
18. Friedrich Nietzsche
19. Ferdinand Canning Scott Schiller
20. Arthur Schopenhauer
21. John Locke
22. Benedictus de Spinoza
23. Edwin Holt
24. Isaac Barrow
25. Johann Gottlieb Fichte
Notes:

Henri Bergson (1859-1941) debuts at #1! What a rock star. (He was #63 in the 1900s list.) We forget how huge he was in his day. Russell, who so far has had much more durable influence, rockets up to #2. It's also interesting to see Bernard Bosanquet (1848-1923), who is now little read in mainstream Anglophone circles, at #9.

Josiah Royce is also highly mentioned in this era (#14 in this list, #15 in the 1900s list), despite not being much read now. F.C.S. Schiller (1864-1937) is a similar case (#19 in this list, #18 in the 1900s list).

1920s:

1. Immanuel Kant
2. Plato
3. Aristotle
4. Bernard Bosanquet
5. Georg Wilhelm Friedrich Hegel
6. F. H. Bradley
7. Bertrand Russell
8. Benedictus de Spinoza
9. William James
10. Socrates
11. John Dewey
12. Alfred North Whitehead
13. David Hume
14. George Santayana
15. René Descartes
16. Henri Bergson
17. Albert Einstein
18. C. D. Broad
19. John Locke
20. Gottfried Wilhelm Leibniz
21. George Berkeley
22. Isaac Newton
23. James Ward
24. Samuel Alexander
25. Benedetto Croce

Notes:

I'm struck by how the 1920s returns to the classics at the top of the list, with Kant, Plato, and Aristotle as #1, #2, and #3. Bergson is already down to #16 and Russell has slipped to #7. Most surprising to me, though, is Bosanquet at #4! What?!

1930s:

1. Immanuel Kant
2. Plato
3. Aristotle
4. Benedictus de Spinoza
5. Georg Wilhelm Friedrich Hegel
6. René Descartes
7. Alfred North Whitehead
8. Bertrand Russell
9. David Hume
10. John Locke
11. George Berkeley
12. Socrates
13. Friedrich Nietzsche
14. Rudolf Carnap
15. William James
16. Gottfried Wilhelm Leibniz
17. John Dewey
18. Isaac Newton
19. Clarence Irving Lewis
20. Arthur Oncken Lovejoy
21. Albert Einstein
22. Charles Sanders Peirce
23. F. H. Bradley
24. Ludwig Wittgenstein
25. Bernard Bosanquet

Notes:

Nietzsche rises suddenly (#13; vs #56 in the 1920s list). Wittgenstein also cracks the list at #24 (not even in the top 100 in the 1920s).

With the exception of Whitehead, top of the list looks like what early 21st century mainstream Anglophone philosophers tend to perceive as the most influential figures in pre-20th-century Western philosophy (see, e.g., Brian Leiter's 2017 poll). The 1930s, perhaps, were for whatever reason a decade more focused on the history of philosophy than on leading contemporary thinkers. (The presence of historian of ideas Arthur Lovejoy [1873-1962] at #20 further reinforces that thought.)

1940s:

1. Immanuel Kant
2. Alfred North Whitehead
3. Aristotle
4. Plato
5. Bertrand Russell
6. John Dewey
7. David Hume
8. William James
9. George Berkeley
10. Charles Sanders Peirce
11. René Descartes
12. Benedictus de Spinoza
13. Edmund Husserl
14. Georg Wilhelm Friedrich Hegel
15. Gottfried Wilhelm Leibniz
16. Thomas Aquinas
17. Socrates
18. Rudolf Carnap
19. Martin Heidegger
20. G. E. Moore
21. John Stuart Mill
22. Isaac Newton
23. Søren Kierkegaard
24. A. J. Ayer
25. John Locke

Notes:

Oh, how people loved Whitehead (#2) in the 1940s!

Edmund Husserl (1859-1938) makes a posthumous appearance at #13 (#31 in the 1920s) and Heidegger (1889-1976) at #19 (#97 in the 1920s), suggesting an impact of Continental phenomenology. I suspect this is due to the inclusion of Philosophy and Phenomenological Research in the database starting 1940. Although the journal is now a bastion of mainstream Anglophone philosophy, in its early decades it included lots of work in Continental phenomenology (as the journal's title suggests).

The philosophers we now think of as the big three American pragmatists have a very strong showing in the 1940s, with Dewey at #6, James at #8, and Peirce at #10.

Thomas Aquinas makes his first and only showing (at #16), suggesting that Catholic philosophy is having more of an impact in this era.

We're also starting to see more analytic philosophers, with G. E. Moore (1873-1958), and A. J. Ayer (1910-1989) now making the list, in addition to Russell and Carnap (1891-1970).

Wittgenstein, surprisingly to me, has fallen off the list all the way down to #73 -- perhaps suggesting that if he hadn't had his second era, his earlier work would have been quickly forgotten.

1950s:

1. Immanuel Kant
2. Plato
3. Aristotle
4. Bertrand Russell
5. David Hume
6. Gilbert Ryle
7. G. E. Moore
8. Willard Van Orman Quine
9. George Berkeley
10. Georg Wilhelm Friedrich Hegel
11. John Dewey
12. Alfred North Whitehead
13. Rudolf Carnap
14. Ludwig Wittgenstein
15. René Descartes
16. John Locke
17. Clarence Irving Lewis
18. Socrates
19. John Stuart Mill
20. Gottfried Wilhelm Leibniz
21. Gottlob Frege
22. A. J. Ayer
23. William James
24. Edmund Husserl
25. Nelson Goodman

By the 1950s, the top eight are four leading historical figures -- Kant, Plato, Aristotle, and Hume -- and four leading analytic philosophers: Russell, Gilbert Ryle (1900-1976), G. E. Moore, and W. V. O. Quine (1908-2000). Neither Ryle nor Quine were among the top 100 in 1940s, so their rise to #6 and #8 was sudden.

Gottlob Frege (1848-1925) also makes his first, long-posthumous appearance.

1960s:

1. Aristotle
2. Immanuel Kant
3. Ludwig Wittgenstein
4. David Hume
5. Plato
6. René Descartes
7. P. F. Strawson
8. Willard Van Orman Quine
9. Bertrand Russell
10. J. L. Austin
11. John Dewey
12. Rudolf Carnap
13. Edmund Husserl
14. Socrates
15. Norman Malcolm
16. G. E. Moore
17. Gottlob Frege
18. Georg Wilhelm Friedrich Hegel
19. George Berkeley
20. R. M. Hare
21. John Stuart Mill
22. Gilbert Ryle
23. A. J. Ayer
24. Karl Popper
25. Carl Gustav Hempel

Wittgenstein is back with a vengeance at #3. Other analytic philosophers, in order, are P. F. Strawson, Quine, Russell, Austin, Carnap, Norman Malcolm (1911-1990), Moore, Frege, R. M. Hare (1919-2002), Ryle, Ayer, Karl Popper (1902-1994), and Carl Hempel (1905-1997).

Apart from pre-20th-century historical giants, it's all analytic philosophers, except for Dewey and Husserl.

Finally, the 1970s:

1. Willard Van Orman Quine
2. Immanuel Kant
3. David Hume
4. Aristotle
5. Ludwig Wittgenstein
6. Plato
7. John Locke
8. René Descartes
9. Karl Popper
10. Rudolf Carnap
11. Gottlob Frege
12. Edmund Husserl
13. Hans Reichenbach
14. Socrates
15. P. F. Strawson
16. Donald Davidson
17. John Stuart Mill
18. Bertrand Russell
19. Thomas Reid
20. Benedictus de Spinoza
21. Nelson Goodman
22. Carl Gustav Hempel
23. John Rawls
24. Karl Marx
25. Saul Kripke

With the continuing exception of Husserl, the list is again historical giants plus analytic philosophers. Interesting to see Marx enter at #24. Hans Reichenbach (1891-1953) has a strong debut at #13. Ryle's decline is striking, from #6 in the 1950s to #22 in the 1960s to off the list at #51 in the 1970s.

At the very bottom of the list, #25, we see the first "Silent Generation" philosopher: Saul Kripke (1940-2022). In a recent citation analysis of the Stanford Encyclopedia of Philosophy, I found that the Silent Generation has so far had impressive overall influence and staying power in mainstream Anglophone philosophy. It would be interesting to see if this influence continues.

The only philosopher born after 1800 who makes both the 1890s and the 1970s top 25 is John Stuart Mill. Peirce and James still rank among the top 100 in the 1970s (#58 and #86). None of the other stars of the 1890s -- Spencer, Herbert, Lotze, Bradley, Green -- are still among the top 100 by the 1970s, and I think it's fair to say they are hardly read except by specialists.

Similar remarks apply to most of the stars of the 1900s, 1910s, and 1920s: Bergson, Bosanquet, Royce, Schiller, C. D. Broad, and George Santayana are no longer widely read. Two exceptions are Russell, who persists in the top 25 through the 1970s, and Dewey who falls from the top 25 but still remains in the top 100, at #87.

Also, in case you didn't notice: no women or people of color (as we would now classify them) appear on any of these lists, apart from "Buddha" in the 1890s.

In my recent Stanford Encyclopedia of Philosophy analysis, the most-cited living philosophers were Timothy Williamson, Martha Nussbaum, Thomas Nagel, Frank Jackson, John Searle, and David Chalmers. However, none of them is probably as dominant now as Spencer, James, Bradley, Russell, Bosanquet, and Bergson were at the peak of their influence.

---------------------------------------

[1] The Edhiphy designers estimate "82%-91%" precision, but I'm not sure what that means. I'd assume that "Wittgenstein" and "Carnap" would hit with almost 100% precision. Does it follow others might be as low as 40%? There certainly are some problems. I noticed, for example, that R. Jay Wallace, born in 1957, has 78 mentions in the 1890s. I spot checked "Russell", "Austin", "James", and "Berkeley", finding only a few false positives for Russell and Austin (e.g., misclassified references to legal philosopher John Austin). I found significantly more false positives for William James (including references to Henry James and some authors with the first name James, such as psychologist James Ward), but still probably not more than 10%. For "Berkeley" there were a similar number of false positives referencing the university or city. I didn't attempt to check for false negatives.

[Bosanquet and Bergson used to be hugely influential]

Tuesday, November 19, 2024

New in Draft: When Counting Conscious Subjects, the Result Needn't Always Be a Determinate Whole Number

(with Sophie R. Nelson)

One philosophical inclination I shared with the late Dan Dennett is a love of weird perspectives on consciousness, which sharply violate ordinary, everyday common sense. When I was invited to contribute to a special issue of Philosophical Psychology in his memory, I thought of his intriguing remark in Consciousness Explained against "the myth of selves as brain-pearls, particular, concrete, countable things", lamenting people's stubborn refusal "to countenance the possibility of quasi-selves, semi-selves, transitional selves" (1991, p. 424-425). As I discussed in a blog post in June, Dennett's "fame in the brain" view of consciousness naturally suggests that consciousness won't always come in discrete, countable packages, since fame is a gradable, multidimensional phenomenon, with lots of gray area and partial overlap.

So I contacted Sophie R. Nelson, with whom I'd published a paper last year on borderline cases of group minds, and we decided to generalize the idea. On a broad range of naturalistic, scientific approaches to consciousness, we ought to expect that conscious subjects needn't always come in determinate, whole number packages. Sometimes, the number of conscious subjects in an environment should be either indeterminate, or a determinate non-whole number, or best modeled by some more complicated mathematical representation. If some of us have commonsense intuitions to the contrary, such intuitions aren't probative.

Our submission is due November 30, and comments are (as always) very welcome -- either before or after the Nov 30 deadline (since we expect at least one round of revisions).

Abstract:

Could there be 7/8 of a conscious subject, or 1.34 conscious subjects, or an entity indeterminate between being one conscious subject and seventeen? Such possibilities might seem absurd or inconceivable, but our ordinary assumptions on this matter might be radically mistaken. Taking inspiration from Dennett, we argue that, on a wide range of naturalistic views of consciousness, the processes underlying consciousness are sufficiently complex to render it implausible that conscious subjects must always arise in determinate whole numbers. Whole-number-countability might be an accident of typical vertebrate biology. We explore several versions of the inconceivability objection, suggesting that the fact that we cannot imagine what it’s like to be 7/8 or 1.34 or an indeterminate number of conscious subjects is no evidence against the possibility of such subjects. Either the imaginative demand is implicitly self-contradictory (imagine the one, determinate thing it’s like to be an entity there isn’t one, determinate thing it’s like to be) or imaginability in the relevant sense isn’t an appropriate test of possibility (in the same way that the unimaginability, for humans, of bat echolocation experiences does not establish that bat echolocation experiences are impossible).

Full draft here.

[Figure 2 from Schwitzgebel and Nelson, in draft: An entity intermediate or indeterminate between one and three conscious subjects. Solid circles represent determinately conscious mental states. Dotted lines represent indeterminate or intermediate unity among those states.]

Friday, November 15, 2024

Three Models of the Experience of Dreaming: Phenomenal Hallucination, Imagination, and Doxastic Hallucination

What are dreams like, experientially?

One common view is that dreams are like hallucinations. They involve sensory or sensory-like experiences just as if, or almost as if, you were in the environment you are dreaming you are in. If you dream of being Napoleon on the fields of Waterloo, taking in the sights and sounds, then you have visual and auditory experiences much like Napoleon might have had in the same position (except perhaps irrational, bizarre, or otherwise different in specific content). This is probably the predominant view among dream researchers (e.g., Hobson and Revonsuo).

Another view, less common but intriguing, is that dreams are like imaginings. Dreaming you are Napoleon on the fields of Waterloo is like imagining or "daydreaming" that you're there. The experience isn't sensory but imagistic (e.g., Ichikawa and Sosa).

These views are very different!

For example, look at your hands. Now close your eyes and imagine looking at your hands. Unless you're highly unusual, you will probably agree that the first experience is very different from the second experience. On the hallucination model of dreams, dream experience is more like the first (sensory) experience. On the imagination model, dream experience is more like the second (imagery) experience. On pluralist models, dream experiences are sometimes like the one, sometimes like the other (e.g., Rosen and possibly Windt's nuanced version of the hallucination model). (Unfortunately, proponents of the hallucination model sometimes confusingly talk about dream "imagery".)

-----------------------------------

I confess to being tempted to the imagination model. My reason is primarily introspective or immediately retrospective. I sometimes struggle with insomnia and it's not unusual for me to drift in and out of sleep, including lying quietly in bed, eyes closed, allowing myself to drift in daydream, which seems sometimes to merge into sleep, then back into daydream, and my immediately remembered dreams seem not so radically different from my eyes-closed daydream imaginations. (Ichikawa describes similar experiences.)

Another consideration is this: Plausibly, the stability and detail of our ordinary sensory experiences depend to a substantial extent on the stabilizing influence of external inputs. It appears both to match my own experience and to be neurophysiologically plausible that the finely detailed, vivid, sharp structure, of say, visual experience, would be difficult for my brain to sustain without the constraint of a rich flow of input information.  (Alva Noë makes a similar point.)

Now, I don't put a lot of stock in these reflections. There's reason to be skeptical of the accuracy of introspective reports in general, and perhaps dream reports in particular, and I'm willing to apply my own skepticism to myself. But by the same token, what is the main evidence on the other side, in favor of the hallucination model? Mainly, again, introspective report. In particular, it's the fact that people often report their dream experiences as having the rich, sensory-like detail that the hallucination model predicts. Of course, we could just take the easy, obvious, pluralist path of saying that everyone is right about their own experiences. But what fun is that?

-----------------------------------

In fact, I'm inclined to throw a further wrench in things by drawing a distinction between two types of hallucination: phenomenal and doxastic. I introduced this distinction in a blog post in 2013, after reading Oliver Sacks's Hallucinations.

Consider this description, from page 99 of Hallucinations:

The heavens above me, a night sky spangled with eyes of flame, dissolve into the most overpowering array of colors I have ever seen or imagined; many of the colors are entirely new -- areas of the spectrum which I seem to have hitherto overlooked. The colors do not stand still, but move and flow in every direction; my field of vision is a mosaic of unbelievable complexity. To reproduce an instant of it would involve years of labor, that is, if one were able to reproduce colors of equivalent brilliance and intensity.

Here are two ways in which you might come to believe the above about your experience:

(1.) You might actually have visual experiences of the sort described, including of colors entirely new and previously unimagined and of a complexity that would require years of labor to describe.

Or

(2.) you might shortcut all that and simply arrive straightaway at the belief that you are undergoing or have undergone such an experience -- perhaps with the aid of some unusual visual experiences, but not really of the novelty and complexity described.

If the former, you have phenomenally hallucinated wholly novel colors. If the latter, you have only doxastically hallucinated them. I expect that I'm not the first to suggest such a distinction among types of hallucination, but I haven't yet found a precedent.

Mitchell-Yellin and Fischer suggest that some "near death experiences" might also be doxastic hallucinations of this sort. Did your whole life really flash before your eyes in that split second during an auto accident, or did you only form the belief in that experience without the actual experience itself? It's not very neurophysiologically plausible that someone would experience hundreds or thousands of different memory experiences in 500 milliseconds.

-----------------------------------

It seems clear from dream researchers' descriptions of the hallucination model of dreams that they have phenomenal hallucination in mind. But what if dream experiences involve, instead or at least sometimes, doxastic rather than phenomenal hallucinations?

Here, then, is a possibility about dream experience: If I dream I am Napoleon, standing on the fields of Waterloo, I have experiences much like the experiences I have when I merely imagine, in daydream, that I am standing on the fields of Waterloo. But sometimes a doxastic hallucination is added to that imagination: I form the belief that I am having or had rich sensory visual and auditory experience. This doxastic hallucination would explain reports of rich, vivid, detailed sensory-like dream experience without requiring the brain actually to concoct rich, vivid, and detailed visual and auditory experiences.

Indeed, if we go full doxastic hallucination, even the imagination-like experiences would be optional.  (Also, if -- following Sosa -- we don't genuinely believe things while dreaming, we could reframe doxastic hallucinations in terms of whatever quasi-belief analogs occur during dreams.)

[The battle at Waterloo: image source]

Monday, November 11, 2024

New in Draft: The Copernican Argument for Alien Consciousness; The Mimicry Argument Against Robot Consciousness

(with Jeremy Pober)

Over the past several years, I've posted a few times on what I call the "Copernican Argument" for thinking that behaviorally sophisticated space aliens would be conscious, even if they are constituted very differently from us (here, here, here, here). I've also posted a few times on what I call the "Mimicry Argument" against attributing consciousness to AI systems or robots that were designed to mimic the superficial signs of human consciousness (including current Large Language Models like ChatGPT and Claude) (here, here, here).

Finally, I have a circulatable paper in draft that deals with these issues, written in collaboration with Jeremy Pober, and tested with audiences at Trent University, Harvey Mudd, New York University, the Agency and Intentions in AI conference in Göttingen, Jagiellonian University, the Oxford Mind Seminar, University of Lisbon, NOVA Lisbon University, University of Hamburg, and the Philosophy of Neuroscience/Mind Writing Group.

It's a complicated paper! Several philosophers have advised me that the Copernican Argument is one paper and the Mimicry Argument is another. Maybe they are right. But I also think that there's a lot to be gained from advancing these arguments side by side: Each shines light on the boundaries of the other. The result, though intricate, is I hope not too intricate for evaluation and comprehensibility. (I might still change my mind about that.)


Abstract:

On broadly Copernican grounds, we are entitled to default assume that apparently behaviorally sophisticated extraterrestrial entities (“aliens”) would be conscious. Otherwise, we humans would be inexplicably, implausibly lucky to have consciousness, while similarly behaviorally sophisticated entities elsewhere would be mere shells, devoid of consciousness. However, this Copernican default assumption is canceled in the case of behaviorally sophisticated entities designed to mimic superficial features associated with consciousness in humans (“consciousness mimics”), and in particular a broad class of current, near-future, and hypothetical robots. These considerations, which we formulate, respectively, as the Copernican and Mimicry Arguments, jointly defeat an otherwise potentially attractive parity principle, according to which we should apply the same types of behavioral or cognitive tests to aliens and robots, attributing or denying consciousness similarly to the extent they perform similarly. Instead of grounding speculations about alien and robot consciousness in metaphysical or scientific theories about the physical or functional bases of consciousness, our approach appeals directly to the epistemic principles of Copernican mediocrity and inference to the best explanation. This permits us to justify certain default assumptions about consciousness while remaining to a substantial extent neutral about specific metaphysical and scientific theories.

Full paper here.


As always, questions/comments/objections welcome here on the blog, on my social media accounts, or by email to my UCR address.

[image source]

Wednesday, October 30, 2024

The Ethics of Harmonizing with the Dao

Reading the ancient Chinese philosophers Xunzi and Zhuangzi, I am inspired to articulate an ethics of harmonizing with the dao (the "way"). This ethics doesn't quite map onto any of the three conceptualizations of ethics that are standard in Western philosophy (consequentialism, deontology, and virtue ethics), nor is it exactly a "role ethics" of the sort sometimes attributed to ancient Confucians.

Xunzi

The ancient Confucian Xunzi articulates a vision of the world in which Heaven, Earth, and humanity operate in harmony:

Heaven has its proper seasons,
Earth has its proper resources,
And humankind has its proper order,
-- this is called being able to form a triad
(Ch 17, l. 34-37; Hutton trans. 2014, p. 176).

Heaven (tian, literally the sky, but with strong religious associations) and Earth are jointly responsible for what we might now call the "laws of nature" and all "natural" phenomena -- including, for example, the turning of the seasons, the patterns of wind and rain, the tendency for plants and animals to thrive under certain conditions and wither under other conditions. Also belonging to these natural phenomena are the raw materials with which humans work: not only the raw materials of wood, metal, and fiber, but also the raw material of natural human inclinations: our tendency to enjoy delicious tastes, our tendency to react angrily to provocations, our general preference for kin over strangers.

Xunzi views humanity's task as creating the third corner of a triad with Heaven and Earth by inventing customs and standards of proper behavior that allow us to harmonize with Heaven and Earth, and with each other. For example, through trial and error, our ancestors learned the proper times and methods for sowing and reaping, how to regulate flooding rivers, how to sharpen steel and straighten wood, how to make pots that won't leak, how to make houses that won't fall over, and so on. Our ancestors also -- again through trial and error -- learned the proper rituals and customs and standards of behavior that permit people to coexist harmoniously with each other without chaotic conflict, without excessive or inappropriate emotions, and with an allocation of goods that allow all to flourish according to their status and social role.

Following the dao can be conceptualized for Xunzi, then, as aligning harmoniously into this triad. Abide by the customs and standards of behavior that contribute to the harmonious whole, in which crops are properly planted, towns are properly constructed, the crafts flourish, and humans thrive in an orderly society.

Each of us has a different role, in accord with the proper customs of a well-ordered society: the barley farmer has one role, the soldier another role, the noblewoman yet another, the traveling merchant yet another. It's not unreasonable to view Xunzi's ethics as a kind of role ethics, according to which the fundamental moral principle is that one adheres to one's proper role in society. It's also not unreasonable to think of the customs and standards of proper behavior as a set of rules to which one ought to adhere (those rules applying in different ways according to one's position in society), and thus to view Xunzi's ethics as a kind of deontological (rule-based) ethics. However, there might also be room to interpret harmonious alignment with the dao as the most fundamental feature of ethical behavior. Adherence to one's role and to the proper traditional customs and practices, on this interpretation of Xunzi, would be only derivatively good, because doing so typically constitutes harmonious alignment.

A test case is to imagine, through Xunzi's eyes, whether a morally well-developed sage might be ethically correct sometimes to act contrary to their role and to the best traditional standards of good behavior, if they correctly see that by doing so they contribute better to the overall harmony of Heaven, Earth, and humankind. I'm tempted to think that Xunzi would indeed permit this -- though only very cautiously, since he is pessimistic about the moral wisdom of ordinary people -- and thus that for him harmonious alignment with the dao is more fundamental than roles and rules. However, I'm not sure I can find direct textual support in favor of this interpretation; it's possible I'm being overly "charitable".

[image source]

A Zhuangzian Correction

A Xunzian ethics of this sort is, I think, somewhat attractive. But it is also deeply traditionalist and conformist in a way I find unappealing. It could use a Zhuangzian twist -- and the idea of "harmonizing with the dao" is at least as Zhuangzian (and "Daoist") as it is Confucian.

Zhuangzi imagines a wilder, more wondrous cosmos than Xunzi's neatly ordered triad of Heaven, Earth, and humankind -- symbolized (though it's disputable how literally) by people so enlightened that they can walk without touching the ground; trees that count 8000 years as a single autumn; gracious emperors with no eyes, ears, nose, or mouth; people with skin like frost who live by drinking dew; enormous, useless trees who speak to us in dreams; and more. This is the dao, wild beyond human comprehension, with which Zhuangzi aims to harmonize.

There are, I think, in Zhuangzi's picture -- though he would resist any effort to fully capture it in words -- ways of flowing harmoniously along with this wondrous and incomprehensible dao and ways of straining unproductively against it. One can be easygoing and open-minded, welcome surprise and difference, not insist on jamming everything into preconceived frames and plans; and one can contribute to the delightful weirdness of the world in one's own unique way. This is Zhuangzian harmony. You become a part of a world that is richer and more wondrous because it contains you, while allowing other wonderful things to also naturally unfold.

In a radical reading of Zhuangzi, ethical obligations and social roles fall away completely. There is little talk in Zhuangzi's Inner Chapters, for example, of our obligation to support others. I don't know that we have to read Zhuangzi radically; but regardless of that question of interpretation, I suggest that there's an attractive middle between Xunzi's conventionalism and Zhuangzi's wildness. Each can serve as a corrective to the other.

In the ethical picture that emerges from this compromise, we each contribute uniquely to a semi-ordered cosmos, participating in social harmony, but not rigidly -- also transcending that harmony, breaking rules and traditions for the better, making the world richer and more wondrous, each in our diverse ways, while also supporting others who contribute in their different ways, whether those others are human, animal, plant, or natural phenomena.

Contrasts

This is not a consequentialist ethics: It is not that our actions are evaluated in terms of the good or bad consequences they have (and still less that the actions are evaluated by a summation of the good minus the bad consequences). Instead, harmonizing with the dao is to participate in something grand, without need of a further objective. Like the deontologist, Xunzi and Zhuangzi and my imagined compromise philosopher needn't think that right or harmonious action will always have good long-term results. Nor is it a deontological or role ethics: There is no set of rules one must always follow or some role one must always adhere to. Nor is it a virtue ethics: There is no set of virtues to which we all must aspire or a distinctive pattern of human flourishing that constitutes the highest attainment. We each contribute in different ways -- and if some virtues often prove to be important, they are derivatively important in the same way that rules and roles can be derivatively important. They are important only because, and to the extent, having those virtues enables or constitutes one's contribution to the magnificent web of being.

So although there are resonances with the more pluralistic forms of consequentialism, and virtue ethics, and role ethics, and even deontology (trivially or degenerately, if the rule is just "harmonize with the dao"), the classical Chinese ethical ideal of harmonizing with the dao differs somewhat from all of these familiar (to professional philosophers) Western ethical approaches.

Many of these other approaches also contain an implicit intellectualism or elitism, in which ideal ethical goodness requires intellectual attainment: wisdom, or a sophisticated ability to weigh consequences or evaluate and apply rules -- far beyond, for example, the capacities of someone with severe cognitive disabilities. With enough Zhuangzi in the mix, such elitism evaporates. A severely cognitively disabled person, or a magnificently weird nonhuman animal, might far exceed any ordinary adult philosopher in their capacity to harmonize with the dao and might contribute more to the rich tapestry of the world.

Perhaps an ethics of harmonizing with the dao can resonate with some 21st-century Anglophone readers, despite its origins in ancient China. It is not, I think, as alien as it might seem from its reliance on the concept of dao and its failure to fit into the standard ethical triumvirate of consequentialism, deontology, and virtue ethics. The fundamental idea should be attractive to some: We each contribute by instantiating a unique piece of a magnificent world, a world which would be less magnificent without us.

Tuesday, October 22, 2024

An Objection to Chalmers's Fading Qualia Argument

[Note: This is a long and dense post. Buckle up.]

In one chapter of his influential 1996 book, David Chalmers defends the view that consciousness arises in virtue of the functional organization of the brain rather than in virtue of the brain's material substrate.  That is, if there were entities that were functionally/organizationally identical to humans but made out of different stuff (e.g. silicon chips), they would be just as conscious as we are.  He defends this view, in part, with what he calls the Fading Qualia Argument.  The argument is enticing, but I think it doesn't succeed.

Chalmers, Robot, and the Target Audience of the Argument

Drawing on thought experiments from Pylyshyn, Savitt, and Cuda, Chalmers begins by imagining two cases: himself and "Robot".  Robot is a functional isomorph of Chalmers, but constructed of different materials.  For concreteness (but this isn't essential), we might imagine that Robot has a brain with the exact same neural architecture as Chalmers' brain, except that the neurons are made of silicon chips.

Because Chalmers and Robot are functional isomorphs, they will respond in the same way to all stimuli.  For example, if you ask Robot if it is conscious, it will emit, "Yes, of course!" (or whatever Chalmers would say if asked that question).  If you step on Robot's toe, Robot will pull its foot back and protest.  And so on.

For purposes of this argument, we don't want to assume that Robot is conscious, despite its architectural and functional similarity to Chalmers.  The Fading Qualia Argument aims to show that Robot is conscious, starting from premises that are neutral on the question.  The aim is to win over those who think that maybe being carbon-based or having certain biochemical properties is essential for consciousness, so that a functional isomorph made of the wrong stuff would only misleadingly look like it's conscious.  The target audience for this argument is someone concerned that for all Robot's similar mid-level architecture and all of its seeming "speech" and "pain" behavior, Robot really has no genuinely conscious experiences at all, in virtue of lacking the right biochemistry -- that it's merely a consciousness mimic, rather than a genuinely conscious entity.

The Slippery Slope of Introspection

Chalmers asks us to imagine a series of cases intermediate between him and Robot.  We might imagine, for example, a series each of whose members differs by one neuron.  Entity 0 is Chalmers.  Entity 1 is Chalmers with one silicon chip neuron replacing a biological neuron.  Entity 2 is Chalmers with two silicon chip neurons replacing two biological neurons.  And so on to Entity N, Robot, all of whose neurons are silicon.  Again, the exact nature of the replacements isn't essential to the argument.  The core thought is just this: Robot is a functional isomorph of Chalmers, but constructed of different materials; and between Chalmers and Robot we can construct a series of cases each of which is only a tiny bit different from its neighbors.

Now if this is a coherent setup, the person who wants to deny consciousness to Robot faces a dilemma.  Either (1.) at some point in the series, consciousness suddenly winks out -- between Entity I and Entity I+1, for some value of I.  Or (2.) consciousness somehow slowly fades away in the series.

Option (1) seems implausible.  Chalmers, presumably, has a rich welter of conscious experience (at least, we can choose a moment at which he does).  A priori, it would be odd if the big metaphysical jump from that rich welter of experience to zero experience would occur with an arbitrarily tiny change between Entity I and Entity I+1.  And empirically, our best understanding of the brain is that tiny, single-neuron-and-smaller differences rarely have such dramatic effects (unless they cascade into larger differences).  Consciousness is a property of large assemblies of neurons, robust to tiny changes.

But option (2) also seems implausible, for it would seem to involve massive introspective error.  Suppose that Entity I is an intermediate case with very much reduced, but not entirely absent, consciousness.  Chalmers suggests that instead of having bright red visual experience, Entity I has tepid pink experience.  (I'm inclined to think that this isn't the best way to think about fading or borderline consciousness, since it's natural to think of pink experiences as just different in experienced content from red cases, rather than less experiential than red cases.  But as I've argued elsewhere, genuinely borderline consciousness is difficult or impossible to imaginatively conceive, so I won't press Chalmers on this point.)

By stipulation, since Entity I is a functional isomorph, it will give the same reports about its experience as Chalmers himself would.  In other words, Entity I -- despite being barely or borderline conscious -- will say "Oh yes, I have vividly bright red experiences -- a whole welter of exciting phenomenology!"  Since this is false of Entity I, Entity I is just wrong about that.  But also, since it's a functional isomorph, there's no weird malfunction going on either, that would explain this strange report.  We ordinarily think that people are reliable introspectors of their experience; so we should think the same of Entity I.  Thus, option (2), gradual fading, generates an implausible tension: We have to believe that Entity I is radically introspectively mistaken; but that involves committing to an implausible degree of introspective error.

Therefore, neither option (1) nor option (2) is plausible.  But if Robot were not conscious, either (1) or (2) would have to be true for at least one Entity I.  Therefore, Robot is conscious.  And therefore, functional isomorphism is sufficient for consciousness.  It doesn't matter what materials an entity is made of.

We Can't Trust Robot "Introspection"

I acknowledge that it's an appealing argument.  However, Chalmers' response to option (2) should be unconvincing to the argument's target audience.

I have argued extensively that human introspection, even of currently ongoing conscious experience, is highly unreliable.  However, my reply today won't lean on that aspect of my work.  What I want to argue instead is that the assumed audience for this argument should not think that the introspection (or "introspection" -- I'll explain the scare quotes in a minute) of Entity I is reliable.

Recall that the target audience for the argument is someone who is antecedently neutral about Robot's consciousness.  But of course by stipulation, Robot will say (or "say") the same things about its experiences that Chalmers will say.  Just like Chalmers, and just like Entity I, it will say "Oh yes, I have vividly bright red experiences -- a whole welter of exciting phenomenology!"  The audience for Chalmers' argument must therefore initially doubt that such statements, or seeming statements, as issued by Robot, are reliable signals of consciousness.  If the audience already trusted these reports, there would be no need for the argument.

There are two possible ways to conceptualize Robot's reports, if they are not accurate introspections: (a.) They might be inaccurate introspections.  (b.) They might not be introspections at all.  Option (a) allows that Robot, despite lacking conscious experience, is capable of meaningful speech and is capable of introspecting, though any introspective reports of consciousness will be erroneous.  Option (b) is preferred if we think that genuinely meaningful language requires consciousness and/or that no cognitive process that fails to target a genuinely conscious experience in fact deserves to be called introspection.  On option (b) Robot only "introspects" in scare quotes.  It doesn't actually introspect.

Option (a) thus assumes introspective fallibilism, while option (b) is compatible with introspective infallibilism.

The audience who is to be convinced by the slow-fade version of the Fading Qualia Argument must both trust the introspective reports (or "introspective reports") of the intermediate entities while not trusting those of Robot.  Given that some of the intermediate entities are extremely similar to Robot -- e.g., Entity N-1, who is only one neuron different -- it would be awkward and implausible to assume reliability for all the intermediate entities while not doing so for Robot.

Now plausibly, if there is a slow fadeout, it's not going to be still going on with an entity as close to Robot as Entity N-1, so the relevant cases will be somewhere nearer the middle.  Stipulate, then, two values I and J not very far separated (0 < I < J < N) such that we can reasonably assume that if Robot in nonconscious, so is Entity J, while we cannot reasonably assume that if Robot is nonconscious, so is Entity I.  For consistency with their doubts about the introspective reports (or "introspective reports") of Robot, the target audience should have similar doubts about Entity J.  But now it's unclear why they should be confident in the reports of Entity I, which by stipulation is not far separated from Entity J.  Maybe it's a faded case, despite its report of vivid experience.

Here's one way to think about it.  Setting aside introspective skepticism about normal humans, we should trust the reports of Chalmers / Entity 0.  But ex hypothesi, the target audience for the argument should not trust the "introspective reports" of Robot / Entity N.  It's then an open question whether we should trust the reports of the relevant intermediate, possibly experientially faded, entities.  We could either generalize our trust of Chalmers down the line or generalize our mistrust of Robot up the line.  Given the symmetry of the situation, it's not clear which the better approach is, or how far down or up the slippery slope we should generalize the trust or mistrust.

For Chalmers' argument to work, we must be warranted in trusting the reports of Entity I at whatever point the fade-out is happening.  To settle this question, Chalmers needs to do more than appeal to the general reliability of introspection in normal human cases and the lack of functional differences between him, Robot, and the intermediate entities.  Even an a priori argument that introspection is infallible will not serve his purposes, because then the open question becomes whether Robot and the relevant intermediate entities are actually introspecting.

Furthermore, if there is introspective error by Entity I, there's a tidy explanation of why that introspective error would be unsurprising.  For simplicity, assume that introspection occurs in the Introspection Module located in the pineal gland, and that it works by sending queries to other parts of the brain, asking questions like "Hey, occipital lobe, is red experience going on there right now?", reaching introspective judgments based on the signals that it gets in reply.  If Entity I has a functioning, biological Introspection Module but a replaced, silicon occipital lobe, and if there really is no red experience going on in the occipital lobe, we can see why Entity I would be mistaken: Its Introspection Module is getting exactly the same signal from the occipital lobe as it would receive if red experience were in fact present.

It's highly doubtful that introspection is as neat a process as I've just described.  But the point remains.  If Entity I is introspectively unreliable, a perfectly good explanation beckons: Whatever cognitive processes subserve the introspective reporting are going to generate the same signals -- including misleading signals, if experience is absent -- as they would in the case where experience is present and accurately reported.  Thus, unreliability would simply be what we should expect.

Now it's surely in some respects more elegant if we can treat Chalmers, Robot, and all the intermediate entities analogously, as conscious and accurately reporting their experience.  The Fading Qualia setup nicely displays the complexity or inelegance of thinking otherwise.  But the intended audience of the Fading Qualia argument is someone who wonders whether experience tracks so neatly onto function, someone who suspects that nature might in fact be complex or inelegant in exactly this respect, such that it's (nomologically/naturally/scientifically) possible to have a behavioral/functional isomorph who "reports" experiences but who in fact entirely lacks them.  The target audience who is initially neutral about the consciousness of Robot should thus remain unmoved by the Fading Qualia argument.

This isn't to say I disagree with Chalmers' conclusion.  I've advanced a very different argument for a similar conclusion: The Copernican Argument for Alien Consciousness, which turns on the idea that it's unlikely that, among all behaviorally sophisticated alien species of radically different structure that probably exist in the universe, humans would be so lucky as to be among the special few with just the right underlying stuff to be conscious.  Central to the Fading Qualia argument in particular is Chalmers' appeal to the presumably reliable introspection of the intermediate entities.  My concern is that we cannot justifiably make that presumption.

Dancing Qualia

Chalmers pairs the Fading Qualia argument with a related but more complex Dancing Qualia argument, which he characterizes as the stronger of the two arguments.  Without entering into detail, Chalmers posits for sake of reductio ad absurdum that the alternative medium (e.g., silicon) hosts experiences but of a different qualitative character (e.g., color inverted).  We install a system in the alternative medium as a backup circuit with effectors and transducers to the rest of the brain.  For example, in addition to having a biological occipital lobe, you also have a functionally identical silicon backup occipital lobe.  Initially the silicon occipital lobe backup circuit is powered off.  But you can power it on -- and power off your biological occipital lobe -- by flipping a switch.  Since the silicon lobe is functionally identical to the biological lobe, the rest of the brain should register no difference.

Now, if you switch between normal neural processing and the backup silicon processor, you should have very different experience (per the assumption of the reductio) but you should not be able to introspectively report that different experience (since the backup circuit interacts identically with the rest of the brain).  That would again be a strange failure of introspection.  So (per the rules of reductio) we conclude that the initial premise was mistaken: Normal neural processing should generate the same types of experience as functionally identical processing in a silicon processor.

(I might quibble that you-with-backup-circuit is not functionally isomorphic to you-without-backup-circuit -- after all, you now have a switch and two different parallel processor streams -- and if consciousness supervenes on the whole system rather than just local parts, that's possibly a relevant change that will cause the experience to be different from the experience of either an unmodified brain or an isomorphic silicon brain.  But set this issue aside.)

The Dancing Qualia argument is vulnerable on the introspective accuracy assumption, much as the Fading Qualia argument is.  Again for simplicity, suppose a biological Introspection Module.  Suppose that what is backed up is the portion of the brain that is locally responsible for red experience.  Ex hypothesi, the silicon backup gives rise to non-red experience but delivers to the Introspection Module exactly the same inputs as that module would normally receive from an organic brain part experiencing red.  This is exactly the type of case where we should expect introspection to be unreliable.

Consider an analogous case of vision.  Looking at a green tree 50 feet away in good light, my vision is reliable.  Now substitute a red tree in the same location and a mechanism between me and the tree such that all the red light is converted into green light, so that I get exactly the same visual input I would normally receive from looking at a green tree.  Even if vision is highly reliable in normal circumstances, it is no surprise in this particular circumstance if I mistakenly judge the red tree to be green!

As I acknowledged before, this is a cartoon model of introspection.  Here's another way introspection might work: What matters is what is represented in the Introspection Module itself.  So if the introspection module says "red", necessarily I experience red.  In that case, in order to get Dancing Qualia, we need to create an alternate backup circuit for the Introspection Module itself.  When we flip the switch, we switch from Biological Introspection Module to Silicon Introspection Module.  Ex hypothesi, the experiences really are different but the Introspection Module represents them functionally in the same way, and the inputs and outputs to and from the rest of the brain don't differ.  So of course there won't be any experiential difference that I would conceptualize and report.  There would be some difference in qualia, but I wouldn't have the conceptual tools or memorial mechanisms to notice or remember the difference.

This is not obviously absurd.  In ordinary life we arguably experience minor version of this all the time: I experience some specific shade of maroon.  After a blink, I experience some slightly different shade of maroon.  I might entirely fail to conceptualize or notice the difference: My color concepts and color memory are not so fine grained.  The hypothesized red/green difference in Dancing Qualia is a much larger difference -- so it's not a problem of fineness of grain -- but fundamentally the explanation of my failure is similar: I have no concept or memory suited to track the difference.

On more holist/complicated views of introspection, the story will be more complicated, but I think the burden of proof would be on Chalmers to show that some blend of the two strategies isn't sufficient to generate suspicions of introspective unreliability in the Dancing Qualia cases.

Related Arguments

This response to the Fading Qualia argument draws on David Billy Udell's and my similar critique of Susan Schneider's Chip Test for AI consciousness (see also my chapter "How to Accidentally Become a Zombie Robot" in A Theory of Jerks and Other Philosophical Misadventures).

Although this critique of the Fading Qualia argument has been bouncing around in my head since I first read The Conscious Mind in the late 1990s, it felt a little complex for a blog post but not quite enough for a publishable paper.  But reading Ned Block's similar critique in his 2023 book has inspired me to express my version of the critique.  I agree with Block's observations that "the pathology that [Entity I] has [is] one of the conditions that makes introspection unreliable" (p. 455) and that "cases with which we are familiar provide no precedent for such massive unreliability" (p. 457).