Category Archives: philosophy

the difference between human and artificial intelligence: relationships

A large-language model (LLM) like ChatGPT works by identifying trends and patterns in huge bodies of text previously generated by human beings.

For instance, we are currently staying in Cornwall. If I ask ChatGPT what I should see around here, it suggests St Ives, Land’s End, St Michael’s Mount, and seven other highlights. It derives these ideas from frequent mentions in relevant texts. The phrases “Cornwall,” “recommended” (or synonyms thereof), “St Ives,” “charming,” “art scene,” and “cobbled streets” probably occur frequently in close proximity, because ChatGPT uses them to construct a sentence for my edification.

We human beings behave in a somewhat similar way. We also listen to or read a lot of human-generated text, look for trends and patterns in it, and repeat what we glean. But if that is what it means to think, then LLM has clear advantages over us. A computer can scan much more language than we can and uses statistics rigorously. Our generalizations suffer from notorious biases. We are more likely to recall ideas we have seen most recently, those that are most upsetting, those that confirm our prior assumptions, etc. Therefore, we have been using artificial means to improve our statistical inferences ever since we started recording possessions and tallying them by category thousands of years ago.

But we also think in other ways. Specifically, as intensely social and judgmental primates, we frequently scan our environments for fellow human beings whom we can trust in specific domains. A lot of what we believe comes from what a relatively small number of trusted sources have simply told us.

In fact, to choose what to see in Cornwall, I looked at the recommendations in The Lonely Planet and Rough Guide. I have come to trust those sources over the years–not for every kind of guidance (they are not deeply scholarly), but for suggestions about what to see and how to get to those places. Indeed, both publications offer lists of Cornwall’s highlights that resemble ChatGPT’s.

How did these publishers obtain their knowledge? First, they hired individuals whom they trusted to write about specific places. These authors had relevant bodily experience. They knew what it feels like to walk along a cliff in Cornwall. That kind of knowledge is impossible for a computer. But these authors didn’t randomly walk around the county, recording their level of enjoyment and reporting the places with the highest scores. Even if they had done that, the sites they would have enjoyed most would have been the ones that they had previously learned to understand and value. They were qualified as authors because they had learned from other people: artists, writers, and local informants on the ground. Thus, by reading their lists of recommendations, I gain the benefit of a chain of interpersonal relationships: trusted individuals who have shared specific advice with other individuals, ending with the guidebook authors whom I have chosen to consult.

In our first two decades of life, we manage to learn enough that we can go from not being able to speak at all to writing books about Cornwall or helping to build LLMs. Notably, we do not accomplish all this learning by storing billions of words in our memories so that we can analyze the corpus for patterns. Rather, we have specific teachers, living or dead.

This method for learning and thinking has drawbacks. For instance, consider the world’s five biggest religions. You probably think that either four or five of them are wrong about some of their core beliefs, which means that you see many billions of human beings as misguided about some ideas that they would call very important. Explaining why they are wrong, from an outsider’s perspective, you might cite their mistaken faith in a few deeply trusted sources. In your opinion, they would be better off not trusting their scriptures, clergy, or people like parents who told them what to believe.

(Or perhaps you think that everyone sees the same truth in their own way. That’s a benign attitude and perhaps the best one to hold, but it’s incompatible with what billions of people think about the status of their own beliefs.)

Our tendency to believe select people may be an excellent characteristic, since the meaning of life is more about caring for specific other humans than obtaining accurate information. But we do benefit from knowing truths, and our reliance on fallible human sources is a source of error. However, LLMs can’t fully avoid that problem because they use text generated by people who have interests and commitments.

If I ask ChatGPT “Who is Jesus Christ?” I get a response that draws exclusively from normative Christianity but hedges it with this opening language: “Jesus Christ is a central figure in Christianity. He is believed to be … According to Christian belief. …” I suspect that ChatGPT’s answers about religious topics have been hard-coded to include this kind of disclaimer and to exclude skeptical views. Otherwise, a statistical analysis of text about Jesus might present the Christian view as true or else incorporate frequent critiques of Christianity, either of which would offend some readers.

In contrast, my query about Cornwall yields confident and unchallenged assessments, starting with this: “Cornwall is a beautiful region located in southwestern England, known for its stunning coastline, picturesque villages, and rich cultural heritage.” This result could be prefaced with a disclaimer, e.g., “According to many English people and Anglophiles who choose to write about the region, Cornwall is …:” A ChatGPT result is always a summary of what a biased sample of people have thought, because choosing to write about something makes you unusual.

For human beings who want to learn the truth, having new tools that are especially good at scanning large bodies of text for statistical patterns should prove useful. (Those who benefit will probably include people who have selfish or even downright malicious goals.) But we have already learned a fantastic amount without LLMs. The secret of our success is that our brains have always been networked, even when we have lived in small groups of hunter-gatherers. We intentionally pass ideas to other people and are often pretty good at deciding whom to believe about what.

Moreover, we have invented incredibly complex and powerful techniques for improving how many brains are connected. Posing a question to someone you know is helpful, but attending a school, reading an assigned book, finding other books in the library, reading books translated from other languages, reading books that summarize previous books, reading those summaries on your phone–these and many other techniques dramatically extend our reach. Prices send signals about supply and demand; peer-review favors more reliable findings; judicial decisions allow precedents to accumulate; scientific instruments extend our senses. These are not natural phenomena; we have invented them.

Seen in that context, LLMs are the latest in a long line of inventions that help human beings share what they know with each other, both for better and for worse.

See also: the design choice to make ChatGPT sound like a human; artificial intelligence and problems of collective action; how intuitions relate to reasons: a social approach; the progress of science.

analytical moral philosophy as a way of life

(These thoughts are prompted by Stephen Mulhall’s review of David Edmonds’ book, Parfit: A Philosopher and His Mission to Save Morality, but I have not read that biography or ever made a serious study of Derek Parfit.)

The word “philosophy” is ancient and contested and has labeled many activities and ways of life. Socrates practiced philosophy when he went around asking critical questions about the basis of people’s beliefs. Marcus Aurelius practiced philosophy when he meditated daily on well-worn Stoic doctrines of which he had made a personal collection. The Analects of Confucius may be “a record of how a group of men gathered around a teacher with the power to elevate [and] created a culture in which goals of self-transformation were treated as collaborative projects. These people not only discussed the nature of self-cultivation but enacted it as a relational process in which they supported one another, reinforced their common goals, and served as checks on each other in case they went off the path, the dao” (David Wong).

To practice philosophy, you don’t need a degree (Parfit didn’t complete his), and you needn’t be hired and paid to be a philosopher. However, it’s a waste of the word to use it for activities that aren’t hard and serious.

Today, most actual moral philosophers are basically humanities educators. We teach undergraduates how to read, write, and discuss texts at a relatively high level. Most of us also become involved in administration, seeking and allocating resources for our programs, advocating for our discipline and institutions, and serving as mentors.

Those are not, however, the activities implied by the ideal of analytic moral philosophy. In that context, being a “philosopher” means making arguments in print or oral presentations. A philosophical argument is credited to a specific individual (or, rarely, a small team of co-authors). It must be original: no points for restating what has already been said. It should be general. Philosophy does not encompass exercises of practical reasoning (deciding what to do about a thorny problem). Instead, it requires justifying claims about abstract nouns, like “justice,” “happiness,” or “freedom.” And an argument should take into consideration all the relevant previous points published by philosophers in peer-reviewed venues. The resulting text or lecture is primarily meant for philosophers and students of philosophy, although it may reach other audiences as well.

Derek Parfit held a perfect job for this purpose. As a fellow of All Souls College, he had hardly any responsibilities other than to write philosophical arguments and was entitled to his position until his mandatory retirement. He did not have to obtain support or resources for his work. He did not have to deliberate with other people and then decide what to say collectively. Nor did he have to listen to undergraduates and laypeople express their opinions about philosophical issues. (Maybe he did listen to them–I would have to read the biography to find out–but I know that he was not obliged to do so. He could choose to interact only with highly prestigious peers.)

Very few other people hold similar roles: the permanent faculty of the Institute for Advanced Study, the professors of the Collège de France, and a few others. Such opportunities could be expanded. In fact, in a robust social welfare state, anyone can opt not to hold a job and can instead read and write philosophy, although whether others will publish or read their work is a different story. But whether this form of life is worthy of admiration and social support is a good question–and one that Parfit was not obliged to address. He certainly did not have to defend his role in a way that was effective, persuading a real audience. His fellowship was endowed.

Mulhall argues that Parfit’s way of living a philosophical life biased him toward certain views of moral problems. Parfit’s thought experiments “strongly suggest that morality is solely or essentially a matter of evaluating the outcomes of individual actions–as opposed to, say, critiquing the social structures that deeply shape the options between which individuals find themselves having to choose. … In other words, although Parfit’s favoured method for pursuing and refining ethical thinking presents itself as open to all whatever their ethical stance, it actually incorporates a subtle but pervasive bias against approaches to ethics that don’t focus exclusively or primarily on the outcomes of individual actions.”

Another way to put this point is that power, persuasion, compromise, and strategy are absent in Parfit’s thought, which is instead a record of what one free individual believed about what other free individuals should do.

I am quite pluralistic and inclined to be glad that Parfit lived the life he did, even as most other people–including most other moral philosophers–live and think in other ways. Even if Parfit was biased (due to his circumstances, his chosen methods and influences, and his personal proclivities) in favor of certain kinds of questions, we can learn from his work.

But I would mention other ways of deeply thinking about moral matters that are also worthy and that may yield different kinds of insights.

You can think on your own about concrete problems rather than highly abstract ones. Typically the main difficulty is not defining the relevant categories, such as freedom or happiness, but rather determining what is going on, what various people want, and what will happen if they do various things.

You can introduce ethical and conceptual considerations to elaborate empirical discussions of important issues.

You can deliberate with other people about real decisions, trying to persuade your peers, hearing what they say, and deciding whether to remain loyal to the group or to exit from it if you disagree with its main direction.

You can help to build communities and institutions of various kinds that enable their members to think and decide together over time.

You can identify a general and relatively vague goal and then develop arguments that might persuade people to move in that direction.

You can strive to practice the wisdom contained in clichés: ideas that are unoriginal yet often repeated because they are valid. You can try to build better habits alone or in a group of people who hold each other accountable.

You can tentatively derive generalizations from each of these activities, whether or not you choose to publish them.

Again, as a pluralist, I do not want to suppress or marginalize the style that Parfit exemplified. I would prefer to learn from his work. But my judgment is that we have much more to learn from the other approaches if our goal is to improve the world. That is because the hard question is usually not “How should things be?” but rather “What should we do?”

See also: Cuttings: A book about happiness; the sociology of the analytic/continental divide in philosophy; does doubting the existence of the self tame the will?

defining state, nation, regime, government

As a political philosopher by training, and now political scientist by appointment, I have long been privately embarrassed that I am not sure how to define “state,” “government,” “regime,” and “nation.” On reflection, these words are used differently in various academic contexts. To make things more complicated, the discussion is international, and we are often dealing with translations of words that don’t quite match up across languages.

For instance, probably the most famous definition of “the state” is from Max Weber’s Politics as Vocation (1919). He writes:

Staat ist diejenige menschliche Gemeinschaft, welche innerhalb eines bestimmten Gebietes – dies: das „Gebiet“, gehört zum Merkmal – das Monopol legitimer physischer Gewaltsamkeit für sich (mit Erfolg) beansprucht.

[The] state is the sole human community that, within a certain territory–thus: territory is intrinsic to the concept–claims a monopoly of legitimate physical violence for itself (successfully).

Everyone translates the keyword here as “state,” not “government.” But this is a good example of how words do not precisely match across languages. The English word “government” typically means the apparatus that governs a society. The German word commonly translated as “government” (“der Regierung“) means an administration, such as “Die Regierung von Joe Biden” or a Tory Government in the UK. (In fact, later in the same essay, Weber uses the word Regierung that way while discussing the “typical figure of the ‘grand vizier'” in the Middle East.) Since “government” has a wider range of meanings in English, it wouldn’t be wrong to use it to translate Weber’s Staat.

Another complication is Weber’s use of the word Gemeinschaft inside his definition of “the State.” This is a word with such specific associations that it is occasionally used in English in place of our vaguer word “community.” A population is not a Gemeinschaft, but a tight association can be. Thus to translate Weber’s phrase as “A state is a community …” is misleading.

For Americans, a “state” naturally means one of our fifty subnational units, but in Germany those are Länder (cognate with “lands”). The word “state” derives from the Latin status, which is as “vague a word as ratio, res, causa” (Paul Shorey, 1910) but can sometimes mean a constitution or system of government. Cognates of that Latin word end up as L’État, el Estado and similar terms that have a range of meanings, including the subnational units of Mexico and Brazil. In 1927, Mussolini said, “Tutto nello Stato, niente al di fuori dello Stato, nulla contro lo Stato (“Everything in the State, nothing outside the State, nothing against the State”). I think he basically meant that he was in charge of everything he could get his hands on. Louis XIV is supposed to have said “L’État c’est moi,” implying that he was the government (or the nation?), but that phrase may be apocryphal; an early documented use of L’État to mean the national government dates to 1799. In both cases, the word’s ambiguity is probably one reason it was chosen.

“Regime” can have a negative connotation in English, but political theorists typically use it to mean any government plus such closely related entities as the press and parties and prevailing political norms and traditions. Regimes can be legitimate, even excellent.

If these words are used inconsistently in different contexts, then we can define them for ourselves, as long as we are clear about our usage. I would tend to use the words as follows:

  • A government: either the legislative, executive, and judicial authority of any entity that wields significant autonomous political power (whether it’s territorial or not), or else a specific group that controls that authority for a time. By this definition, a municipality, the European Union, and maybe even the World Bank may be a government.

(A definitional challenge is deciding what counts as “political” power. A company, a church, a college, an insurgent army, or a criminal syndicate can wield power and can use some combination of legislative, executive, and/or judicial processes to make its own decisions. Think of canon law in the Catholic Church or an HR appeals processes inside a corporation. Weber would say that the fundamental question is whether an entity’s power depends on its own monopolistic use of legitimate violence. For instance, kidnapping is a violent way to extract money, but it does not pretend to be legitimate and it does not monopolize violence. Taxation is a political power because not paying your taxes can ultimately land you, against your will, in a prison that presents itself as an instrument of justice. Not paying a private bill can also land you in jail, but that’s because the government chooses to enforce contracts. Your creditor is not a political entity; the police force is. However, when relationships between a government and private entities are close, or when legitimacy is controversial, or when–as is typical–governments overlap, these distinctions can be hard to maintain and defend.)

  • A state: a government plus the entities that it directly controls, such as the military, police, or public schools. For example, it seems most natural to say that a US government controls the public schools, but not that a given school is part of the government. Instead, it is part of the state. Likewise, an army can be in tension with the government, yet both are components of the state.
  • A regime: the state plus all non-state entities that are closely related to it, e.g., political parties, the political media, and sometimes the national clergy, powerful industries, etc. We can also talk about abstract characteristics, such as political culture and values, as components of a regime. A single state may change its regime, abruptly or gradually.
  • A country: a territory (not necessarily contiguous) that has one sovereign state. It may have smaller components that also count as governments but not as countries.
  • A nation: a category of people who are claimed (by the person who is using this word) to deserve a single state that reflects their common identity and interests. Individuals can be assigned to different nations by different speakers.
  • A nation-state: a country with a single functioning and autonomous state whose citizens widely see themselves as constituting a single nation. Some countries are not nations, and vice-versa. People may disagree about whether a given country is a nation-state, depending on which people they perceive to form a nation.

See also: defining capitalism; avoiding a sharp distinction between the state and the private sphere; the regime that may be crumbling; what republic means, revisited etc.

Analyzing Political Opinions and Discussions as Networks of Ideas

This is a talk that I have prepared for the Universidad Carlo III in Madrid tomorrow. It is a summary of recent work that I have been conducting with colleagues at Northeastern, Wisconsin, and Oxford and that I’m beginning to develop into a book manuscript.

In the model that I present, an individual holds potentially connected beliefs about political or moral issues, which we can represent with nodes and links (an “idiodictuon”). Whether and how the various ideas are linked in the person’s network influences that individual’s actions and opinions. When people discuss political or ethical issues, they share portions of their respective networks of which they are conscious at the time and may bring ideas from their interlocutors into their own idiodictuons.

Some network structures are better than others for discussion: overly centralized or scattered networks are problematic. Individuals tend to demonstrate similar network structures on different issues, so that having a proclivity for a certain form of network is a character trait.

People, with their respective networks of ideas, are also embedded in social networks. An idea is more likely to spread depending on features of both the social network and the idea networks of the people who interact. Specifically, the odds that an idea will spread from a given person depend on how many people receive communications from that person and how much they trust the communicator. It is reasonable to take into account the trustworthiness of a source when assessing an idea.

As a whole, a population may develop a shared network structure. An idea that is widely shared and frequently central in individuals’ networks becomes a norm. Such norms play important roles in institutions. A community or a culture is a single network or phylodictuon that encompasses disagreement. Ultimately, all such networks interconnect to form a network of human ideas.

does doubting the existence of the self tame the will?

I like the following argument, versions of which can be found in many traditions from different parts of the world:

  1. A cause of many kinds of suffering is the will (when it is somehow excessive or misplaced).
  2. Gaining something that you desire does not reduce your suffering; you simply will something else.
  3. However, one’s will can be tamed.
  4. Generally, the best way to manage the will is to focus one’s mind on other people instead of oneself. Thus,
  5. Being ethical reduces one’s suffering.

In some traditions, notably in major strands of Buddhism and in Pyrrhonism, two additional points are made:

  1. The self does not actually exist. Therefore,
  2. It is irrational to will things for oneself.

Point #7 is supposed to provide both a logical and a psychological basis for #4. By realizing that I do not really exist, I reduce my attachment to my (illusory) self and make more space to care about others, which, in turn, makes me happier.

Point #6 is perfectly respectable. Plenty of philosophers (and others) who have considered the problem of personal identity have concluded that an ambitious form of the self does not really exist. (For instance, David Hume.)

But if the self doesn’t exist, does it really follow that we should pay more attention to other people? We might just as well reason as follows:

  1. The self does not really exist. Therefore,
  2. a. Other people do not really exist as selves. Therefore,
  3. a. It is irrational to be concerned about them.

Or

  1. The self does not really exist. Therefore,
  2. b. It is impossible for me to change my character in any lasting way. Therefore,
  3. b. There is no point in trying to make myself more ethical.

Striving to be a better or happier person is not a sound reason for doubting the existence of the self. This doubt may do more harm than good. If there actually is no self, that is a good reason not to believe in one. But then we are obliged to incorporate skepticism about personal identity into a healthy overall view. The best way might be some version of this:

  1. The self does not really exist. Nevertheless,
  2. c. I would be wise to treat other people as if they were infinitely precious, durable, unique, and persistent things (selves).

I think it is worth getting metaphysics right, to the best of our ability. For example, it is worth trying to reason about what kind of a thing (if anything) a self is. However, I don’t believe that metaphysical beliefs entail ways of life in a straightforward way, with monotonic logic.

Any given metaphysical view is usually compatible with many different ways of being. It may even strongly encourage several different forms of life, depending on how a person absorbs the view. Thus I am not surprised that some people (notably, thoughtful Buddhists) have gained compassion and equanimity by adopting the doctrine of no-self, even though the same doctrine could encourage selfishness in others, and some people may become more compassionate by believing in the existence of durable selves. In fact, many have believed in the following argument:

  1. Each person (or sentient being) has a unique, durable, essential being
  2. I am but one out of billions of these beings. Therefore,
  1. It is irrational to will things for myself.

The relationship between an abstract idea and a way of being is mediated by “culture,” meaning all our other relevant beliefs, previous examples, stories, and role-models. We cannot assess the moral implications of an idea without understanding the culture in which it is used. For instance, the doctrine of no-self will have different consequences in a Tibetan monastery versus a Silicon Valley office park.

We cannot simply adopt or join a new culture. That would require shedding all our other experiences and beliefs, which is impossible. Therefore, we are often in the position of having to evaluate a specific idea as if it were a universal or culturally neutral proposition that we could adopt all by itself. For instance, that is what we do when we read Hume and Kant (or Nagarjuna) on the question of personal identity and try to decide what to think about it. This seems a respectable activity; I only doubt that, on its own, it will make us either better or worse people.

See also: notes on religion and cultural appropriation: the case of US Buddhism; Buddhism as philosophy; how to think about the self (Buddhist and Kantian perspectives); individuals in cultures: the concept of an idiodictuon. And see “The Philosophic Buddha” by Kieran Setiya, which prompted these thoughts.