2 Empirical resolutions to metaethical debates

2.1 Sentimentalism vs. rationalism and externalism

Let’s begin with the first question on the metaethics decision tree: are moral judgments affect-laden? No question in ethics has received more empirical attention than this. Dozens of studies have attempted to determine whether emotions play a central role in morality, and the evidence has consistently shown that they do. Let me begin with an unpublished study of my own and then offer a brief review of the empirical literature.

To begin with, let’s consider folk intuitions. Do ordinary people use emotions as evidence when attributing moral judgments? To test this, I conducted a simple vignette study, which pitted emotions against verbal testimony. A group of college undergraduates taking an introductory-level philosophy class responded to the following probe:

Fred belongs to a fraternity and his brothers in the fraternity sometimes smoke marijuana. Fred insists that he thinks it’s morally acceptable to smoke marijuana. He says, “You guys are not doing anything wrong when you smoke.” But Fred also feels disgusted with his frat brothers when he sees them smoking. One day, to prove that he thinks smoking is okay, he smokes marijuana himself. Afterwards, he feels incredibly ashamed about smoking the drug.

Which of the following seems more likely:

  1. Fred says he morally approves of marijuana smoking, but in reality he thinks it is morally wrong.

  2. Fred feels badly about smoking marijuana, but in reality he thinks it is morally acceptable.

In my small sample, 68.4% chose answer 1, suggesting that the majority of them take emotions as evidence for moral values, even when that directly contradicts self-report. This suggests that many people take emotions to be a sufficient evidence for attribution moral attitudes. An even more dramatic result was obtained when another twenty participants assessed this scenario:

Frank belongs to a fraternity and his brothers in the fraternity sometimes smoke marijuana. Frank insists that their actions are morally unacceptable. He says, “You guys are doing something wrong when you smoke.” But Frank does not feel any anger or disgust when he sees his frat brothers smoking. One day, when they are not around, he smokes marijuana himself. Afterwards, he doesn’t feel any shame about smoking the drug.

Which of the following seems more likely:

  1. Frank says he morally opposes marijuana smoking, but in reality he thinks it is morally acceptable.

  2. Frank doesn’t feel badly about smoking marijuana, but in reality he thinks it is morally wrong.

Here, 89.5% of participants chose response 1, indicating that they take emotions to be necessary for the attribution of moral attitudes. Absent the right feelings, verbal testimony is treated as an unreliable indicator of a person’s values.

This study has at least four serious limitations: people may not trust self-reports; the results were far from unanimous; it fails to distinguish evidence for moral attitudes and essence of moral attitudes; and folk beliefs about moral judgments may be wrong. To get around these limitations we must move beyond experimental philosophy, and look for more direct evidence that emotions actually are sufficient and necessary for moral judgments. But the study is still revealing, because it shows that emotions are used as evidence in moral attribution. Most participants make attributions that fall in line with sentimentalism.

To show that emotions actually do contribute to moral cognition, we can look at three kinds of evidence: cognitive neuroscience, behavioral psychology, and pathology. In each domain, sentimentalism finds support. There have now been dozens of neuroimaging studies on moral judgment tasks, and every one of them, to my knowledge, shows an increase in activation in brain structures associated with emotion, when moral decisions are compared to non-moral decisions. Key stuctures include the posterior cingulate, temporal pole, orbitofrontal cortex, and ventromedial prefrontal cortex. There are only two groups of studies that even appear to depart from this pattern. Joshua Greene et al. (2001) report that emotions play more of a role in deontological judgments than in consequentialist judgments, but their data show that, as compared to non-moral judgments, emotions are involved in both (see their Figure 1). Moreover, Greene et al. use moral dilemmas in which the common denominator is saving lives—they manipulate the nature of the harm necessary in order to save five people in danger. Thus, each moral judgment condition presumably elicits the judgment that it would be good to help people in need. This positive moral judgment may be emotionally grounded, but the neuroimaging method subtracts away this emotional information, because it is present in each scenario, and imaging results of this kind report only contrasts between different conditions. Thus, a major dimension of moral emotions may be systematically concealed by the method. The other study that fails to show an increase in emotional responses during moral judgment is one condition in a series of imagining experiments performed by Jana Borg et al. (2006). But, in that condition, a moral scenario is compared to a scenario about an encroaching fire that threatens one’s property, and it is unsurprising that moral judgments produce less of an emotional response than a case of personal loss.

Brain science resoundingly links moral judgment to emotion, but the method is correlational. Moral rationalists and externalists could concede that moral judgments excite emotional responses, while denying that these are the basis of moral judgment. Imagine the following view: we use reason to arrive at moral judgments, but morality matters to us, so when we arrive at those judgments emotions normally kick in. By analogy, reason might be used to determine that certain life activities (smoking, high fat diets, sleep deprivation) are harmful, and, upon drawing that reason-based conclusion, we tend to experience corresponding emotions, such as anxiety when contemplating lighting a cigarette. Neuroimaging results showing responses to cigarettes might confirm this, showing emotion areas active when cigarettes are seen, but that wouldn’t refute a rationalist theory of how we arrive at the judgment that cigarettes are dangerous.

To adjudicate between the thesis that emotions are constitutive of moral judgments, as opposed to mere consequences, we need behavioral evidence. Numerous studies now establish a causal link between emotion and moral judgment. When emotions are induced, they influence how good or bad things seem. Induction methods have been widely varied: hypnosis, dirt, film clips, autobiographical recall, and smells. In one recent study, Kendal Eskine, Natalie Kacinik, and I induced bitterness by giving people a bitter beverage and found that moral judgments became more severe (Eskine et al. 2011). In other recent studies Angelika Seidel and I use sound clips to induce specific emotions, and we have shown that different emotions have different moral effects: anger induces more stringent wrongness judgments about crimes against persons; disgust induces greater stringency on crimes against nature (such as cannibalism); and happiness induces stronger judgments that it is both good and compulsory to help the needy (Seidel & Prinz 2013a, 2013b). There is also evidence that we feel different emotions when judging our own actions than when judging others. When another person commits a crime against nature, we tend to feel disgust, but when we perform an act deemed by others to be unnatural, the most common response seems to be shame. Conversely, when others commit crimes against persons, we feel angry, but guilt is the natural response when we perform such acts ourselves. To test this hypothesis, I conducted a forced-choice study in which a group of college undergraduates had to pick guilt or shame in response to mildly “unnatural” acts (“Suppose your roommate catches you masturbating”), and mildly harmful acts (“Suppose you take something from someone and never return it”). 80% chose shame for the first case, and over 90% picked guilt for the second.

Such findings demonstrate that different emotions play different roles. I mentioned three distinctions that are currently receiving empirical attention: the split between positive and negative emotions (praise and blame), between two kinds of blame (crimes against nature and crimes against persons), and between self- and other-directed blame. The self/other distinction may be particularly important because it helps us see how moral emotions differ from their non-moral analogues. Anger (or at least irritation) and disgust can both occur in non-moral contexts, but they take on a moral cast, I submit, when paired with dispositions to feel guilt and shame, respectively. If I find eating insects physically revolting, I will experience disgust when I see others eat insects, and disgust when I inadvertently eat them myself. But if I found insect eating immoral, it would not be disgust that I experience in the first-person case, but shame. This feeling of shame would motivate me to make amends for my actions, or to conceal my wrongdoing from others, not simply to repel the unwanted food from my body. The self-directed emotions round out the punitive cast of our moral attitudes. We see morally bad acts as not just worth aggressing against, but as worthy of apology. This need not be a second-order belief. Rather it is implicit in the fact that moralized behaviors carry emotional dispositions toward self and other that together promote a punitive attitude: a disposition to issue and submit to punishment.

Putting this together, I propose that standing moral values (the values that a given individual has for an extended period of time) consist in dispositions to feel the self- and other-directed emotions that I have been discussing. Such an emotional disposition can be called a sentiment. On any given occasion when a standing value becomes active in thought—i.e., when a moral judgment is made—these dispositions result, all else being equal, in an emotional state. The emotion that is felt depends on who is doing what to whom. For example, if I recall a situation in which I hurt someone’s feelings, I will have a feeling of guilt regarding that event, because a person was harmed and I was the culprit. This feeling of guilt toward an event constitutes my judgment that the action was wrong, and I gain introspective access to this judgment by feeling guilt well up inside me. If this is right, then emotions are not merely effects of moral judgments, but essential components of them.

Against this picture, one might object that emotions are merely a heuristic that can be used in certain circumstances, but not strictly necessary for making moral judgment. Following the analogy mentioned before, anxiety might be used as a heuristic when deciding whether to smoke, but the judgment that smoking is dangerous does not depend on fear, and was initially arrived at by the light of reason.

To establish that emotions are not merely helpful heuristics, one must see what happens when emotions are reduced or eliminated. To look into this, Eskine (2011) gave people the bitter taste manipulation and then warned them not to let the feelings caused by that beverage interfere with the moral judgments. In this condition, he found that moral judgments were considerably less severe than a control condition, suggesting that, when we ignore emotions, it is harder to see things as wrong. The finding indicates, in other words, that moral judgments subside when emotions are absent. The study cannot confirm this strong claim, however, because people cannot suppress emotions completely. More powerful evidence comes from the clinical populations who suffer from emotional deficits. For example, psychopaths, who suffer from deficit in guilt and other negative emotions, notoriously fail to appreciate what is wrong with their actions (Hare 1993). Similarly, people with Huntington’s disease, which impairs disgust, show high incidence of paraphelias, suggesting that they cease to see deviant sexual behavior as wrong (Schmidt & Bonelli 2008). Kramer (1993, p. 278) argues that anti-depressants can flatten affect in a way that results in a “loss of moral sensibility.” There is also a positive relationship between alexithymia and Machiavellianism, suggesting that a reduction in emotional competence may act in ways that are more instrumental than moral (Wastell & Booth 2003). For better or worse, there is no clinical condition in which all emotions are absent and behavioral function remains, but these findings suggest that selective or global dampening of the emotions leads to corresponding deficits in moral judgment. That is, people with diminished emotions seem to be insensitive to corresponding parts of the moral domain, suggesting that they may not be forming moral judgments.

The evidence summarized here suggests that emotions arise when we make moral judgments, that emotions are consulted when reporting such judgments, and that moral judgments are impaired when emotions are unavailable. Some of this evidence is preliminary, but, for present purposes, let’s assume that the findings hold up to further and more stringent testing. By inference to the best explanation, such findings suggest that emotions are components of moral judgments. The idea is that, when people say something is morally bad, the thought they are expressing on that occasion consists of a negative emotion directed towards the thing judged bad. Emotions, on this view, function like predicates in thought. That is what traditionally sentimentalists, such as Hume, seem to have maintained. Hume thought ideas—the components of thoughts—were stored copies of impressions, and the idea of moral badness consisted in a stored copy of the impression of disapprobation.

Traditional sentimentalism, which says that emotions (or sentiments) are actually components of moral judgments, differs conspicuously from neo-sentimentalism. Neo-sentimentalists theories say that moral judgments are judgments about the appropriateness of emotions. These theories do not straightforwardly predict that emotions come on line when we make moral judgments, nor that a reduction in emotions should interfere with our ability to moralize. Instead, they predict that people will think about emotions when they make moral judgments. Correlatively, they also predict that people with limited metacognitive abilities will lose their ability to make moral judgments; this is not the case (Nichols 2008). Thus, given the current state of evidence, traditional sentimentalism outperforms neo-sentimentalism empirically. Traditional sentimentalism predicts a robust pattern of empirical findings.

Rationalists and externalist moral realists might baulk at this point and say that the empirical evidence lacks the adequate modal strength to support sentimentalism. The evidence shows that emotions are often consulted when making moral judgments, but this leaves open the possibility that we might also make moral judgments dispassionately under circumstances that have not yet been empirically explored. So stated, this objection is just an expression of faith. It suffers from both conceptual and empirical weaknesses. Conceptually, opponents of sentimentalism must say what moral judgments are, such that they can be had dispassionately. What thought is a dispassionate person conveying, when she says, “Killing the innocent is morally bad?” Any attempt to give a reductive answer will be vulnerable to open-question worries. No descriptive substitute for the phrase “morally bad” leaves us with a sentence that is conceptually synonymous with the third.

Arguably, the open-question argument does not threaten sentimentalism. Let’s distinguish two kinds of open questions. First, given a certain attitude towards killing, one can still wonder whether killing really is morally bad. Second, given a certain attitude toward killing, one can wonder whether one is thereby regarding it as morally bad. Reductive theories of value leave both questions open. If I form the attitude that killing cannot be willed as a universal law, I can still wonder both whether killing is bad and whether I am judging that it is bad. Sentimentalism leaves the first question open, but not the second. When experiencing outrage at killing, it seems impossible to wonder I am regarding killing as bad. I can of course wonder whether killing really is as bad as it seems. Such doubts can arise because I may not know the true source of the emotion I am feeling. Perhaps my outrage comes from some extraneous source (such as a bitter beverage), for example. But this open question does not threaten the thesis that moral judgments are constituted by sentiments. The only open question that poses such a threat would be one about what my attitude is, not one about whether my attitude is true. The fact that some sentiments are experienced as condemnatory effectively closes the question about whether someone experiencing those sentiments is adopting a moral stance. By analogy, imagine tasting a wine and wondering whether it really is delicious, while experiencing gustatory pleasure. We can have this thought (a thought about truth), because we can’t be sure where the pleasure came from (was it the wine or the company?). But we can’t experience gustatory pleasure and wonder whether we are, at that moment, finding the experience delicious. Thus, gustatory pleasure is plausible a component of deliciousness judgments.

The foregoing may look like a conceptual argument for sentimentalism. But it can also be construed as an empirical claim. The argument hangs on the premise that people experiencing outrage take themselves to be making moral judgments. This can be empirically tested. Indeed, all the evidence about people consulting their emotions when making moral judgments stands as evidential support. Merely making someone mad results in more negative moral attitudes. This can be interpreted as showing that, when people are angry, there is no question for them about whether they are holding something in negative moral regard. Conversely, it would be easy to show that people do not necessarily draw this inference when they form the judgment that something cannot be willed as a universal law. Opponents of sentimentalism owe us a positive account of evaluative thoughts that avoids open-question worries as successfully as sentimentalist accounts.

Opponents of sentimentalism might try to bypass this demand by offering a non-reductive account of moral judgments, treating thin moral concepts as primitives. That possibility, which was attractive to Moore, looks unmotivated given the empirical evidence for an emotional foundation. Every study suggests that emotions arise when we make moral judgments. All evidence also suggests that when emotions are eliminated, judgments subside as well. This does not prove that we can make moral judgments without emotions, but, by induction, it provides evidence. Some have argued that extant evidence is ambiguous about whether emotions are essental components of moral judgments or mere accompaniments, but I have suggested here that the former may provide a better explanation (and certainly better predictions) of the total pattern of data (Huebner et al. 2009; Waldmann et al. 2012). Until opponents of sentimentalism can identify some clear cases of moral judgments without emotions, they will be on the losing side of the debate. At the moment, there is no empirical evidence that this ever happens.

Notice too, that it would be relatively uninteresting to show that, under as-yet-unidentified and highly unusual conditions, people can make what look like moral judgments in the absence of emotions. The sentimentalist will reply that the vast majority of ordinary moral judgments are emotionally based. If moral vocabulary is occasionally used dispassionately, sentimentalists can ask whether the thoughts expressed on such occasions are of the same kind that we find, in study after study, in the usual cases. Upon finding a class of dispassionate judgments, one might do best to posit an ambiguity in the category. The sentimentalist can content herself with the project of providing a metaethics for garden-variety moral judgments, while leaving open the possibility that there may be psychological exotica, which conform to the theories of their opponents. At the moment, there is no empirical evidence for such exotica.

More modestly, the empirically-minded sentimentalist might welcome an attempt to find evidence for opposing views. Little effort has been put into this task, though empirical claims for emotion-free moralizing are occasionally advanced. The most publicized example is Koenigs et al.’s (2007) study, which shows intact consequentialist judgments in patients who suffer from ventromedial prefrontal brain injuries, which are thought to impair emotion. But this description is misleading. As the authors note, ventromedial patients are highly emotional, and their most notorious symptom is that they are insensitive to costs when seeking rewards. Presumably, reward- seeking is an affectively grounded behavior. The fact that these patients make normal consequentialist judgments does not entail that they rely on reason alone, but rather on their positive emotions. Since these emotions cannot be easily regulated by negative feedback in ventromedial patients, they tend to be more consequentialist than healthy populations—that is, they are more willing to push a heavy man in a trolley’s path in order to save five.

Will better empirical evidence for rationalism or externalist moral realism be forthcoming? I doubt it. Rationalists hold that we can arrive at moral judgments through reasoning. Unlike some sentimentalists, I think reasoning is important to morality. It is likely that we use reasoning to extrapolate from basic values to novel cases. But it is unlikely that we could use reasoning to derive basic moral values. Philosophers have tried to do this for centuries with no consensus behind any view. This might be described as a strong empirical argument by induction: thousands of smart, trained moral experts have failed to identify a line of reasoning that is widely recognized as providing adequate rational support for basic moral propositions. Moreover, when moral debates arise, there is little evidence that reasoning is efficacious on its own. Instead, societal transformations in values seem to arrive with political upheavals, economic revolutions, and generational change. Attitudes towards slavery changed with the industrial revolution, women’s suffrage came with a world war, and increase in support for gay rights correlates with the dissolution of traditional social roles and economic transformations that have made procreation more costly than abstinence. I don’t mean to imply that there are no rational arguments for these liberation movements. Rather, I am suggesting that those arguments take hold only when social conditions are right. It is noteworthy, for example, that scientific racism appeared very late in the history of slavery, suggesting that slavery was not simply based on false beliefs about racial inequality. In fact many societies have enslaved their own people, and many proponents of scientific racism have been against slavery. Rather, advocacy of slavery seems to reflect a set of basic moral values that changed in recent history: values that say social standing can be determined by the lottery of birth. With industrialization, models of labor based on the idea of self-determination took hold, and the idea that birth should determine social standing began to wane. Of course, it hasn’t disappeared altogether, but it has been tempered by the the emergence of a new norm. Before industrialization, the idea that human beings are born equal and free might have seemed manifestly false, and thus it could have played no effective role in any argument against slavery. With industrialization, this premise gained appeal, and became the foundation of compelling arguments. Arguments are not inert, but they are only as good as the premises on which they are based, and the plausibility of those premises may depend on factors other than reasonging. It is possible that reasons have little role in driving basic values. And if so, then the recent broadening moral umbrella is not the result of a rational inference to the conclusion that our basic values cover more cases than we thought, but rather an irrational shift in basic values.

A realist might concede that such considerations threaten rationalism, but vie instead for a kind of intuitionist perspective, according to which basic moral truths are simply obvious. To me, this looks like a magical moral epistemology—one wonders what moral facts could be such that our moral sense could simply lock on to them. It is also open to a damaging empirical objection. Phenomenologically, it is true that moral intuitions often seem immediate and unbidden, but this can be readily explained on a sentimentalist account. Emotions are conditioned (by training or evolution) to arise automatically and often intensely when certain actions, such as torturing babies, are considered. This gives an impression of immediacy without postulating any special contact with moral reality. Moreover, these intuitions vary from group to group. For example, there is empirical evidence that liberals and conservatives have divergent basic values (Graham et al. 2009). The presence of such foundational intuitions can be explained demographically, and their lack of convergence casts doubt on the existence of a moral faculty that reveals universal moral truths. In other words, intuitionism is vulnerable to a debunking argument. Social science coupled with sentimentalism provides a good explanation of deeply-held intuitions, so there is no need to suppose that these intuitions reflect anything deeper.

This point about moral variation, to which I will return, also counts against some forms of externalist moral realism. Advocates of that position sometimes suggest that objective moral facts can be established by identifying the external factors that best explain human moral behavior or judgments. If moral behavior and judgments vary from group to group, however, it is unlikely that we will find an external common denominator underlying these practices. Such a search also seems unnecessary given that we already have good explanations of moral behavior and judgments in terms of socially-conditioned sentiments.

None of these arguments are the nail in the coffin for externalist realist or rationalist theories. They merely illustrate the relevance of empirical results. The findings mentioned here must be explained. It is my contention that sentimentalism provides the best explanation of the findings I have reviewed, but further arguments and evidence could tip the balance in another direction.

2.2 Cognitivism vs. non-cognitivism

Let’s move on to the second question on the metaethics decision tree: Are moral judgments truth-apt? As positioned on the tree, this is a question that arises for sentimentalists, raised pressingly by the conclusion that moral judgments have a basis in the emotions. It is that conclusion that seems to put truth-aptness in jeopardy, since emotions have not traditionally been regarded as having accuracy conditions in the way regarded as allowing for truth. But, it should be noted that the question of truth-aptness could also be raised independently of sentimentalism. There are non-sentimentalist theories that deny truth-aptness (for example, one might say that moral judgments are imperative, while denying that they need be passionate), and there are non-sentimentalist theories that accept truth-aptness (the vast majority fall in this category). To keep things as neutral as possible, I will begin by asking whether there is any empirical evidence that moral judgments are non-cognitive, whether or not they are affect-laden.

The posing of this question is itself a degree of philosophical progress, because non-cognitivists too rarely reflect on the predictions of their view. Indeed, the most obvious empirical prediction fails resoundingly. If moral judgments do not aim at truth, we might expect them to have a non-declarative syntactic. For example, we might expect them to take the form of imperatives or exclamations. But they do not. In every language that I know of, moral judgments are expressed using declarative sentences, which should stand as a profound embarrassment to the theory. Granted, non-cognitivists sometimes propose elaborate logics to accommodate this fact, but it is surprising that they should have to do so. One would expect the surface grammar to reflect the non-cognitive form.

To push things further, one might look for more subtle linguistic evidence in favor of non-cognitivism. For example, some non-cognitivists assume that moral utterances have the illocutionary force of directives, such as orders, requests, or demands. Directives often occur in speech contexts that contain words that play a role in persuasion, such as “come!”, “let’s”, or “we encourage you…” To empricially test this kind of non-cognitivism, Olasov (2011) ingeniously used this technique for sociolinguistics, called corpus analysis. He used a set of such linguistic elements that correlate with directive speech, such as those just mentioned, and he searched corpora of spoken and written texts for co-variance between these elements and moral terms. He calls the directive elements “suasion markers,” and the correlations between these and other linguistic items a “suasion score.” Non-cognitivism seems to predict a high suasion score, given the postulated directive function of moral judgments. This prediction fails. Not only is there no positive correlation between moral vocabulary and suasion markers, there is actually a negative correlation, which approaches significance. This negative relationship was observed in seventeen out of nineteen different categories of text that he examined. These results are preliminary—a first foray into empirical ethics—but they provide compelling evidence that moral discourse is not directive in nature.

Non-cognitivism entails that moral discourse does not aim to refer to facts in the world. This carries another linguistic prediction that can be readily tested. Certainly adverbs are used to indicate a focus on how things are in the world. These include “really,” “truly,” and “actually.” These words have other uses (“really” can be a term of emphasis), but they often play a role in emphasizing the factive nature of the modified phrase. Therefore, if non-cognitivism were true one might expect these words to rarely be used as modifiers for moral terms. To test this, I used Google search engine to search for and note the frequency of three phrases: “really immoral,” “truly immoral,” and “actually immoral.” To do this, I needed a baseline, and chose to compare “immoral” to a word widely believed to designate a objective feature of the world. I chose “triangular,” a classic primary quality, on a Lockean scheme. The results are as follows (as of March, 2013):

 

“really triangular”: 6,500 hits

“really immoral”: 10,600 hits

“truly triangular”: 4,920 hits

“truly immoral”: 32,000 hits

“actually triangular”: 21,600 hits

“actually immoral”: 61,600 hits

 

Clearly, the adverbs that indicate a real-world focus are used more frequently for moral terms than for terms designating objective physical features—over six times as common in the case of “truly.” I also tried the phrases “in truth,” “truthfully,” and “in actual fact”:

 

“truthfully triangular”: 6 hits

“truthfully immoral”: 44 hits

“in truth triangular”: 46 hits

“in truth immoral”: 1,350 hits

“in actual fact triangular”: 2 hits

“in actual fact immoral”: 133 hits

 

These truth-tracking phrases modify “immoral” between 7 and 166 times more frequently than they modify “triangular.” Moreover, these differentials are misleadingly small because the baserate for “immoral” is far lower than “triangular” (6,910,000 hits as compared to 11,600,000). This was just an exploratory study, but there is a simple implication. Non-cognitivism makes linguistic predictions, and when those are tested, they do not seem to pan out. Non-cognitivists owe us evidence, or they must deny that their theory makes predictions, in which case it would cease to be falsifiable.

In response, non-cognitivists might claim that there is one crucial line of evidence in favor of the view, and it’s a line of evidence that we have already seen. In the previous section, I surveyed studies suggesting that morality is affect-laden. At the start of this section, I said the non-cognitivism is orthogonal to affect-ladeness, but some non-cognitivists would vehemently disagree with this. They would say that non-cognitivism follows from affect-ladeness. Emotions are traditionally regarded as feelings, and feelings are not traditionally believed to be representations of anything. If the thought that killing innocents is wrong is really a bad feeling about killing, then why think this thought has any truth conditions? Does a feeling of indigestion or irritation really refer?

This move might have been compelling in the early part of the twentieth century, but the last fifty years of emotion research have emphasized the intentionality of affect. Some philosophers have adopted cognitive theories of the emotions, according to which emotions are identical to judgments. Elsewhere I have argued against such theories, in favor of the view that emotions are bodily feelings (Prinz 2004), but contemporary feeling theorists still insist that emotions aim to refer. Feeling sad, for example, can be understood as a downtrodden bodily state that represents loss. To say that the feeling represents loss is to say that it has the function of arising in response to losses, and hence carries the information that there has been a loss to a person who experiences it. In a like manner, pain may indicate tissue damage and fatigue may indicate energy depletion, even though pain and fatigue are bodily feelings. None of these feelings are arbitrary. They prepare an organism to cope with specific conditions or events. Emotions qua feelings are in the business of keeping us abreast about how we are faring. Each emotion has a different significance, and any one of them can misfire. I might be sad when there is no loss, or frightened when there is no threat. Such emotions would qualify as erroneous.

If emotions are in the business of representing, then there is no difficulty supposing that moral judgments are truth-apt. When we sincerely assert that, “Killing innocents is bad,” we express a negative feeling towards killing, and that feeling functions as a kind of visceral predicate. It attributes a property to killing (I will have more to say about this property below). In this sense, moral discourse may be much like other forms of emotional discourse. If we say that some food is icky, we express a feeling, while also attributing a property. For example, the feeling of ickiness might represent the property of noxiousness, or perhaps something more subjective, such as the property of causing nausea in the speaker. Someone who calls something “icky” need not know what property that feeling represents, but most language users probably recognize that in using this term we are attempting to say something about whatever it is that elicits the feeling. By analogy to “icky,” moral assertions can be understood as both expressive and predicative. It is a mistake, based on overly simplistic theories of emotions, to assume that feelings cannot play a semantic function. Once we see that feelings can represent properties and function as predicates, non-cognitivism no longer looks like a serious option.

2.3 Realism vs. the error theory

It is one thing to say that moral assertions aim to represent and quite another to say that they succeed in doing so. It is possible that when we say that an action is immoral, we aim to ascribe a property to it, but we do not succeed in doing so. This is precisely what defenders of the error theory have claimed. So, even if the forgoing case for cognitivism succeeds, we must now descend the decision tree and ask whether moral judgments are ever true.

The error theory, which states that moral judgments are truth-apt but always false, was first promulgated by J. L. Mackie (1977). Mackie’s argument begins with the premise that moral predicates aim to represent properties with two important features. The first is objectivity: moral properties are supposed to be the kinds of things that can obtain independent of our beliefs, desires, inclinations, and preferences. The second is action-guidingness: moral properties are supposed to be the kinds of things that compel us to act when we recognize them. Mackie’s second premise is that these two features are difficult to reconcile. Objective properties are usually the kinds of things about which we can be indifferent. Mackie uses the term “queer” to describe properties that are both objective and action-guiding, and he also suggests that such queer properties would require an odd epistemology. For these reasons, he thinks we shouldn't postulate objective action-guiding properties. But, Mackie thinks that moral concepts commit to the existence of such properties, and, thus, that moral judgments posit properties that don’t exist. Therefore, moral judgments are systematically false.

In recent years, the error theory has become popular among evolutionary ethicists (Ruse 1991; Joyce 2006). Mackie’s theory leaves us with a puzzle. Why do people make moral judgments if they are incoherent? Evolutionary ethicists purport to have an answer. They say that morality is an illusion that has been naturally selected because it confers a survival advantage. For example, if we believe that cheating others is objectively bad and that belief is action-guiding, then we will hold others accountable when they cheat, and we will resist cheating even when it might seem advantageous to do so. This reduces the likelihood of free riders and leads to an evolutionarily stable strategy—one that can foster cooperation and collective works. Evolutionary ethicists also typically endorse sentimentalism, suggesting that moral emotions have evolved to motivate such things as punishment and altruism. Mackie himself is not explicit about the role of emotions in his view, which makes it unclear what he means when he says that we perceive the discovery of alleged moral facts to be action-guiding. The link between judgments and emotions, emphasized by evolutionists, provides one answer.

The evolutionary addendum to Mackie’s argument may look like an empirical reason for siding with the error theory. Natural selection is a well-confirmed process, emotions have some basis in evolution, and evolutionary models confirm that emotionally-grounded moral instincts would be adaptive. But there are empirical reasons for doubting the evolutionary story, and for doubting the key premises in Mackie’s argument. Consequently, I think the case for the error theory fails.

The evidence for an evolved moral sense is underwhelming. A thorough critique cannot be undertaken here, but let me offer two broad reasons for doubt (for more discussion, see Prinz 2007a). First, there is little evidence for a moral sense in closely related species. Recall that moral judgments are underwritten by emotions such as anger, disgust, guilt, and shame. There is no evidence that the last three of these emotions exist in chimpanzees, and the anger they exhibit might better be described as reactive aggression, because there is little reason to believe chimps form robust tendencies to be angry about third party offences when they are not directly involved. Evolutionists point out that chimps engage in reciprocal altruism, and other forms or prosocial behavior, but these behaviors may not depend on any moral judgments. Indeed, psychopaths engage in reciprocal altruism (Widom 1976), and chimps often behave in ways that seem psychopathic; they can be extremely violent (Wrangham 2004) and indifferent to each others welfare (Silk et al. 2005).

Evolutionary ethicists might concede this and argue that morality evolved in the human species after we split from other primates. But this position is vulnerable to a second objection: there is good reason to think that morality in humans is learned. Moral judgments derive from emotions that originate outside the moral domain, such as disgust, which is first applied to noxious agents and later expended to the social domain, through conditioning (Prinz 2007a). Even guilt and shame may be learned byproducts of non-moral emotions: shame is related to embarrassment and guilt may be a blend of sadness and anxiety brought on by violating a social norm (Prinz 2005). These emotions and their range of application depend on extensive conditioning in childhood. Moral variation across cultures is considerable, as we will see, and shared moral values can be attributed to widespread constraints on building a stable society (for example, stable societies must prohibit wanton murder within the in-group). Moreover, there is no poverty-of-the-stimulus argument for morality; children receive ample “negative data” in the form of punishment, and they directly imitate values in their communities. As I argue in greater detail elsewhere, arguments for innate moral norms have been unconvincing (Prinz 2007a). This suggests that morality is learned, not evolved.

If morality is acquired through learning, then one cannot bolster Mackie’s argument by assuming that morality is the product of evolution. This alone does not undermine the error theory, however. Error theorists might abandon the evolutionary approach and try to explain systematic error by appeal to a learning story. There is some evidence that people tend to treat certain rules as universally binding, regardless of operative conventions. When asked whether it would be okay to hit a classmate if the teacher granted permission, children tend to say “no.” Turiel (1983, Ch. 7) who made this discovery, denies that such objectivist leanings are innate. Rather, he thinks children learn to distinguish moral and conventional rules. Some subsequent authors have argued that the learning in question involves emotional conditionism (Blair 1995; Nichols 2004). Moral rules are aquired through the inculcation of emotions such as anger, guilt, and shame. There are strong negative feelings associated with hitting that don’t disappear when children imagine the teacher saying it is okay to hit. Violating social conventions may lead to other emotions, such as embarrassment, but these are mitigated when we move from one social setting to another. For example, wearing a hat at the dinner table might be frowned on in some circumstances, but not when wearing a birthday hat at a birthday party. The idea that moral rules are learned by emotional conditioning could also explain their motivational impact; emotions impel us to act, so emotionally grounded rules seem to carry practical demands. This analysis would explain both features emphasized by Mackie—action-guidingness and objectivity—without assuming that moral rules actually are objective. Thus, the error theory could get off the ground without assuming that morality is a product of evolution.

On closer scrutiny, however, this argument is not strong enough to rescue the error theory. It conflates objectivity with authority independence. It is true that children think hitting is wrong even when it is permitted, but that does not mean they think moral truths exist independently of subjective responses. Many of our subjective responses seem independent of what authorities happen to say—our preferences for food and music, for example. But we don’t necessarily infer that these things are objective. So it is a further empirical question whether objectivity is an essential feature of how we understand moral properties.

This brings us to the heart of Mackie’s argument. Should we grant his first premise that moral assertions entail objectivity? Empirically, the answer is a bit messy. When polled, many people assume that morality is objective, but many reject this assumption (Nichols 2004; Goodwin & Darley 2008). In survey studies, there is a nearly even split between objectivists and their opponents. Strikingly, belief in objectivity correlates with religiosity. Goodwin and Darley report that religious beliefs were the strongest predictor of objectivity that they were able to find. This suggests that beleif in objectivity is not an essential part of moral competence, but is, rather, an explicitly learned add-on that most often comes with religious education. The authors also found that belief in objectivity goes down in cases of moral issues about which there is considerable public debate, such as abortion. This might be interpreted as showing, again, that objectivity is not a conceptual truth about the moral domain, but rather a negotiable add on, which can be abandoned in light of counter-evidence. Faith in objectivity goes up with certain relgious beliefs (e.g., divine command theory), and goes down when confronted with the fact that decent, intelligent people have very different moral convictions. In Quine’s terms, moral objectivism, when it is found, may be collateral information rather than an analytic truth—a belief about morality that we are willing to revise.

To test this hypothesis, I conducted a survey study in which I compared a moral predicate (immoral) to two natural kind terms (beetle and tuberculosis), which paradigmatically aim to designate objective properties, and to two terms that are often said to represent secondary qualities (red and humorous). If natural kind terms have a presumption of objectivity, then any threat to that presumption should lead people to conclude that those terms don’t refer. Things are a little trickier with terms such as red and humorous: many people believe that they designate objective properties, but are willing to give up this assumption when presented with countervailing evidence. When told that there is no unifying essence to humor, people do not conclude that nothing is funny; they conclude that humorousness is a property that depends on our responses. In other words, objectivity is not analytically entailed by humorous or red. It is collateral information. My study was designed to see if immoral followed this same pattern.

A group of college undergraduates read the following vignette for the immoral case, with comparable vignettes for the other terms:

Suppose scientists discover that there are two kinds of things that people call immoral. Would it be better to say:
(a) The term “immoral” is misleading, and it might be better to replace it with two terms corresponding to the two kinds of cases.
Or
(b) The fact that there are different cases is interesting, but doesn’t affect the word. The fact that we react the same way to these two things is sufficient for saying they are both members of the same category; they are both immoral.

When given these options, 75% chose option (b) for immoral, resisting the first option which is tantamount to an error theory. Exactly as many chose option (b) for red, and a few more picked (b) for humorous (90%). In contrast, (a) was the dominant answer for the natural kind terms, tuberculosis and beetles (55% and 65% respectively). This suggests that people do not treat moral terms the way that they treat natural kind terms. Even if many people happen to think that morality is objective (as the studies by Nichols 2004, and Goodwin & Darley 2008, suggest), they are willing to give up on this belief without abandoning their moral concepts. They are willing to treat those concepts as response-dependent.

I think these results can be best interpreted as follows. Moral concepts are neutral about moral objectivity. People can acquire these concepts without any beliefs about what kinds of properties they designate. This neutrality begets a kind of resistance to error. If there are no objective moral properties, then it wouldn’t follow that moral judgments fail to refer; it would mean only that they refer to response-dependent properties. Thus, it is all but guaranteed that some moral judgments will come out true, and to this extent the evidence favors moral realism (defined as the view that there are truthmakers for some moral judgments). Mackie mistakes a popular but dispensable belief about morality for an analytic truth. His error theory rests on an error. In fact, his argument for the error theory may rest on two mistakes, the second of which we will come to presently. Of course, this is just one study, and other interpretations may be available, but it provides some evidence against Mackie’s conceptual claim and shows how empirical findings might be used to explore whether moralizers are, as he suggests, committed to objectivism. Extant empirical evidence suggests otherwise.

2.4 Sensibility vs. moral sense

The survey study just described suggests that one can possess moral concepts without knowing whether moral judgments refer to properties that are objective. The survey also brings out the possibility that people are willing to accept the conclusion that moral truth depends on our responses. But the survey does not settle whether a response-dependent theory is true. This is the next question on the decision tree. As we have seen, Mackie thinks action-guidingness and objectivity are incompatible. This may suggest that he sees no room for a theory that combines moral objectivity with the view that moral judgments have motivational pull. This, however, is Mackie’s second mistake. The hypothesis that morality has an emotional basis reveals a way out of Mackie’s argument for incompatibility. Emotions are action-guiding in that they motivate us to act. But some emotions may also represent objective features of the world. Fear, for example, may represent danger, and danger may be an objective property. Emotions can represent objective properties in a motivating way: they simultaneously pick up on information while compelling us to respond adaptively. The fact that fear is action-guiding does not rule out the possibility that it is designed by evolution to track objective threats. Likewise, disgust is action-guiding but it may register real sources of contamination.

This brings us back to “icky.” This emotionally-expressive term may refer to something objective, like contamination, or to something subjective, such as the tendency to cause feelings of nausea. We can ask whether ickiness is objective or subjective, even if we grant that the word “icky” is expressive. Expressive terms can have objective referents. Likewise, we can ask this question about moral terms. This question frames a historical debate between Francis Hutcheson, who may have believed that our moral sentiments track objective moral truths, and David Hume, who suggests that morality depends on human responses. The claim that moral judgments track objective properties is called the moral sense theory. It seems to have been defended by Francis Hutcheson in the eighteenth century. It may even have been Kant’s considered view, since he had an objective procedure for arriving at moral truth, but also insisted that every moral judgment is associated with a moral feeling. The moral sense view finds an analogue in contemporary authors who combine external standards of moral truth with motivationally charged moral psychologies (e.g., Campbell 2007; Copp 2001; see also Railton 2009, who makes a modest move in that direction). The alternative view, which says that moral judgments refer to response-dependent properties, has been called the sensibility theory (McDowell 1985; Wiggins 1987). We can now ask whether there is any way to decide between these options empirically.

I think there is some reason to favor sensibility over moral sense. For the moral sense theory to be true, there would have to be a candidate objective property to which our moral concepts could refer. Unfortunately, I cannot undertake a review of modern moral sense theories here, but I will offer, instead, a more general line of empirically-informed resistance. Moral rules are emotionally conditioned, and communities condition people to avoid a wide range of different behaviors. Within a given society, the range of things that we learn to condemn is remarkably varied. Examples include physical harm, theft, unfair distributions, neglect, disrespect, selfishness, self-destruction, insults, harassment, privacy invasions, indecent exposure, and sex with the wrong partners (children, animals, relatives, people who are married to other people). One might think that all of these wrongs have a common underlying essence. For example, one might propose that each involves a form of harm. But this is simply not true. Empirical evidence shows that people condemn actions that have no victims, such as consensual sex between adult siblings and eating the bodies of people who die in accidents (Murphy et al. 2000). Furthermore, harm itself is a subjective construct. It cannot be reduced to something like physical injury. Privacy violations are regarded as a kind of harm, even though they don’t hurt or threaten health, whereas manual labor is not considered a harm, but it threatens the body more than, say, theft. Similar problems arise if we try to define moral wrongs in terms of autonomy violations. Mandatory education violates autonomy, but it is considered good, and consensual incest is an expression of autonomy, but is considered bad.

Realists would no doubt resist some of these claims, but theirs is an uphill battle. On the face of it, morality lacks a common denominator. Empirical surveys of human values suggest that moral rules are a potpourri, which can be extended and contracted in any number of ways, with no fixed ingredients. Or rather, the common denominator is not a property shared by the things we condemn, but rather by the condemning itself. Moral sense theorists liken morality to perception, and, in so doing, they imply that there is an external feature of the world that our moral sentiments pick up on. But there is little reason to believe this. Unlike perception, there is massive variation in what we moralize, and there is a perfectly good explanation for this: the content of morality is determined by social conditioning rather than by the mind-independent world. Morality is not something we get by simply observing.

The foregoing is offered as an empirical challenge to moral sense theories, not a decisive refutation. Too often philosophers stick with examples of moral norms that clearly concern harm or violations of autonomy. This inflates optimism about a unifying essence. If one uses empirical methods to discover the full range of things that people actually moralize (such as victimless harms), the task of finding a unified essence looks much harder. Moral sense theorists might reply that this diversity is illusory. They might say, for example, that people would stop condemning victimless crimes on reflection. That claim is amenable to empirical testing, and so far the tests provide little support. For example, Murphy et al. (2000) presented people with cases of incest and cannibalism where it was extremely salient that no one was harmed. They invited people to revise knee jerk moral intuitions and rule that, on reflection, these victimless actions are permissible. A piddling 20% revised accordingly, but 80% stuck to their original view. Moral sense theories seem to place their bets on the 20%. The challenge is to explain why the stubborn and considered opinions of the majority are performance errors of some kind.

Given the diversity of things about which people moralize, I think the sensibility theory is more promising than the moral sense theory. Wrongness is projected, not perceived. The property of being wrong is the property of causing negative sentiments, not a response-independent property that those sentiments are designed to detect. This conclusion follows from an inference to the best explanation. Empirically it looks as if there is no common essence to the things that we find morally wrong—a finding that is difficult to explain on the moral sense model, but easy to explain on the assumption that wrongness is response dependent. By analogy, imagine that we catalogue the things that make people laugh, and find that they lack a shared essence. This would imply that laughter does not pick up on an objective property. The things that we find funny are unified by the very fact that we are amused by them. Likewise for the things we find immoral: disapprobation carves the moral landscape.

2.5 Relativism vs. ideal observers

I have just been arguing that moral truth is response-dependent. Moral judgments can be true, but their truth depends on our sentiments. Something is immoral if it causes anger, disgust, guilt, and shame in us. But now we can ask, who does “us” refer to here? Whose sentiments determine moral truth? This brings us to the final question in the metaethics decision tree. Can divergent responses have equal claim to truth?

Empirical evidence strongly suggests that moral sentiments vary, both within and across cultures. Within a culture, the clearest divisions are between political orientations. Liberals and conservatives have interminable debates, even when they are exposed to the same science and education. Research suggests that these debates come down to fundamental differences in moral values. Conservatives are much more likely than liberals to emphasize purity, authority, and preservation of the in-group in justifying their moral norms (Haidt 2007). These things are foundational for conservatives and largely irrelevant to liberals.

Across cultures, differences are even greater. Everything that we condemn is accepted somewhere else (such as slavery and torture), and things that have been condemned by other cultures (such as women’s suffrage) have been embraced by us. There are cultures whose moral outlooks are dominated by considerations that we tend to downplay in the post-industrial West (sanctity and honor, for example), and ideals that are central to our moral outlook appear to be modern inventions (rights and the idea of human equality).

Descriptively, then, people do not seem to have the same moral values, within or across cultures. There is divergence in our sentiments. Some of this divergence might diminish if we filtered out cases where people were reasoning badly or on poor evidence, but there is ample evidence that disagreements remain among people who reason carefully and draw on the same factual knowledge. Indeed, if we filter for good reasoning, divergence might increase rather than decrease: consider professional normative ethicists, who are experts at reasoning but nevertheless arrive at varied and novel moral perspectives that neither converge with each other nor with the communities to which they belong.

I think such descriptive moral relativism provides support for metaethical moral relativism. This would be a terrible inference on its own, as every metaethics textbook points out, but the inference gains plausibility if bolstered by a premise I argued for above: moral truth is dependent on our responses. If responses vary, even under favorable epistemic conditions, and responses determine truth, then the truth of a moral judgment can vary depending on whose values are being expressed.

The ethical universalist can resist this conclusion by offering an antidote to moral variation. The most natural strategy would be to defend universality by developing an ideal observer theory, and to argue that, under ideal epistemic conditions (which might include external factors as well as being an epistemically ideal agent), judges would arrive at the same set of moral values. This strikes me as woefully unlikely. Once we grant that sentimentalism is true, and that our sentiments track response-dependent properties, it’s not clear how to settle on which observer is ideal. Two people who have the same factual knowledge may have different sentiments as a result of differences in temperament (Lovett et al. 2012), reward sensitivity (Moore et al. 2011), gender (Fumagalli et al. 2010), class (Côté et al. 2013), and age (Truett 1993). Whose sentiments are right? Moreover, the standard traits associated with ideal observation may be problematic in the moral domain. Should we consult someone who is disinterested when we know, empirically, that distance from a situation can lead to moral indifference? Should we consult someone who has not been conditioned by a particular culture when we know that innate sentiments are unlikely to deliver moral attitudes? Should we consult someone who attends to every detail of a case, when we know that framing, vivid description, and concreteness can alter moral judgments? These problems strike me as insuperable. There are no clear criteria for ideal observation and no reason to believe that careful observers would converge.

In posing this challenge, I am inviting ideal observer theorists to look at empirical findings and propose epistemic standards that would overcome the sources of variation mentioned here. Some ideal observer theories try to be empirically responsive in this way. For example, Smith (1994) advances the hypothesis that ideal rational agents would converge, but he also realizes that some readers might be reluctant to share his optimistic outlook. To quell these doubts he makes three empirical observations (p. 188): there is considerable moral convergence already (he cites the existence of thick concepts as evidence: we all think brutality is bad and honesty is good); there has been moral progress (he cites slavery, among other examples); and entrenched disagreements often reflect faulty rationality, such as religious beliefs. Here, I think further empirical scrutiny would weaken Smith’s case. Divergence is rampant, and people disagree on the scope of thick concepts (is torture brutal? is espionage dishonest?). Cases of (what we consider to be) moral progress are, I’ve noted, often driven by economic upheavals and other irrational factors, with reasoning playing a post-hoc role. Finally, disagreements remain after bad reasoning and religiosity are controlled for; the examples mentioned, in formulating the challenge include things such as temperament and framing effects. I think empirical evidence provides little reason to expect that rational and informed observers would deliver consistent verdicts.

In light of such worries, universalists might abandon the ideal observer theory and offer instead a procedural approach to consensus, arguing that people would and should converge if they arrived at their sentiments in the right way. For example, many people might agree that it is good to arrive at decisions democratically, taking multiple sentiments into consideration, and we might sentimentally endorse the outcome of democratically-resolved moral disputes. Though I cannot make the case here, I suspect the problems with such a procedural approach outweigh its prospects. Democratic decision-making does not result in moral consensus; it can even polarize. When such procedures increase consensus it is often through power and prestige rather than sentimental convergence. Our faith in democratic procedures may also be an expression of moral relativism rather than a solution. Democratic procedures are an historical anomaly, which emerged in the modern period with the rise of capitalism, and they have often been used to oppress minorities and to impose the values of the many over the few. Perhaps such procedures are an improvement over totalitarian forms of decision-making, but they do not remedy relativism. Indeed, as societies move towards consensus-building procedures, they may actually promote variation, leading to an endless proliferation of values and an ever widening gulf between those who cherish diversity and those who reside in more traditional societies. From a social science perspective, the prospects for a universal morality look grim.

Once the case for relativism is established, the question arises: relative to what? Are moral judgments relative to value systems? Are those systems individuated at the scale of cultures and subcultures or do they vary across individuals? Little empirical work has been done to address this question, but let me end with a suggestion about how to proceed. When examining the semantics of natural kind terms, philosophers have sometimes appealed to a linguistic division of labor (Putnam 1975). We defer to experts and thereby license them to adjudicate the boundaries between natural kinds. Now we can ask, is there such a thing as moral expertise? Do we appeal implicitly or explicitly to moral experts? Would we change our moral judgments if the designated members of our community told us we were morally mistaken? We don’t know the answers to such questions, because moral expertise has not been intensively studied. I suspect there will be considerable individual differences, with members of more traditional societies showing more willingness to defer. But I also suspect that deference in the moral domain will be less prevalent than for natural kinds; we are more inclined to take ourselves as having morally authoritative insight. What is most clear, however, is whether the scope of the relativity depends ultimately on how we use moral concepts and terms; and this is something that can be investigated empirically. Naturalizing relativism will require the marriage of cultural anthropology and sociolinguistics. From the armchair, it is tempting to think there is a single true morality; introspective reflection tends towards solipsism.