1. Animal consciousness and welfare
What is the relation between consciousness and animal welfare? Let us note from the outset that there are various things one might mean by “consciousness.” My main focus here is subjective experience, or what philosophers tend to call phenomenal consciousness. To say that a state of an animal is phenomenally conscious is to say that there’s something it feels like to be in that state (Block Reference Block1995). Ideally, we would have a better definition than this, but phenomenal consciousness is a notoriously difficult-to-define property, one that resists definition in explicitly functional terms. What we can do is point to examples (Schwitzgebel Reference Schwitzgebel2016). When you’re in a state of dreamless sleep, in a coma, or under general anesthetic, there’s nothing it feels like to be you. But as we go through our waking lives, there is something it feels like to experience odors, tastes, colors, sounds, pleasures, and pains. This is the sense of “consciousness” that gives rise to the notorious “hard problem of consciousness,” the problem of trying to explain how neural activity could give rise to states that feel like something (Chalmers Reference Chalmers1995).
Animal welfare scientists often take it for granted that consciousness in this sense is intimately related to animal welfare. Definitions of animal welfare typically appeal to sentience, consciousness, experience, subjective feeling, or related ideas. These definitions matter insofar as they shape the aims of the field. For example, Donald Broom (Reference Broom2008) defines welfare as an animal’s “state as regards its attempts to cope with its environment” but adds that “welfare includes health and the extent of positive and negative feelings.” Whereas Broom considers subjective feelings “an extremely important part” of welfare, Ian Duncan (Reference Duncan1993) has argued for a definition of welfare that considers only subjective feelings: “Neither health nor lack of stress nor fitness is necessary and/or sufficient to conclude that an animal has good welfare. Welfare is dependent on what animals feel” (Duncan Reference Duncan1993).
In the 1980s, Marian Stamp Dawkins was a pioneer of this way of thinking about animal welfare, writing that “to be concerned about animal welfare is to be concerned with the subjective feelings of animals, particularly the unpleasant subjective feelings of suffering and pain” (Dawkins Reference Dawkins1988). More recently, however, Dawkins (Reference Dawkins2012, Reference Dawkins2021) has argued for the importance of defining welfare without appealing to consciousness or related ideas. Dawkins now favors a view in which an animal with good welfare is one that is healthy and “has what it wants” (Reference Dawkins2021, 11). I aim to charitably reconstruct what I see as Dawkins’s strongest argument against consciousness-involving definitions of welfare before turning to Dawkins’s alternative and its problems. I reflect on the strain on the concept of animal welfare the debate reveals and consider how the field of animal welfare science should respond.
2. Two unpersuasive considerations
Dawkins has offered various reasons for stripping the concept of consciousness from the concept of welfare, not all of them persuasive. Perhaps the least persuasive is a direct appeal to the hard problem, as if the hard problem immediately led to skepticism about animal consciousness:
Even with our own consciousness, we still do not understand how the lump of nervous tissue that makes up our brain gives rise to private subjective experiences…. And because we do not understand how the human brain makes us conscious, we do not know what to look for in other species to decide if they, too, have conscious experiences like us. Perhaps they do, but how would we know? (Dawkins Reference Dawkins2021, 5)
This is too quick. If evidence-based attributions of consciousness relied on solving the hard problem, we would not be able to make evidence-based attributions of consciousness to other humans. Although there may be limits to the level of certainty we can achieve in the human case, human verbal reports of experiences provide good abductive evidence—if they did not, there could be no science of human consciousness. The key question is whether nonverbal indicators provide enough evidence to get a parallel science of animal consciousness off the ground. The mere existence of the hard problem does not imply that they cannot.
Also unpersuasive is an appeal to substantial disagreement in the science of human consciousness. As Dawkins notes, there is great disagreement about the correct theory of consciousness, with currently fashionable options including the global workspace theory (Mashour et al. Reference Mashour, Roelfsema, Changeux and Dehaene2020) and the integrated information theory (Oizumi, Albantakis, and Tononi Reference Oizumi, Albantakis and Tononi2014), and great disagreement about the neural basis of human consciousness, too, with proponents of different theories tending to favor different brain areas (Boly et al. Reference Boly, Massimini, Tsuchiya, Postle, Koch and Tononi2017; Mashour et al. Reference Mashour, Roelfsema, Changeux and Dehaene2020). Yet substantial disagreement about the nature of a property is not a compelling reason to deny that the property is a constituent of welfare. If disagreement were itself a reason, we would also need to avoid defining welfare in terms of preferences or health because there is plenty of disagreement about the nature of these properties too.
3. The challenge from unconscious affect
Dawkins, however, also has an empirical motivation for defining welfare in a consciousness-free way, and I see this as the strongest part of her case. Dawkins (Reference Dawkins2021, 9–11) points to a body of work on human affect in which conscious and unconscious motivational states are apparently dissociated (LeDoux and Pine Reference LeDoux and Pine2016).Footnote 1 In a typical experiment of this genre, an emotionally salient stimulus (e.g., an angry or happy face) will be presented to subjects in two ways: subliminally (i.e., below the subjective threshold of conscious perception) and supraliminally (i.e., above that threshold). It turns out that some arousal responses interpreted as “fear responses,” such as breaking out in a sweat (as measured by skin conductance), are evoked by the subliminal stimulus. When the stimulus is supraliminal, these responses tend to be weaker, perhaps indicating top-down inhibition of arousal (Tamietto and de Gelder Reference Tamietto and de Gelder2010). Moreover, in cases of “affective blindsight,” a subject is unable to consciously perceive the presented face as a result of damage to the visual cortex but is nonetheless able to guess its valence (e.g., happy or angry) at a better-than-chance rate. Evidence of this type has led Joseph LeDoux and collaborators to posit a “two-system” model of affective processing, in which subjective feelings depend on a cortical pathway, whereas “defensive survival circuits” can be activated by a subcortical, entirely unconscious pathway (LeDoux and Pine Reference LeDoux and Pine2016; LeDoux and Brown Reference LeDoux and Brown2017).
The evidence for the two-system model is suggestive but not overwhelming. Granting that the stimuli are indeed unconsciously perceived (cf. Peters et al. Reference Peters, Kentridge, Phillips and Block2017), a remaining methodological challenge is to show that the entire pathway from stimulus to defensive response, and not just the perceptual part of it, occurs nonconsciously. The possibility of unconscious vision eliciting amygdala activity that is experienced as conscious affect before producing behavioral consequences is hard to rule out.
Indeed, there is some support for this latter possibility from cases in which subjects with affective blindsight have been able to verbally report the valence of the feeling evoked by an unseen stimulus in addition to guessing correctly in forced-choice tasks (Anders et al. Reference Anders, Birbaumer, Sadowski, Erb, Mader, Grodd and Lotze2004). There is also some evidence pointing the other way from a study in which some participants (presented with subliminal faces) were instructed to use their feelings as a guide to the valence of the face and yet failed to perform any better than controls (Bornemann, Winkielman, and van der Meer Reference Bornemann, Winkielman and van der Meer2012). However, although it is important to LeDoux and colleagues’ case, the evidence from this experiment is weak because only 4/19 subjects in the group instructed to use their feelings actually attempted to do so, according to a postexperiment questionnaire (Bornemann et al. Reference Bornemann, Winkielman and van der Meer2012, 121).
The evidence for the two-system view of affect need not be overwhelming or even strong, however, to present animal welfare scientists with a challenge. We need only grant that it is credible. To the extent that it is credible, then it is credible that many defensive responses to stimuli in animals do not indicate conscious affect. We should take seriously the possibility that such responses in nonhuman animals are controlled by an analogue of a subcortical pathway that is nonconscious in humans. Intuitively, the idea that the conscious pathway could be absent altogether deserves increasingly serious consideration as the evolutionary distance from humans increases. Even among mammals, there are some differences in neocortical organization, but nonmammals do not possess a neocortex at all, so they either have no conscious pathway or else have evolved a different neural route to the same result.
I see this as the basis for Dawkins’s most persuasive argument. Without reliable markers of conscious as opposed to unconscious affect, or a reason to consider unconscious affect unlikely, we have no way of confidently validating putative indicators of subjective feelings. We can reconstruct the argument like this:
The challenge from unconscious affect:
-
1. Animal welfare should be defined in such a way as to make it feasible (with current methods) to validate animal welfare indicators with high confidence.
-
2. We do not currently have reliable methods for distinguishing conscious from unconscious affect in animals.
-
3. If we define animal welfare in terms of conscious affect, then in the absence of reliable methods for distinguishing conscious from unconscious affect, we will be unable to validate animal welfare indicators with high confidence.
-
4. Animal welfare should not be defined in terms of conscious affect.
To give one example, a currently popular indicator of animal welfare is judgment bias, where an animal is presented with an ambiguous stimulus and its response is taken to indicate an “optimistic” or “pessimistic” expectation of reward (Mendl et al. Reference Mendl, Burman, Parker and Paul2009). Optimistic expectations are taken as a sign of good welfare, whereas pessimistic expectations are taken as a sign of poor welfare. Dawkins’s basic point is that for all the talk of “pessimism” and “optimism,” we have no reason to think this indicator tracks subjective feelings rather than unconscious affective states, and we have no reliable way of finding out (Dawkins Reference Dawkins2012, 109–10). So, if welfare constitutively involves subjective feelings, such an indicator cannot be confidently validated.
4. What’s the alternative?
Suppose we accept this argument. What is the alternative? Dawkins (Reference Dawkins2021) proposes that we define animal welfare in terms of health and “valenced states,” where the concept of a valenced state is characterized functionally in a way that is neutral between conscious and nonconscious processing.
The idea is that all animals have at least a minimal capacity for preference. When they are presented with options, their behavior will often show a systematic bias. Moreover, there will typically be some capacity for reinforcement learning: some states will lead to positive reinforcement and others to negative reinforcement. Although this departs from current usage, we could stipulate that the terms positive valence and negative valence are to refer to the functional significance of a state for preference and reinforcement learning without saying anything about whether or not the state is subjectively felt. Then we could say that good welfare for an animal consists of its (a) being healthy and (b) having many positively valenced states and few negatively valenced states. The best balance of positive and negative valence is open for debate because negative valence has value for learning. I take this to be the idea that is captured in Dawkins’s slogan: an animal has good welfare if it is healthy and “has what it wants.”
Intuitive motivation for the view comes from cases such as a bird that flies continually into the bars of its cage (Dawkins Reference Dawkins2021, 8, 49). As Dawkins argues, it is obvious that the bird has compromised welfare, but it is not obvious that the bird is experiencing conscious affect. The proposed account of welfare can make sense of this because it is obvious that the bird is in a state that motivates escape behavior. The intuition is that we should not want a science of animal welfare that makes it difficult, perhaps even impossible, to assess confidently whether the bird’s welfare has been compromised.
5. The normative challenge to Dawkins’s alternative
Dawkins’s alternative has a significant downside: it does not fulfill the normative role of the concept of animal welfare. It fails what we might call the criterion of “normative aptness.” Animal welfare science aims to achieve normatively significant results. It is what David Fraser (Reference Fraser2008, following Salter Reference Salter1988) has called a “mandated science”: a science with a societal mandate to inform animal welfare policies and improve welfare. I contend that if animal welfare science is to fulfill this mandate, then the construct of animal welfare should denote a property that deserves high ethical priority and is a genuine source of ethical constraint on our treatment of animals.Footnote 2
To assess whether a construct can play this role, we cannot avoid substantive ethical questions. We cannot place a firewall between the science and ethics of animal welfare. We must ask which properties are the source of ethical constraints. I contend that health and valence (in the previously described sense) are not enough: they miss the distinctive importance of experienced valence. If a valenced state is not experienced by the animal, then it either has no ethical significance at all or else it has far less significance than an experienced state. This is implied by ethical theories that give a central role to subjective experience (e.g., hedonic utilitarianism and Korsgaard’s [Reference Korsgaard2018] neo-Kantian view), but I take it to be a principle that is more plausible than any particular theory.
As a motivating example, consider Paramecia aurelia, a single-celled ciliate. A paramecium has no nervous system, but it does have health, and it has states that drive defensive/escape responses—valenced states, in Dawkins’s stripped-back sense. As Michel (Reference Michel2019) has noted, “if a paramecium encounters a potentially dangerous concentrated salt solution or acetic acid, it will back away and swim in a different direction or engage in defensive behavior by discharging trichocysts” (Michel Reference Michel2019, 2418). Yet failing to consider these states is either not an ethical failing or, at worst, it is a failing far less grave than failing to consider the states of a conscious animal. The presence or absence of consciousness makes a large ethical difference.
If consciousness were ethically significant in theory yet irrelevant to current animal welfare policy, it might be tempting to dismiss the practical importance of its omission. But consider the question of whether to protect any invertebrate animals—and if so, which ones. Cephalopod mollusks (such as octopuses) are protected in science in the EU and UK but not in any nonscientific context. Switzerland recently brought in legislation to ban the live boiling of decapod crustaceans (such as crabs and lobsters), and my team’s recent report to the UK government recommended the same move (Birch et al. Reference Birch, Burn, Schnell, Browning and Crump2021). The report led to the UK legally recognizing the sentience of cephalopods and decapods in the Animal Welfare (Sentience) Act 2022, but not (so far) to a ban on controversial practices such as live boiling. It is part of the mandate of animal welfare science to inform policy decisions like these by providing evidence about which slaughter and processing methods create high risks of experienced suffering. There is no escaping the entanglement of these questions with questions of consciousness.
In short, Dawkins’s proposed definition sacrifices normative aptness to enable confident validation of welfare indicators with current methods, whereas definitions that emphasize subjective feeling opt for normative aptness at the cost of making welfare indicators challenging to validate. We would ideally like a definition that is normatively apt and reliably tracked by well-validated indicators, but in fact, these two considerations seem to pull in opposite directions, at least given our current state of knowledge (Figure 1). Moreover, this is not merely a semantic dispute—it is a clash between two ways of thinking about the aims and priorities of animal welfare science.
6. Benefits and risks of conceptual splitting
How to resolve such a dispute? I think it may help to reflect on an analogy with the concept of consciousness itself and its role in human consciousness science. Unlike animal welfare science, human consciousness science is not a mandated science; it does not generally aim to inform policy decisions. But it does face a comparable problem, in that disputes about the meaning of its central term reflect bigger-picture disputes over the aims and priorities of the field.
Ned Block (Reference Block1995) highlighted one persistent source of strain. To study consciousness, scientists rely on the idea that verbal reports and other voluntary actions will reliably track conscious states. This pulls in the direction of defining consciousness in a way that would secure this link with voluntary reporting. Yet what brings scientists to the topic of consciousness in the first place is often a drive to understand “what it’s like” to be a conscious subject—a drive to address the hard problem. This pulls in the direction of defining consciousness in a way that puts the hard problem and “what it’s like” center stage, avoiding any a priori commitments regarding functions.
Block’s solution was to split the concept of consciousness (Figure 2). Access consciousness (sometimes also called conscious access) refers to the availability of information for planning, reasoning, and voluntary action, whereas phenomenal consciousness refers to “what it’s like,” subjectively, to be you. Block claimed to be analyzing the ordinary concept, but the move is often viewed in retrospect as a piece of conceptual engineering rather than analysis (Chalmers Reference Chalmers2020). The distinction is now widespread, but not universal, in the field. A similar splitting move is possible for animal welfare. We could take Dawkins’s concept and relabel it functional welfare while relabeling a purely experience-based definition such as Duncan’s as experienced welfare (Figure 3).
But should we? I think we could expect a similar profile of risks and benefits to those of the access/phenomenal split. In consciousness science, the split has facilitated the development of detailed mechanistic theories of conscious access. The global neuronal workspace theory is usually presented by its proponents as a theory of access (Mashour et al. Reference Mashour, Roelfsema, Changeux and Dehaene2020). The hypothesis that it is also a theory of phenomenal consciousness is a further step beyond the theory’s core commitments (Carruthers Reference Carruthers2019). More generally, the phenomenal/access distinction makes it possible to debate how phenomenality relates to access and, in particular, whether phenomenality “overflows” access. This has led to a rich and complex empirical controversy (Phillips Reference Phillips2018).
The conceptual split has, perhaps unintentionally, induced a subtle division of labor. Scientific work exploring the mechanisms of conscious access can now be conducted, and published, with little or no discussion of whether the work makes any progress toward understanding phenomenal consciousness. That issue can be left for articles reflecting on the wider implications of the scientific work. The adoption of the conscious access terminology by researchers in the global workspace program suggests they, at least, find this division of labor productive. Others who disavow that terminology (e.g., proponents of the integrated information theory) disagree.
Why might one find the division of labor counterproductive? Suppose one thinks that phenomenal consciousness should be the primary object of study of consciousness science but suspects that access/availability of information is unlikely to be the basis of phenomenal consciousness. From this perspective, the construction of an “access consciousness/conscious access” concept is bad engineering: it has led scientific efforts to be misdirected toward access/availability and away from the field’s proper object of study. Moreover, it has encouraged a widespread presumption that the function of phenomenal consciousness must have something to do with access, a presumption that is only occasionally challenged (Scott et al. Reference Scott, Samaha, Chrisley and Dienes2018). Worse, it has encouraged the view that phenomenal consciousness, being allegedly “distinct from any cognitive, intentional, or functional property” (Block Reference Block1995, 230), is beyond the reach of scientific inquiry altogether. That was not Block’s intended message, but it is the message some have taken from his distinction. More speculatively, it may also have encouraged consciousness scientists to neglect the ethical implications of their work, a problem documented by Mazor et al. (Reference Mazor, Brown, Ciaunica, Demertzi, Fahrenfort, Faivre, Francken, Lamy, Lenggenhager, Moutoussis, Nizzi, Salomon, Soto, Stein and Lubianker2021), for it allows the following line of thinking: “Phenomenal consciousness is what matters for ethics, but I study conscious access, so my work has no direct ethical implications.”
Let us turn back to animal welfare. Human consciousness science gives us a case study of how conceptual splitting can affect the trajectory of a research program, providing a basis for thinking about the risks and benefits of a split between “functional” and “experienced” welfare. On the plus side, we could expect to see progress in the direction of more detailed theories of functional welfare and better-validated indicators of it. We could expect a division of labor to emerge between a body of empirical work measuring functional welfare and a separate body of work about how functional welfare relates to experienced welfare (to some extent regimenting a division of labor that already exists informally). We could expect those immersed in research into functional welfare to find the division of labor productive.
On the negative side, if one thinks experienced welfare is the proper object of study of animal welfare and that functional welfare matters only insofar as it correlates with experienced welfare (the view expressed by Duncan Reference Duncan1993), then one will naturally fear that the proposed conceptual split risks dragging the field away from its proper object of study. It will risk creating a default presumption that functional welfare tracks experienced welfare, discouraging serious empirical interrogation of this assumption. Worse, it will encourage the view that experienced welfare, being allegedly distinct from any functional property, is beyond the reach of scientific inquiry. And it risks creating too much distance between the bulk of empirical research in animal welfare science and its normative mandate. If the field’s central empirical project is guided by a welfare concept that is normatively inapt, that is still a serious problem, even if we retain a normatively apt concept for use in explicitly ethical and policy-relevant discussions.
I see the phenomenal/access split as a cautionary tale: on balance, I think the negatives have outweighed the positives. Human consciousness science should have accepted strain on its central concept as a sign of its immaturity as a scientific discipline rather than bifurcating the concept. Splitting the concept is a shortcut to easing strain, but it is a costly one. It is better, on balance, to retain a single concept and tolerate disagreement over its meaning than to split the concept in a way that creates a potentially counterproductive division of labor.
This is, I suggest, a lesson for animal welfare science, another young scientific field in the process of maturing. My tentative recommendation is that the field would be better served by retaining a single concept of animal welfare, defined partly in terms of consciousness—along the lines of Broom’s concept—while recognizing that the demands of empirical validation and normative relevance currently put that concept under severe strain. Over the long run, we need to try to ease the strain by improving our empirical methods so that the challenge from unconscious affect is no longer so severe. Splitting the concept is a tempting shortcut solution but one that comes with significant risks.
Acknowledgments
I thank Kristin Andrews, Heather Browning, Andrew Crump, Simon Fitzpatrick, Simona Ginsburg, Eva Jablonka, Michael Mendl, Elizabeth Paul, Eva Read, Henry Shevlin, Walter Veit, and audiences at the 2021 Philosophy of Science Association (PSA) Meeting and the University of Edinburgh for discussion. This research is part of a project that has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation program, grant number 851145.