-
PDF
- Split View
-
Views
-
Cite
Cite
Camilo Miguel Signorelli, Joanna Szczotka, Robert Prentner, Explanatory profiles of models of consciousness - towards a systematic classification, Neuroscience of Consciousness, Volume 2021, Issue 2, 2021, niab021, https://doi.org/10.1093/nc/niab021
Close - Share Icon Share
Abstract
Models of consciousness aim to inspire new experimental protocols and aid interpretation of empirical evidence to reveal the structure of conscious experience. Nevertheless, no current model is univocally accepted on either theoretical or empirical grounds. Moreover, a straightforward comparison is difficult for conceptual reasons. In particular, we argue that different models explicitly or implicitly subscribe to different notions of what constitutes a satisfactory explanation, use different tools in their explanatory endeavours and even aim to explain very different phenomena. We thus present a framework to compare existing models in the field with respect to what we call their ‘explanatory profiles’. We focus on the following minimal dimensions: mode of explanation, mechanisms of explanation and target of explanation. We also discuss the empirical consequences of the discussed discrepancies among models. This approach may eventually lead to identifying driving assumptions, theoretical commitments, experimental predictions and a better design of future testing experiments. Finally, our conclusion points to more integrative theoretical research, where axiomatic models may play a critical role in solving current theoretical and experimental contradictions.
Models of consciousness constitute hypotheses on how consciousness relates to the physical domain.
Numerous competing frameworks build on different philosophical grounds and favour particular scientific methodologies.
We compare models based on three analytical dimensions: mode of explanation, mechanisms of explanation and target of explanation.
Differences in the conception of explanation have empirical consequences: the futility of localization approaches, the potential of second-person methodologies and different senses of ‘content’ of consciousness.
Future research needs integrative approaches, where mathematical and axiomatic models may play a critical role.
Introduction
Models of consciousness set out to provide a principled description of how the physical domain relates to conscious experience (Seth 2007; Seth 2009; Durham et al. 2020). In the last decades, consciousness researchers put forward an abundance of conceptual and formal proposals, drawing from neuroscience, physics, mathematics, philosophy and experimental psychology. In an early scientific phase, it is natural to expect many competing models to develop in parallel to each other. A more mature stage should entail a substantial cross-talk between them, aiming at distilling critical similarities and differences between them, extracting precise empirical predictions (Boly et al. 2017) and lastly, eliminating falsified frameworks through a set of crucial empirical experiments, as presently envisioned (Reardon 2019; Melloni et al. 2021).
An alternative would be to demand that competing theories need to ‘converge’ to a unified, synthesized account in order to make progress (Northoff and Lamme 2020; Wiese 2020). In the following paragraphs, we posit that there are currently several, serious impediments to both crucial experiments but also convergence approaches. Arguably, more conspicuous and empirically tangible differences in the theories (such as their postulated neural correlates: prefrontal cortex or posterior hot zone) derive from much deeper, implicitly held deviations in theoretical and philosophical assumptions. In particular, proponents of different theories seem to substantially disagree on what would constitute a ‘satisfactory explanation’ of consciousness in the first place. Therefore, the aims of these theories are sometimes different. Once all these discrepancies are fully made explicit, the major models can start to enrich each other in a meaningful way. At least in some cases, theoretical misalignment between the models might boil down to different angles of looking at the same problem.
Although navigating through such a highly diversified theoretical landscape remains challenging, there is hope that one could account for the large variation in the field by using only relatively few axes of comparison. To this date, there have been only a couple of systematic attempts to thoroughly compare the contemporary models of consciousness (Klink et al. 2015; Block 2009; Northoff and Lamme 2020). Arguably, however, all these endeavours have not gone beyond simply collating the theories’ different explanatory targets and their main employed paradigms. In addition, the way the theories have been classified so far [such as targeting either ‘phenomenal’ or ‘access’ consciousness (Northoff and Lamme 2020) or whether they are related to the ‘pre-’ or ‘post-’stimulus neural activity] might have created more confusion than elucidation (Rosenthal 2002a; Block 2007; Rosenthal 2020).
In order to identify the most critical points of contention in the field, we first introduce the philosophical landscape (see The problem of consciousness Section), then we explicate three crucial directions in which frameworks diverge most noticeably with respect to their explanatory pretence (see Method of classification Section). Among these directions, we distinguish mode of explanation (mechanistic vs. unificationist), mechanism of explanation (functional vs. causal) and target of explanation (quality vs. quantity of consciousness). These dimensions are discussed along contemporary models of consciousness, which creates a multi-dimensional explanatory profile for each. We also discuss three empirical consequences of this classification (see Classifying models of consciousness Section). We conclude by advocating for a more integrative approach, hinting at already existing empirical and theoretical (mathematical) tools (see Conclusions: integrative methods for the future Section).
We narrow our scope to 14 influential models. While being aware that our selection of models is not exhaustive, we do hope that our work would still spark meaningful discussions and inspire orderly and structured comparisons in the field.
The problem of consciousness
Philosophy does not solve problems; it rather helps to emphasize and re-conceptualize them, making problems amenable to scientific investigation. Therefore, we first look at some of the main concepts relevant to the question of consciousness, which are often implicitly held by current scientific theories.
‘What-it-is-like’ to have an experience
The idea that consciousness comprises an inner subjective feeling poses perhaps the biggest challenge to any model of consciousness. Explanatory projects in the contemporary neuroscience of consciousness thus often seem to target the notion of ‘phenomenal consciousness’ that goes back to work by Thomas Nagel from the 1970s, emphasizing the ‘what-it-is-likeness’ of conscious experience. Nagel argued that the purely objective study of an entity, such as the one science provides, does not allow any inference about the subjective character of being such an entity. This has sometimes been misconstrued as the claim that a purely subjective phenomenon such as consciousness cannot be studied at all within the objective framework of science and should best be left alone, an idea which has been forcefully disputed by John Searle (Searle 1998; Searle 2000).
However, the position that phenomenal consciousness should be the end-all of any scientific theory of consciousness has been contested. Most prominent here is the illusionism proposed early on by Daniel Dennett (Dennett 1988; Dennett 1991), which got traction recently but still comes with its own issues (Frankish 2017; Chalmers 2018; Dennett 2018). A frequent criticism of such illusionist approaches is the charge of being anti-realist about consciousness. Yet, this charge can be finessed (Fallon 2020). What is typically denied by illusionists is the existence of some non-physical essence of conscious experience – the seemingly qualitative nature of experience – but not the reality of experience itself. This can, e.g., be illustrated with the ‘multiple drafts model’ of consciousness put forward by Dennett (Dennett 1991). If the illusionist claim is understood in this weaker way, the majority of neuroscientists would likely agree. However, it is then not clear how this position would contribute to a scientific understanding of consciousness, effectively arguing against a philosophical straw man.
Another related distinction is the one between the notions of ‘phenomenal’ and ‘access’ consciousness (Block 2005; Block 1995). The former refers explicitly to its what-it-is-likeness, sometimes assumed to correspond to a ‘minimal sense’ of conscious experience without necessarily requiring reportability (Metzinger 2020; Changeux 2006). In particular, some theories of consciousness claim to be about exactly this phenomenal aspect and thus carry a distinct explanatory pretence. The latter notion, ‘access’ consciousness, corresponds to centralized availability for processing of information and the reportability of a conscious experience. It also refers to phenomena that are closely related to consciousness in other aspects (e.g. attention or meta-cognition). It is an open debate in the scientific study of consciousness whether an explanation purely in terms of access consciousness is truly satisfactory or whether it is not in fact the ‘only’ scientifically rigorous approach (Naccache 2018; Fazekas and Overgaard 2018).
Much of the prominence that phenomenal consciousness received as a potential scientific topic can be traced back to the work of David Chalmers who introduced the notion of the ‘hard problem of consciousness’ (Chalmers 1995a), the difficulty to explain why certain forms of physical information processing should feel like anything at all or how physical and phenomenal facts are related to each other. Chalmers argued against the reducibility of consciousness and initially advocated a ‘natural dualism’ (Chalmers 1997). Natural dualists assume that reality is composed of two sets of (irreducible) properties. Chalmers, moreover, assumed that those refer to two properties of ‘information’, and that the functional (psychological) structure of information processing in the brain is mirrored in the phenomenal structure of consciousness. Chalmers later made the case for ‘panpsychism’ (Strawson 2006; Chalmers 2013a; Goff 2019), the view that consciousness is ubiquitous in nature and serves as the irreducible ‘intrinsic nature’ that grounds physical properties. Panpsychism is related to ‘dual-aspect monism’ that considers consciousness and physics merely as two aspects of a single underlying reality (Atmaspacher 2014). Some dual-aspect monists are panpsychists, although not necessarily so [other dual-aspect monists assume the single underlying reality to be neither of mental nor material but of ‘neutral’ nature (Stubenberg 2018)].
An alternative way to approach the problem of consciousness is the ‘biological naturalism’ of John Searle (Searle 2000). While Searle acknowledges the phenomenal character of consciousness, he finds it ‘obvious’ that it emerged from the brain similar to the way bile is produced by the liver. However, to date, no viable mechanism has been identified for this process. The problems for Searle’s approach are representative of all materialistic or ‘physicalist’ approaches to consciousness that are discussed later in the paper, at least where they pretend to shed light on phenomenal consciousness. The majority of approaches in the scientific study of consciousness endorses the position of physicalism.
Another monistic response is to invert the hard problem and argue that the physical world is a product of consciousness. Many idealists, roughly, state that matter exists only insofar as it is represented in consciousness – although idealism in fact comprises a set of many different (heterogeneous but related) views. The idealistic position was once the dominant world view in much of Western and Eastern culture, but received a massive blow in the 20th century. Most philosophers and scientists do not take this option seriously anymore, although the climate seems to change with more and more scholars advocating such a view (Marshall et al. 2001; Hoffman 2008; Hoffman and Prakash 2014; Kastrup 2017; Chalmers 2019).
Finally, one could try to bracket metaphysical issues in the study of consciousness and instead follow an approach advocated by Francisco Varela. Whereas Chalmers postulated that it seems as if there is a need for ‘extra ingredients’ to physical theory and whereas the various monisms express certain metaphysical assumptions, Varela suggested to regard consciousness and brain processes as ‘mutually constraining phenomena’ (Thompson 2007; Rodríguez 2008; Varela 1996) that ground an ‘empirical’ approach to consciousness. More generally, Varela’s ‘neurophenomenology’ is an adaption of an earlier continental approach [known as ‘phenomenology’ (Gallagher and Zahavi 2008; Kaufer and Chemero 2015)] to cognitive neuroscience that seeks to uncover the necessary structures of all experiences (including the ones that give rise to scientific knowledge). Neurophenomenology thereby suspends any metaphysical judgement if possible, thus realizing a methodological desideratum of earlier phenomenologists, who intended to go ‘back to the things themselves’, i.e. back to experience. Whether or not consciousness is in fact an emergent phenomenon, this leaves unanswered the question how explanations in terms of brain dynamics and conscious experience mutually constrain each other, e.g., to what extent consciousness is able to influence its physical substrate (Thompson and Varela 2001) or how any supposed ‘backreaction’ would manifest itself in scientific data.
All these proposals have their own problems, even though they come in very different guises. Illusionists need to explain why the illusion of consciousness appears as something real and vivid [the ‘illusion problem’ (Frankish 2017)] and argue that there is in fact no problem of consciousness over and above the problems of ‘access’ consciousness (problems which still need to be solved though). Alternatively, one would need to specify the reductive relation between ‘phenomenal’ consciousness and matter (e.g. solve the ‘hard problem’ and give a model of how consciousness actually emerges from the brain) or explain the causal efficacy of conscious experience.
From philosophy to scientific models of consciousness
A framework for the study of consciousness refers to a group of premises and assumptions to guide experiments and interpret results. More specifically, a model of consciousness conveys concrete hypotheses, predictions, mechanisms, and explanations of the associated phenomena. A proper theory for consciousness consists, by contrast, of a set of explicit (and often formalized) systematic premises plus a concrete model to enable the testing of (empirical or theoretical) predictions and eventually its implementation and manipulation. The approaches discussed in this paper correspond to models that mostly operate on implicit assumptions. Even though they do not resemble proper theories, in the sense just outlined, we will use both the terms ‘theory’ and ‘model’ interchangeably in the remainder of this article to better conform to the literature.
A first step to better understand models of consciousness is to make explicit their underlying philosophical assumptions. These assumptions inform and influence models of consciousness. In Fig. 1, we summarize the main relationships between philosophy and early and modern models of consciousness.
Philosophy and models of consciousness. Relationships and influences between philosophy and early and contemporary models of consciousness.
In this article we consider early models such as the mechanistic model of Crick and Koch (CK) (Crick and Koch 1998; Crick and Koch 2003), Dynamical Core (DC) (Edelman 2003), Multiple Drafts Model (MDM) (Dennett 1991), Orchestrated Objective Reduction (OrchOR) (Hameroff and Penrose 2014), Global workspace (GW) (Baars 1988; Baars 2005), Thalamo-Cortical loops and Sensorimotor Couplings (TCL) (Llinas et al. 1998; Llinás 2003), the dualist proposal by Beck and Eccles (BE) (Beck and Eccles 1992) and Enactive and Radical Embodiment (ERE) (Varela 1996; Thompson and Varela 2001; Varela et al. 2001; Lutz et al. 2002). Modern models are Global Neuronal Workspace (GNW) (Dehaene and Naccache 2001; Dehaene and Changeux 2011), Higher-Order Thought Theory (HOT) (Rosenthal 2002b; Rosenthal 2008), Recurrent Processing Theory (RP) (Lamme 2003; Lamme 2010; Fahrenfort et al. 2017), Predictive Processing and Interoception (PP&I) (Wiese and Friston 2021; Friston et al. 2020), Integrated Information Theory (IIT) (Oizumi et al. 2014; Tononi et al. 2016; Haun and Tononi 2019), Attention Schema Theory (AST) (Graziano and Kastner 2011; Graziano et al. 2020; Webb and Graziano 2015), Conscious Agent Networks (CAN) (Hoffman and Prakash 2014; Fields et al. 2018) and Temporo-spatial Theory of Consciousness (TTC) (Northoff 2013; Northoff and Huang 2017).
Early models typically inform later ones, sometimes via direct succession (e.g. GW/GNW), or via integration of concepts developed in previous models (e.g. TTC). Models may also initiate dialogues and remain under dynamic influences with each other. For example, both postulates of GNW and IIT do remain consistent to some extent with mechanisms put forward by the predictive coding approach (PP&I). Sometimes, the relation between different models or theories has more of an implicit nature. For example, IIT is sometimes thought to express some form of panpsychism, although this is done rather implicitly, inferred via analysing IIT’s basic premises.
Method of classification
Our initial examination of the current models of consciousness involves placing every theory within three orthogonal dimensions corresponding to the positions they espouse with regard to explanation. The first axis stands for the ‘mode of explanation’ assumed by the model (mechanistic vs. unificationist), the second for the ‘mechanism of explanation’ (functional vs. causal) and the third for the ‘target of explanation’ (quantity vs. quality of consciousness). This classification results in ‘explanatory profiles’ for each theory, which accounts for a substantial amount of variance in the theoretical landscape.
Importantly, the mode of explanation is not to be conflated with a mechanism of explanation, i.e. while the mode informs about the ultimate ‘aim’ of the explanation (what constitutes a satisfactory answer to the ‘why’ question), the mechanism demonstrates which particular ‘tool’ can get us towards that predefined aim (i.e. the ‘how’ question).
Mode of explanation
One of the most important hallmarks of a successful theory is its explanatory power. Nonetheless, the very notion of explanation (what it means to successfully ‘explain’ something), despite its deceptive simplicity, can be unpacked as highly heterogeneous (Nagel 1961; Salmon 1990; Woodward 2019; Woodward 2013; Woodward 2004; Strevens 2004; Colombo 2017; de Regt and Baumberger 2019). In particular, the very notion of ‘explaining consciousness’, although ubiquitous in the literature, constitutes a deceptive ‘umbrella term’, under which one can identify strikingly different theoretical goals. That ‘consciousness’ is frequently used as an umbrella term for various mental phenomena has been often stated in the literature [e.g. (Van Gulick 2018)]. Here we wish to emphasize that the same is true for the seemingly innocent notion of its ‘explanation’. The field of consciousness science has been flooded with numerous works contemplating whether the full explanation of this phenomenon is even feasible (McGinn 1988; Chalmers 1995b; Hohwy and Frith 2004). Therefore, we focus on two alternatives that seem particularly relevant in the domain of consciousness studies: ‘mechanistic approaches’ and ‘unification’ (Salmon 1990; Strevens 2004).
A mechanistic approach posits that a particular occurrence is explained once we demonstrate how it ‘fits into’ the presumed spatio-temporal causal structure of the world, i.e. to explain an event is to identify its cause (Reutlinger 2017; Salmon 1990). This is typically achieved via referring to the transmission of a counterfactual and local modification (a ‘mark’) of the associated physical, typically spatio-temporally continuous, process (Woodward 2004). Mechanistic accounts typically subscribe to a realist notion of causation that grounds the explanation, prediction and other inferences (Nathan 2020). A theory of consciousness that holds the mechanistic view would accordingly imply that the satisfactory explanation of subjective experience requires meticulous unfolding of the chain of causes and effects occurring in the nervous system that leads to a conscious experience of some kind. Such a position imposes a strictly empirical agenda, investigating which neurophysiological events precede and ‘give rise’ to a conscious experience (Craver 2007). Importantly, the driving force behind such accounts is a philosophical ‘assumption’ that such a chain of causes is ‘independently given’ at some objective level of description at some spatio-temporal scale (e.g. of the brain). Naturally, these frameworks tend to cluster together under branches of ontological or methodological reductionism.
On the other hand, the unificationist stance seeks to provide a ‘unified account’ of a range of different phenomena or laws (Glymour 1980; Kitcher 1981), given causal or non-causal explanations (Reutlinger 2017; Salmon 1990; Batterman and Rice 2014), which were previously thought to be unrelated – or related in a mysterious or seemingly arbitrary way (core historical examples are Maxwell’s unification of electricity and magnetism or Newton’s unification of terrestrial and celestial motion). Explaining consciousness under the unificationist framework would give priority to demonstrating how the phenomenon of consciousness is embedded into a parsimonious, coherent framework. As a result, the unificationist seeking a satisfactory explanation would be inclined to associate less with empiricism, and more with formal tools, mathematics and non-reductive philosophy. On the other hand, unificationist explanations have been argued to be most relevant for physics but at odds with biology and neuroscience (Anderson and Chemero 2013; Bayne 2018).
While mechanistic accounts, implicitly or explicitly, always assume the realist notion of causation in space and time, unificationist accounts postulate that explaining consciousness cannot be exhausted by studying the spatio-temporal chain of causes and effects localized in the brain. In particular, assumptions about the causal order and spatio-temporal descriptions are often being put on hold and treated as an ‘explanandum’ (a phenomenon to be explained) rather than as ‘explanans’ (the grounding of explanation) (Barnes 1992). Causes, at this point, can be regarded as explanatory postulates or theoretical hypotheses which do not exist independently of the explanations that describe them (Nathan 2020; Barnes 1992). Specifically for the science of ‘consciousness’, it should also be made clear how or why consciousness is related to the explanation of such causal chains (e.g., by arguing that causal orders exist only ‘within’ consciousness).
Numerous examples show that both types of explanation have contributed to significant progress in science (Barnes 1992; Woodward 2004). Our aim is not to assess which approach produces better explanations, but to simply recognize that different frameworks of consciousness will be inclined to differ already on this very basic assumption.
Mechanism of explanation
We define the term ‘mechanism’ after Illari and Williamson as ‘entities and activities organized in such a way that they are “responsible” for the phenomenon’ (Illari and Williamson 2012). This way of defining the term is independent from any specific theory of causation. Numerous examples exist where explanations can be purely mathematical, the role of causation being denied or simply ignored (Batterman and Rice 2014; Reutlinger 2017).
In the context of consciousness science, an increasingly popular division, introduced by Doerig et al. (Doerig et al. 2019) and specifically addressing different kinds of mechanisms, distinguishes explanations as either ‘functional’ or ‘causal’. The stance of functionalism primarily states that consciousness can be generated as long as a particular function is realized (Block 1996), without any specific constraint on the exact causal machinery behind it. In principle, any system may become conscious as long as it executes the functions associated with conscious experience. On the opposite side, causal theories do not necessarily unsubscribe from the view that a particular function might be typically associated with consciousness, but the burden of explanation is placed on ‘how’ such a function is implemented: establishing what and how elements of a system interact. These theories seek relevant causal relationships within the model’s target system(s) (Batterman and Rice 2014). In those terms, causal models support the idea that only a system with the right causal relationships will lead to conscious experience.
One could further illustrate this distinction by introducing the idea of a structure-preserving map (M) between two objects. Causal models insist that the causal system’s structure (S) explains consciousness, |$M: S\rightarrow C$|, where the arrow refers to a mapping that preserves system structure (whether or not a particular and objective spatio-temporal structure is the most relevant for analysis). Functional models, however, would argue that the functional structure of a system (F) explains consciousness, |$M: F\rightarrow C$|, and thus it is the function that is preserved by the arrow. These two types of models appear exclusive, in the sense that S and F are different objects. However, if we focus on the nature of the map, independently of their objects, the difference becomes a question of ‘degree’. Causal theories usually assume there is only one way to preserve the causal structure and the phenomenology of subjective reports. This is tantamount to postulating an isomorphism between domains (there exists only one such arrow) (Tsuchiya et al. 2016), whereas functional theories would claim/assume that there are multiple ways (arrows) to preserve the function and therefore giving rise to consciousness, i.e. |$M^{^{\prime}}, M^{^{\prime\prime}}, M^{^{\prime\prime\prime}}$|, etc.
Superficially, mechanistic accounts and causal models seem to overlap, as both axes refer to the vocabulary of causation. To avoid any potential confusion or apparent contradiction, it is critical to recall our distinction between the ‘aim’ and ‘tool’ of explanation (which could be referred to as ‘metaphysical’ and ‘pragmatical’ commitments of the theory, respectively). While mechanistic explanations subscribe to the realist notion of causation, causal models do not necessarily do so and could merely ‘use’ causes (rather than functions) to unfold the relevant interactions within the target system. In principle, they could still maintain that causes are not ‘the only’ explanandum, in the sense that explaining them still would not suffice to explain consciousness. Thus, a model can be non-mechanistic and causal at the same time without contradiction.
Target of explanation
Lastly, the target of an explanation corresponds to the aspects of a particular phenomenon that scientists intend to explain. The most basic distinction that emerged over the last decades is the one between ‘quality’ and ‘quantity’. The quality of consciousness is what makes consciousness feel ‘the way’ it does (cf. also ‘What-it-is-like’ to have an experience Section), and the quantity corresponds to what makes the system conscious rather than unconscious. A model targeting quality should therefore account for why any stimulus should feel a ‘particular way’ and what makes an experience spatial, visual, auditory, painful or temporal, while a model targeting quantity intends to account for global markers differentiating conscious vs. unconscious systems.
A full-fledged theory of consciousness needs to explain both the quantity and quality aspects of subjective experience. Several theories tend to focus exclusively on global markers differentiating conscious vs. unconscious systems, and the problem of quality in such cases is delegated to the external world (i.e. sensory cues feel the way they do solely because they ‘carry’ their quality from the environment, or stimulate ‘correct’, labelled receptors). However, there are many, conceptual discussions, each pointing to a specific problem of such a ‘delegation’; see arguments on brains in vats (Horgan et al. 2004), inverted spectrum (Shoemaker 2000), actual cases of perceptual variation (Block 1999), the fact that sensory characters correlate much more with neural patterns than anything else in the external world (Pautz 2014), dramatic perceptual alterations in psychedelic experiences (Bayne and Carter 2018), perceptual illusions, the generic non-preservation of phenomenal and environmental structure (Prakash et al. 2020), etc. A much more challenging project would strive to explain why an experience feels the way it does, e.g. based on the internal architecture and dynamics of the brain, without referral to the external world as something that ‘stores’ or ‘produces’ any qualities [but see (Graziano et al. 2020)].
Classifying models of consciousness
According to the analytical dimensions defined in the previous section, we now classify and discuss selected models of consciousness in a three-dimensional map (Fig. 2). This classification intends to illustrate our framework. It is a provisional attempt based on quotations and a rough score, which provide support to our discussion. We have first focused on reviewing the relevant articles that describe the selected theories. Consequently, we have scoured representative works for statements pinpointing what a satisfactory explanation would look like according to each theory. We have then determined numerical values along the dimensions specified in the previous section (see sections Supplementary: Relevant Citations, and Supplementary: Scoring Method).
Models Classification. The selected models inhabit different explanatory spaces according to our three analytical dimensions. After scanning the literature, summarizing and discussing main tendencies in light of three-axis classification in Table 1, each author independently scored models between 1 and 3: 1 corresponds to a slightly commitment to the direction of the axis and 3 a strong commitment. 0 value was used any time that the model does not qualify for the + or − direction. This quantification is provisional; it only intents to illustrate our framework and it is not based on statistical data (see details in section Supplementary data: Scoring Method).
Axes for a three-dimensional classification map with two opposite directions per dimension (represented by positive or negative values). These axes were later used to score models of consciousness, see details in section Supplementary data: Scoring Method.
| Axes . | First direction (+) . | Second direction (−) . |
|---|---|---|
| Mode of explanation | Mechanistic | Unificationist |
| Mechanism of explanation | Causal | Functional |
| Target of explanation | Quality | Quantity |
| Axes . | First direction (+) . | Second direction (−) . |
|---|---|---|
| Mode of explanation | Mechanistic | Unificationist |
| Mechanism of explanation | Causal | Functional |
| Target of explanation | Quality | Quantity |
Axes for a three-dimensional classification map with two opposite directions per dimension (represented by positive or negative values). These axes were later used to score models of consciousness, see details in section Supplementary data: Scoring Method.
| Axes . | First direction (+) . | Second direction (−) . |
|---|---|---|
| Mode of explanation | Mechanistic | Unificationist |
| Mechanism of explanation | Causal | Functional |
| Target of explanation | Quality | Quantity |
| Axes . | First direction (+) . | Second direction (−) . |
|---|---|---|
| Mode of explanation | Mechanistic | Unificationist |
| Mechanism of explanation | Causal | Functional |
| Target of explanation | Quality | Quantity |
Mode of explanation
Based on our classification, we unpack the explanatory profile of each theory. We have found individual discrepancies, with most of the theories leaning clearly towards one or another mode. Some frameworks, including GNW and HOT, gave us consistent clues to classify them under the mechanistic cluster. A statement that would drive us to the mechanistic classification could be, e.g., the following statement by Rosenthal, a proponent of HOT: ‘we understand something only when we can explain it, and explaining a natural phenomenon typically if not always means locating it in its distinctive causal nexus’ (Rosenthal 2008), or by Dehaene, a proponent of GNW: ‘tools of cognitive psychology and neuroscience may suffice to analyze consciousness’ (Dehaene and Naccache 2001). On the other hand, unificationist propensities are more apparent in frameworks such as IIT: ‘IIT provides a “principled explanation for several seemingly disparate facts” about the PSC [Physical Substrate of Consciousness]’; or PP: ‘... it could unify existing approaches under a single overarching principle (i.e., the FEP)’ (Wiese and Friston 2021). Such unifying tendencies are also expressed by proponents of AST, who aim for a ‘standard model of consciousness’ (Graziano et al. 2020) by taking a deflationary stance on the reality of consciousness or, on the other extreme, by proponents of CAN: ‘if we want to go beyond this “applied science” and understand the true nature of the mind and the reality beyond it, we can’t look to neurons or brains’ (Fields et al. 2018). The full list of relevant quotes, justifying each model’s classification, can be found in Supplement Table S2.
This first dimension clearly constitutes a parsimonious dividing force, introducing tensions between the models’ individual goals. Arguably, identifying which mode of explanation a theory supports can also predict how a certain theory would pragmatically proceed in its investigations. Although both groups of theories would not deny the relevance of empirical research, the mechanistic models would take on a cautious ‘bottom-up’ approach, withholding the drawing of any firm conclusions without substantial accumulation of incoming data (Michel et al. 2019). Unification accounts, on the other hand, would be much more prone towards a ‘top-down’ approach, giving priority to those empirical predictions which can make sense of seemingly disparate phenomena from the perspective of the framework itself.
Mechanism of explanation
There is a marked difference in which explanations can provide explanatory power, by either concentrating on overall functional roles or by describing network parts and the interactions between them. Pertinent examples for functional theories are HOT and GNW, according to which, consciousness can arise in a physical system as long as it realizes ‘meta-representation’ or ‘global broadcasting’, respectively. On the other hand, models such as IIT and RPT lean towards elucidating the structure of causal interactions at the level of network analysis. For a full list of relevant quotes disclosing either functional or causal inclinations in explaining consciousness, see Supplement Table S3.
Although we believe that most theories do explicitly or implicitly differ in these explanatory assumptions, Fahrenfort and van Gaal point out that ‘most’ empirical theories would eventually aim at explanations involving causal implementations, rather than functions (Fahrenfort and van Gaal 2021). In line with that view, we have indeed come across many examples in which models of consciousness, even those labelled as functional, in actuality resort frequently to the language of ‘causal interactions’. For example, as stated by Rosenthal: ‘On the HOT hypothesis, a conscious state is a compound state, consisting of the state one is conscious of together with a HOT. So the causal role a conscious state plays is actually the interaction of two causal roles: that played by the state itself and that played by the HOT’ (Rosenthal 2008). Another relevant instance could be predictive processing theory (PP). Despite having been classified as functional in (Doerig et al. 2019), one variant of PP clearly states that implementing the adequate computational/functional principles is only a necessary, but not sufficient condition for consciousness. As explicitly mentioned by Wiese and Friston (Wiese and Friston 2021), computations need to be physically instantiated in the right architecture and not all virtual machines that realize approximate Bayesian inference should be considered conscious. The reason they give is that a Markov blanket of the physical system must be based on the system’s dynamics, and the dynamics strictly depends on the system’s structure. That drives us to conclude that a popular framework has recently moved significantly away from functionalism and can now be classified as a causal theory, although different interpretations might exist [e.g. (Clark 2016)].
Additionally, some models seem difficult to classify as either functional or causal in the above sense. For example, TTC aims to accommodate both functional and causal types of explanations, making it difficult to identify its commitments (although according to some key quotes it seems to lean towards causal approaches). Other models such as ERE, CAN and BE are neither functional nor causal in the sense of Doerig et al. (Doerig et al. 2019). This is because they do not share the underlying philosophical assumptions typically made by neuroscientific theories of consciousness (cf. ‘What-it-is-like’ to have an experience Section). ERE assumes a form of dynamical co-emergence, i.e. consciousness co-arises with the system in such a way that there is a contextual constraint between the biological system (living body) and the experience (lived body), making both interdependent (Thompson 2007). In other words, according to ERE consciousness does not only emerge from the brain and the body, but it also actively constrains them, creating a causally reciprocal relationship between neural events and experience (Thompson and Varela 2001). Lastly, other non-trivial cases include CAN (consciousness in this case is fundamental and the physical realm is what emerges from it) and BE (which is premised on dualist).
Target of explanation
The final dimension corresponds to the target of explanation, i.e. quality vs. quantity of consciousness. We analysed relevant papers looking for statements revealing ‘what’ the theory aims to explain. Theories that focus mainly on the contrast between wakefulness and other impaired conditions (e.g. TCL) or on what neural activity might underlie the transition from the stimulus being ‘unseen’ to ‘seen’ (e.g. GNW, HOT and CK) can be classified as theories of quantity. We encourage the reader to notice that the ‘why’ question posed in case of such theories is not ‘why a stimulus feels this way’, but rather ‘what constitutes the switch rendering content visible’. The stimulus here can be replaced with any other stimulus, as the quality (although not necessarily the structure) of content remains irrelevant and out of the explanatory target. Importantly, answering the first question leads us to ‘phenomenal content’, while the second question corresponds to ‘access content’.
An attempt to explain the quality rather than sole quantity is exemplified by IIT. IIT tries to identify the mechanisms behind the phenomenal character of experience, i.e. the quality of its content (‘what makes an experience visual, auditory, colourful, painful?’). In that vein, it poses that the quality of consciousness is in one-to-one correspondence with the geometry, concepts and relations encapsulated by the Maximally Irreducible Conceptual Structures (Haun and Tononi 2019). IIT predicts that the contents of consciousness are entirely specified by the internal workings of elementary mechanisms of the main complex. Notably though, IIT’s agenda to target quality is still in its nascent stages and has not gone beyond trying to explain the spatiality of experience yet (Haun and Tononi 2019).
Other models, to a greater or lesser extent, also strive to address quality. One example is ERE. In this theory, first-person reports inform the interpretation of neural signatures, thus co-determining the actual explanandum of the theory (Lutz et al. 2002; Lutz and Thompson 2003; Lutz et al. 2015; Windt et al. 2016; Petitmengin et al. 2019; Poletti et al. 2021). In one particular example, experienced meditators were able to wilfully influence neural activity (both short and long term) (Lutz et al. 2004). This reveals a more complex relationship between experiential content and its neural substrate: quantitative measures constrain experienced qualities, but quality might in turn constrain quantity (Varela 1996). Another example, RPT, largely focuses on questions related to phenomenal awareness. There are appreciable ways in which it can explain some qualitative aspects of consciousness, such as the figure-ground segmentation, i.e. how one might sometimes perceive a texture to be a surface that lies on top of a background (Roelfsema et al. 2002; Scholte et al. 2008; Fahrenfort et al. 2017). Functional models also claim to target quality, however in a different way, typically using recognition and detection paradigms. For example, HOT, although it tends to delegate ‘qualities’ to largely unspecified mechanisms of low-level representation or the external world, puts forward explanations why we might sometimes ‘feel confident’ of the perceptual experience (Brown et al. 2019)).
It is, therefore, evident that different theories adopt different strategies in addressing the problem of quality, depending on their underlying philosophical assumptions. Arguably, these discussions might pave the way for other theories’ proponents to appreciate different kinds of explanatory aims and recognize that a mature theory of consciousness should address both the quantity and quality of experience. Novel questions inspired by the focus on quality for theories of quantity could be, inter alia: How exactly does the ‘pattern’ of global ignition inform phenomenal content? What distinguishes visual or auditory re-entrant processing, or in other words, what mechanism makes these two modalities ‘feel’ differently? Notice that circumventing the problem by referring to the function of the primary sensory cortices (e.g. pre-labelled as visual or auditory) is not tenable, because it does not inform ‘what precisely makes’ a primary cortex visual or auditory in the first place. Alternatively, for models such as AST, what, on the level of modelling or re-representing, ‘specifically’ correlates with the (illusory) attribution of qualia?
Empirical consequences
The fact that different theories subscribe to different modes, mechanisms and targets of explanation has several empirical consequences. One of them pertains to the localization of neural correlates of consciousness (NCCs) (Boly et al. 2017; Klink et al. 2015). First off, there is a substantial disagreement between the so-called first order (e.g. IIT and RPT) and higher-order theories (e.g. HOT) insofar that the former typically assign NCCs to the early sensory cortices while the latter to the frontal-parietal network. It is worth pointing out that the ‘frontal theories’ tend to systematically cluster under the mechanistic mode of explanation, often interested in questions of quantity rather than quality (i.e. under what conditions a particular stimulus would be classified as consciously perceived or not) (see Fig. 2). Notwithstanding these distinctions, most of the models reviewed above are dynamical global network approaches and therefore non-localist in principle (Dehaene and Naccache 2001). The network is relevant; the nodes alone are not. For example, GNW is often misunderstood as a fixed architecture encompassing the fronto-parietal cortices. In actuality, it comprises dynamic neural contributions that define this hypothetical global network (Dehaene and Changeux 2005). The only anatomical constraint is that relevant regions should be connected by long axons of pyramidal neurons. During decades of research, GNW identified pyramidal neurons in Layers II and III as candidate mechanisms. In light of recent relevant evidence (Suzuki and Larkum 2020), they have put forward Layer V as a more likely GW substrate (Mashour et al. 2020). On the contrary, IIT postulates that the true NCC lies in the posterior hot zone; one of the main points of contention with GNW being the counterintuitive role of inactive units, contributing to the cause–effect information just as the active ones do (Oizumi et al. 2014; Siclari et al. 2017). HOT seems not to specify network mechanisms in sufficiently rigorous details. Even though the original framework has been worked out empirically (Lau and Passingham 2006; Brown et al. 2019), this family of models leave a lot of space to accommodate ambiguous experimental data.
Other inconclusive findings include the study of posterior hot zone involvement in dreams (Siclari et al. 2017; Mashour et al. 2020) and Perturbational Complexity Index (PCI) (Casarotto et al. 2016). In the first case the reduction of low-frequency activity in posterior zones of the brain correlates with dreams during rapid-eye-movement (REM) sleep and non-REM sleep (Siclari et al. 2017). At the same time, content-specific dreams involve high-frequency activity in the frontal and prefrontal cortices. Some authors interpret these results as evidence of posterior zones for phenomenal consciousness and support of IIT, while others view them as a clear evidence of the role of prefrontal cortex and GNW (Mashour et al. 2020). In the second case, the PCI index inspired by the IIT framework seems compatible with the global ignitory activity of GNW (Mashour et al. 2020), as well as with other models such as ERE. Another recent study inspired by GNW found that the thalamic nucleus of monkey brains under deep brain stimulation restores signatures of consciousness and reactivate nodes of the GNW that remain inactive under anaesthesia (Tasserie 2020). However, as the authors also pointed out, these results are also compatible with thalamo-cortical loops theories and IIT.
Therefore, the postulation of either prefrontal regions and posterior regions to be the primary locus of the NCC does not by itself speak to different explanatory approaches and probably merely reflects the availability of experimental techniques and methodological choices (Yaron et al. 2021). Moreover, the question of localization makes sense only if one endorses a mechanistic and causal explanatory framework simultaneously, which is only the case for CK and OrchOR models. If one, by contrast, looks for unification or function, then the fact that, say, frontal or posterior regions are most relevant for the NCC, is contingent. This is particularly true for those views which do not adhere to the explanatory primacy of causal chains within space-time (e.g. CAN). As such, experiments that try to identify the NCC do not provide evidence for or against such a theory. But even bracketing questions of explanatory mode, localization approaches are problematic: evidence for some mechanisms at any particular level does not falsify the relevance of other mechanisms at other scales. An interesting example is a controversial recent comparison between IIT and GNW at the single-unit level. At first glance, results suggest that GNW is supported by the evidence, while IIT is not (Noel et al. 2019). However, considering the active single-unit level as the optimal spatio-temporal scale for testing IIT remains problematic. This would force an assumption of GNW (i.e. that consciousness is correlated to only active neurons) onto IIT, which is prima facie not warranted. But this is a conceptual issue having to do with explanatory commitments – and not an empirical one.
Another empirical consequence of the way a theory emphasizes either quality or quantity has implications for the methodology used. In general, it has been long recognized that the scientific study of consciousness utilizes first-person and third-person approaches (Olivares et al. 2015). The former includes subjective reports and phenomenological interviews (Chalmers 2013b) and the latter refers to objective measures of physical states, using different techniques such as electroencephalography, functional magnetic resonance imaging and magnetoencephalography, among others. Most models of consciousness claim to employ both subjective and objective accounts. However, their assigned importance varies across the models. In particular, those theories which seek to explain quality will wish to utilize a method that is specifically suited to make the qualitative aspect of experience precise. Particularly promising accounts are called ‘second-person’ methods, referring to interview techniques that incorporate verbal and non-verbal reports in order to obtain a well-informed subjective report (Olivares et al. 2015; Lutz et al. 2015; Petitmengin et al. 2019). This approach is motivated by earlier research in neurophenomenology (Thompson 2004; Varela 1996; Lutz et al. 2002). The second-person method is different from the first-person method in that the former is guided by an interviewer who reads and interprets various indicators from the first-person subjective report. Given these indicators, the interviewer is able to ask more refined questions that force a subject to closely specify her reports.
Related to the quality versus quantity distinction is the one between levels and contents of consciousness (Bachmann and Hudetz 2014; Bayne et al. 2016; Storm et al. 2017). Levels of consciousness convey global signatures of consciousness, from which different paradigms contrast awake neural activity against non-awake or disrupted conditions such as sleep, chronic disorders of consciousness and anaesthesia, among others (Signorelli et al. 2021a; Signorelli 2021). By contrast, paradigms looking for contents of consciousness survey conscious experiences through contrasting perceptual analysis (perceived vs. unperceived) and multiple psychophysical reporting paradigms. Examples include masked stimuli, high-contrast figures, binocular rivalry, flash suppression, motion-induced blindness and attentional paradigms, among others (Klink et al. 2015). However, one must not conflate the study of contents with the study of the specific phenomenology of such contents. Yet, approaches that study contents quantitatively might still be understood as studying the structure ‘between’ instances of (qualitative) consciousness. For example, it has recently been argued that postdictive studies in the auditory domain could uncover the temporal structure of perception (Herzog et al. 2020) – something that has previously been discussed extensively in the phenomenological literature. Thus, in general, it is more appropriate to distinguish between an ‘access content’ and a ‘phenomenological content’, the former allowing an indirect (often dynamical) inference of the structure of consciousness and the latter specifically targeting the ‘what-it-is-likeness’ of experience.
Conclusions: integrative methods for the future
In light of our previous sections, the empirical testing of models of consciousness is far from trivial. Comparing models of consciousness is not only difficult due to experimental limitations, but also due to the fact that different models operate on very different and mostly implicit assumptions about modes, mechanisms and targets of explanation. We thus introduced a classification scheme to make the different explanatory profiles of leading models explicit. To our knowledge, this is the first time that all these models and perspectives are reviewed systematically, focusing on different explanatory aspects and analytical dimensions, and organized in one single and comprehensible classification.
Our provisional classification framework serves as an invitation for theorists to weigh in on how their own models might be classified. In future attempts, a representative number of experts, performing statistics, principal component analyses and test–retest reliability should be added, among others. One might also speculate that new models will occupy empty spaces in our taxonomy and current models of consciousness will change their location as new versions of them emerge, either in light of new evidence or more refined theoretical discussions.
Having perused the relevant literature around 14 popular frameworks, we suggest that a number of disputes in the field of consciousness studies might stem from differently set explanatory goals and targets. Although both mechanistic and unificationist accounts have their advantages, there are certain areas of research where one of them might turn out to be more suitable than the other. Some types of explanations might require to step back from investigating the specific empirical details and look more at the overall mathematical structure: the reason ‘why’ a person cannot untie a particular knot may stem from a topological fact about the knot, rather than from a detailed causal trajectory illustrating the attempt of its disentanglement. Several examples show that a full-fledged explanation of a phenomenon might sometimes require more than a causal story (Reutlinger 2017). On the other hand, unificationist accounts face their own problems, such as the ‘problem of asymmetry’: if A explains B, B does not explain A. It is heavily contested whether unificationist approaches could accommodate this intuition (Barnes 1992).
It is yet to be established which type of explanation is most adequate for the science of consciousness or whether even different modes of explanation would be required [e.g. manipulationist accounts (Woodward 2004; Woodward 2019)]. Nonetheless, one of the messages of this paper is to emphasize that the first step towards a more mature science of consciousness is the recognition that the question of why brain activity is correlated to subjective experience can be understood in a variety of ways. We suggest that to fully explain ‘what it means to be conscious’, one needs to first be precise about what it would mean to ‘explain’ something.
Remaining aware of the disagreements within the field (and its early history), one might also try to extract what most models agree on. Some empirical approaches intend to follow this kind of pragmatism. For example, the concept of criticality in dynamical system theory shows to be compatible with both access and phenomenal consciousness signatures (Tagliazucchi 2017). The analytical method of connectome harmonics also aims to unify different signatures of consciousness, from a more general perspective of brain functioning and physical system theory (Atasoy et al. 2017; Atasoy et al. 2019; Luppi et al. 2020). Large-scale models using different anatomical, functional and molecular layers of description also present promising features to integrate different mechanisms at different scales (Kringelbach et al. 2020), as well as various signatures of consciousness (Signorelli et al. 2021a; Signorelli 2021). Recently, optogenetic experiments demonstrated that the biophysics of pyramidal neurons in cortical layer V integrates two contentious mechanisms associated with consciousness, i.e. cortico-cortical loops and higher-order thalamo-cortical loops (Suzuki and Larkum 2020; Aru et al. 2020). More research aiming at synthesizing different findings is currently underway.
Recent mathematical works too have recognized the need for integration within a sound theoretical foundation. This new trend, ‘mathematical consciousness science’ (AMCS 2021), employs formal and rigorous methods to explore ways to distinguish various models and derive new empirical predictions. Some examples are the mathematical developments, based on IIT (Oizumi et al. 2016; Tsuchiya et al. 2016; Kleiner and Tull 2020) and mathematized phenomenology (Yoshimi 2007; Ehresmann and Gomez-Ramirez 2015; Prentner 2019; Signorelli et al. 2021b); other approaches are based on symmetry (Kleiner 2020), category theory (Northoff et al. 2019; Tsuchiya and Saigo 2020) or the compositionality of processes (Signorelli and Meling 2021; Signorelli et al. 2021c; Signorelli and Boils 2021; Signorelli et al. 2021b). Some models explicitly address different metaphysical starting points, such as idealism (Hoffman and Prakash 2014) or decompositional approaches of dual-aspect monism (Atmanspacher 2020). Common to all these approaches is that, inspired by the transparency of mathematics, they explicitly define their core assumptions. In the end, whether they are of any value will be determined by how much explanatory power they bring into the constantly accruing experimental evidence.
The science of consciousness needs integrative frameworks, and integrative frameworks by definition are multidisciplinary. In the future, and going beyond empirical methods and mathematics, a dialogue with artists, meditators and proponents of the humanities may help us to think outside the box and rediscover some aspects of conscious experience that have largely been unattended to.
Supplementary data
Supplementary data is available at NCONSC Journal online.
Acknowledgement
The authors thank Athena Demertzi, Johannes Fahrenfort, Chris Fields, Don Hoffman, Bechir Jarraya, Lucia Melloni, Anil Seth, Jacobo Sitt, Wanja Wiese and two anonymous referees for helpful comments and constructive feedback.
Data availability
There is no experimental data involved in this research.
Funding
This work received financial support by Comisión Nacional de Investigación Ciencia y Tecnología (CONICYT, currently ANID) through Programa Formacion de Capital Avanzado (PFCHA), Doctoral scholarship Becas Chile: CONICYT PFCHA/DOCTORADO BECAS CHILE/2016 - 72170507, Polish Ministry of Science and Higher Education’s Diamond Grant DI2016 020046, and the project Comparative investigation of the cortical circuits in mouse, NHP and human (CORTICITY).
Conflict of interest statement
None declared.
References
Author notes
Camilo Miguel Signorelli, Joanna Szczotka and Robert Prentner contributed equally to this work

