Hi Dan! This is a really nice paper! I very much enjoyed reading it, and I learnt a lot!
Here are three random comments on a terrific paper. Please feel free to ignore them if they are not useful.
Hi Dan,
Thanks very much for this!
I’m not sure I see the sense in which H-cognition is a heuristic
I just meant that talk of H-cognition is useful in helping us to get into the ballpark of making theoretically interesting claims (like talk of ‘innateness’). Talking in this way serves a useful heuristic purpose in guiding research. But once one starts doing serious work in the area, this informal talk should should be dropped in favour of the more precise notions that you develop.
I somewhat cagily said that “the hierarchy of higher cognitive faculties generally” can be seen as constituting psychological kinds. On one reading, that would be satisfied in the event that each type of H-cognition plays an interesting taxonomic and explanatory role.
Thanks—that is really helpful. I didn’t pick that up that subtlety on my first reading of the paper.
I don’t think H-cog should simply be defined in terms of nonmodularity. Defining it in terms of informational integration avoids this.
I can see your point, but it still seems a fine line to tread. On many understandings of modularity, modules require information encapsulation, and that at least on the surface, seems to preclude informational integration.
Is your way out of this to say that although a single module would not satisfy information integration, there is nothing to stop a collection of interacting modules from satisfying it? Therefore, a MM architecture implementing H-cognition is not being ruled out by fiat. Is that the right way to think about it?
But if it is, then the old modularity proposal in Sect 2 still seems to have some life left:
A related way of drawing the higher/lower distinction says that lower cognition takes place within modular systems, while higher cognition is non-modular.
Suppose we understand ‘non-modular’ as taking place outside a single model—say, involving the interaction between multiple modules. If so, this proposal is no longer rules out a MM architecture implementing H-cognition by fiat. L-cognition would take place inside modules, and H-cognition would be implemented in the interactions between multiple modules (as Carruthers describes). So your original objection to the modularity proposal—that it precludes an MM architecture from implementing H-cognition—no longer goes through?
Thanks very much for your response to (3) and comments to Bryce, really helpful!
Hi Dan
I really enjoyed the paper and I like your three dimensions for concepts a lot. I think they have the potential to be deployed in many useful ways. Clearly one can envisage how they can determine whether faculty A is higher than faculty B. But I guess the points made by the commenters above got me thinking about the practicalities of setting a bar for higher cognition simpliciter as Bryce puts it and just what would motivate that setting.
Clearly faculties using human style concepts would be HC as these score nearly maximum on your dimensions. And faculties in a subsumption architecture might make use of concepts/representations scoring nearly zero so these could be ruled out as engaging in higher cognition (although getting fans of such architectures to agree might be trickier!).
But how about in between? What might be rough guidelines for setting the bar? For example, would a score of, say, over 50% on all three dimensions be a useful first pass? Would a score of zero (or below a set minimum) on any one dimension rule it out for HC? Would the three dimensions carry equal weight or differ in importance for HC, as sometimes I felt you were suggesting?
The worry then becomes what would make even rough answers to these questions principled rather than arbitrary? Anyone whose favoured architecture got ruled out could simply claim that there was no motivation for setting the bar in that way. For example a purely (as opposed to a partially) connectionist architecture might not exceed the threshold because it scored so badly on the combinability dimension but its advocates would say that it may turn out that we discover in the future that combination can be achieved without the vehicles being combined.
On the other hand without some detail on these kinds of matters it seems we would only be sure about our taxonomizations in cases that were close to maximum/minimum scores on all three dimensions.
So I guess my question is whether you think there is a way to set the bar in a principled manner or, if not, why you don’t think there needs to be?Hi Dan,
Thanks for sharing your excellent paper, which has tremendously helped my own thinking on a bunch of closely related issues. In particular, I think that your taxonomic scheme can be fruitfully employed to think about current debates over 4E-cognition, in addition to the applications you mention in Section 4. So let me offer a few musings on your taxonomy from a 4E-perspective.
First, we can obviously use your scale to derive a much more fine-grained taxonomy of “ideal types” akin to Dennett’s (1995) “Tower of Generate and Test,” defined purely in terms of meta-architectural constraints (rather than any specific cognitive architectures that are used to meet those constraints). Let’s say that Popperian creatures score relatively low on all 3 dimensions, Fregean creatures score fairly high on all 3 dimensions, and Gregorian creatures score extremely high on all 3. Once could view the resulting hierarchy of mental complexity as a trend towards increasingly greater amounts of “symbol un-grounding.” From that perspective, the main question is to identify the main drivers in un-grounding and decoupling cognition from its embodied/embedded roots, including lower, non-conceptual forms of mental representations. More specifically, what kinds of tweaks to the biologically basic architecture of our brains does it take to reach the upper echelon of H-cognition?
Clearly, the cogitations of a socially, culturally, and technologically scaffolded mind can score very high on the W-scale. [Note: we can stay neutral with respect to HEC vs. HEMC interpretations of scaffolding – Sterelny’s modest notion of “scaffolded minds” will do just fine.] For example, we can rely on sophisticated mathematical symbolisms to represent Hilbert spaces. We can use computer simulations of your tinkertoy model that we can start and pause, rewind, and re-run at will, with little cognitive effort, to engage in counterfactual reasoning about chemical compounds. And we use fully and transparently compositional natural deduction systems to perform complex reasoning tasks. In fact, we can use the W-scale to rank various types of scaffolding (cf. M. Wilson 2002). (1) Rotating falling blocks in Tetris, or moving furniture around the room to generate possible solutions of where to put things, involve the exploitation of spatial relationships among elements in the world in order to solve spatial problems. Since the elements do not represent anything other than themselves, they presumably score relatively low on H-cognition (“situated & concrete scaffolding”). (2) Arranging tokens of army soldiers on a map for the purpose of military decision-making involve the exploitation of spatial relationships, but applied to a more abstract task; so those activities score higher (“situated & abstract scaffolding”). (3) Arranging diagrammatic systems such Venn diagrams to solve syllogistic reasoning problems exploits the correspondences between the spatial/topological properties of circles with the mathematical properties of sets (“spatially grounded & abstract scaffolding”). (4) Being able to use natural deduction systems to solve inferential problems (spatially ungrounded & abstract scaffolding”) would presumably be a very advanced form of H-cognition.
Many other current debates in this arena can naturally be framed in the terms of this account. For example, starting from a Popperian creature with a perceptually grounded symbol systems architecture (Barsalou, etc), how Fregean can you get without natural language? Then, how far can you get with the incorporation of natural language (which would itself be perceptually grounded, rather than translated into amodal symbols)? Finally, how much further un-grounding work had to be accomplished by the cultural evolution of external representational systems from the emergence of graphical art to the invention of written language? The answers to those questions will potentially yield empirical generalizations that are couched in terms of H-cognition, thus providing further evidence for the robustness of your taxonomic kinds: (e.g., “Human brains can’t become Fregean without natural language,” Human talking brains can’t become Gregorian without such-and-such forms of external scaffolding” etc.)
What other implications does a 4E-friendly application of your account have for the topics that have already been discussed in this forum? First, consider the robustness of H-cognition as a psychological kind. According to Dehaene’s (1997) “triple-code” model, our adult mathematical competence is a hybrid mental faculty that arises from the informational integration of three separate systems: a biologically basic analog system for understanding magnitude, the representation of elementary arithmetic facts in the verbal system, and our visually based acquaintance with purely symbolic numerical representations. As an integrated faculty, our adult mathematical competence scores high on H-cognition; whereas each subsystem taken individually arguably does not. As you point out, our intuitive grasp of magnitude may not even be H-cognition. Verbal rote-learning of the multiplication table is not (obviously) compositional; it takes the child a while to realize that these operations can be arbitrarily applied to any given pairs of natural numbers. The ability to manipulate (eg) algebraic notations rides piggyback on the ability to read, which itself depends on the invasion and redeployment of areas of the visual cortex that were already well-adapted for certain aspects of visual shape recognition (Dehaene 2009). The evolution of writing systems towards greater readability by our biological brains was no doubt an important prerequisite for the development of highly scaffolded H-cognition; however, without the proper informational integration into the other systems, the ability to manipulate meaningless squiggles would presumably not be H-cognition. Does the emergence of hybrid mental faculties undermine the claim that H-cognition is a robust functional kind? No: as long as we can make important empirical generalizations about our adult mathematical competence in terms of H-cognition, this is compatible with causal heterogeneity and, at least potentially, large amounts of variation in the levels of H-cognition at the level of implementing mechanisms.
Next, the related issue of modularity. According to your account, modular systems score low on the W-scale, and thus have to be considered as L-cognition. Now consider the literature on situated problem-solving which is replete with examples in which people rely on the use of concrete, special-case oriented, idiosyncratic strategies which cannot be generalized beyond a very narrow task environment (Kirsh 2009). The classic example here is a dieter’s situated strategy for figuring out for her lunch portion how much cottage cheese two-thirds of her daily allotment is, which are three-quarters of a cup. What the person did was to turn over a one-cup tub of cottage cheese, criss-cross it to mark four quarters, remove three quarters, and then consume two of those three, which of course was exactly one half. As Kirsh (ibid.) points out, this strategy works only in a very narrow, activity-specific environment: e.g., if the dieticians might have decided that the daily allotment was 3/5 of cup, and the lunch portion was 7/16, such a strategy would fail, and dieters would have to undergo the general-purpose but more error-prone strategy of calculating fractions. However, this is not a problem because our problem-solving strategies typically co-evolve with the task environments in which we tend to solve them (e.g., if the dieticians’ recommendations were to change along the above lines, manufactures would presumably change the tub size to make the calculations easier). Is this H-cognition or L-cognition? Or, on a much grander scale, we have the vast range of highly specialized, special-purpose, “modular” character of most iPhone apps. For example, when I shazam a song that I hear in a bar to identify the singer, title, length, and album of a song, is this an instance of L-cognition or H-cognition?
On the one hand, one could argue that episodes of soft-assembled, modular problem-solving are not cases of H-cognition, similar to the biologically basic, perhaps hardwired forms of modular cognition. For example, Peretz and Coltheart (2006) have argued that our basic music faculty is highly modular, comprising a set of neurally isolable, functionally distinct processing components. If they are right, your account implies that our biologically basic music faculty is clearly not a case of H-cognition. On the other hand, there are important differences between the two cases. The augmented music faculty that is composed by the integrated User+Shazam system seems to deserve a significantly “higher” score on the W-scale than its biologically basic counterparts: it employs greater representational resources, as well as categories of greater abstraction; supports higher degrees of causal autonomy, and makes available inferences that take advantage of a great deal of what we know about these songs (though not necessarily known by the user herself before her use of Shazam.). If we accept this intuition, does this lead to a collapse of the difference between higher and lower mental faculties? How can the “music faculty” receive a low score when it is located inside the head, yet receive a much higher score when it is partly located in the environment?
There is, of course, an important difference between biologically hardwired modules and culturally soft-assembled “modular” components. Suppose it is true, as it seems quite plausible to me, that only Gregorian minds that already score very high on the W-scale are able to enhance their repertoire with a practically unlimited variety of special-purpose skills, practices, and artifacts that can be used for specific types of problem-solving. In other words, the problem of mental synthesis must already have been solved before extra layers of relatively modular (yet, at least in some cases, potentially knowledge-rich) forms of expertise can be gainfully added to the pre-existing cognitive repertoire of a Gregorian creature. Perhaps MM-architectures could never get to this point courtesy of biological evolution alone. But once our minds became sufficiently Gregorian, culturally advanced forms of “modular” problem-solving are capable of yielding relatively high forms of H-cognition, provided that they are informationally integrated with our overall cognitive architecture. In that case, Shazam and its cognates would also provide evidence that your taxonomy of H-cognition does not discriminate “by fiat” against (partly) modular systems.
Let me know what you think. Again, I really learned a great deal from engaging with your paper.
Cheers,
-Georg
Hi Dan
Thanks for your clarification. I absolutely do think what you set out is a worthwhile goal. I have been writing myself about the combination dimension (with more giving you greater cognitive power) so I found it fascinating to be thinking about the abstractness and autonomy dimensions at the same time.Georg:
Thanks for the highly detailed and thoughtful suggestions. These will be useful in revising the latter sections of the paper. I particularly appreciate the cases illustrating how achieving certain kinds of H-cognition is only possible given certain preconditions—as I pointed out in response to Robert above, this is where some of the relevant explanatory power in the account should be coming from.
I did have something like Dennett’s model in mind when developing mine (as well as other exemplars, such as Sterelny and Liz Camp). You’re also correct that H-cognition generally involves ways of decoupling cognitive activity from various constraints: of the categories of perception and action, of the immediate causal impingements of the world, of the local informational context, etc. While I don’t doubt that our symbolic and cultural activities in the world play a massive role in amplifying what we can do, H-cognitively speaking, I personally am on record as opposing both embodied and extended cognition. Even so, a host of (by my reckoning) extra-psychological structures certainly play a role in normal development of these capacities. It’s a nice point that the taxonomy can be used to cover ‘extended’ cognitive capacities as well as standardly interpreted intracranial ones; that’s exactly the kind of neutrality I was aiming at. I wouldn’t want, as you point out at the end, to have the view itself to discriminate between extended and intracranial capacities. That takes a separate argument.
Your music recognition case is a bit of a puzzler. The reason it’s hard to place is, I think, because whereas when I identify a piece of music by hearing it, it’s my own transducers and perceptual processes that are implicated in the process, whereas it’s odd to think of the phone’s mic as being one of *my* sensory surfaces. The phone’s processing seems to be a ‘sideways’ adjunct to my cognition, deployed in the event that I can’t myself pick out who sang that song. Playing along with the idea that this is really one of my extended cognitive capacities for a moment, devices like this function a bit like oracular voices. If I can’t figure something out, I can just ask the oracle for the answer, although how it does it is entirely obscure to me; I only see the results it dumps out. The relevant informational interface involves only my awareness of my musical ignorance, my decision to consult the oracle, and my knowledge of the oracle’s output—all of which are relatively H-cognitively mediated.
But I don’t think it should bother us that there is one L-cognitive and one H-cognitive way to recognize music. For one thing, there is a way of individuating the faculties on which they take different inputs and outputs. More importantly, though, the nature of their interface with the rest of cognition is different. They are placed differently relative to the perceptual systems, are initiated by different cognitive acts, send output to different systems, and so on. These locational or relative facts matter. This highlights the point that there is no fact about whether a capacity is H-cognitive or L-cognitive outside of the context of a particular animal’s cognitive architecture. The distinction is senseless, or at least undefined, without such information.
So, does that seem convincing? I don’t see any reason a 4E theorist in particular should object to it, anyway.
D
Hi Dan,
Thanks for these clarifications. Again, I’m mostly sympathetic to the issues you’re raising, including your point that whether a capacity is H-cognitive or L-cognitive depends on the context of a particular animal’s cognitive architecture (although it’s likely that I have a more plastic vision of what “the” human cognitive architecture looks like). At any rate, let me return to the point about the emergence of H-cognition from co-opting previously segregated L-cognitive components, and integrating them into one’s cognitive architecture. I’m interested in what further generalizations we can potentially derive based on your taxonomy (cf. Section 4). These are mostly empirical issues, but it’s worth speculating about them a bit.
First of all, I totally agree with the point that you made earlier in response to Robert, that it is certainly misguided to think that there would be/has been a uniform evolutionary trend towards greater degrees of H-cognition. You already cited several problems with this idea in your discussion of evolutionary-based proposals; Mithen’s (1996) theory that during hominid evolution, selective advantages have oscillated between favoring specialized, hardwired, and modularized intelligence and favoring general intelligence would be another example of this sort. Still, each time the pendulum swings towards cognitive fluidity, it seems to undergo a similar type of trajectory – a spiral leading to increasingly higher levels of mental complexity.
Example: Let’s suppose, for the sake of the argument, that Mithen is right and the “cultural explosion” that occurred 30,000-60,000 yrs ago happened when homo sapiens sapiens crossed a certain threshold of cognitive fluidity, after reversing a trend towards increased modularity that dominated early hominid evolution. This trend continued, leading to greater feats of mental synthesis concerning biologically basic forms of intracranial cognition. With a certain temporal (cultural) delay, this pattern was recapitulated, at a higher level of complexity, by the way in which our tools evolved with us. Consider the cultural evolution of ancient systems of writing, counting, and numbering, each of which followed the patterns towards increased representational abstraction, context-independence, and potential for informational integration. During the early (concrete, task-dependent, situated) stages of their development, who would have been able to predict the enormous amounts of H-cognition that you can squeeze out of biological brains – even those of our ancestors who had already fully achieved fairly high forms of biologically basic H-cognition! – if they dovetail with the right kind of symbolic structures.
Thus, if we limit your meta-architectural taxonomy to the description of particular types of advanced cognitive evolution, rather than trying to cover the entire realm of cognitive creatures, I can easily see how we coul gainfully precisify the metrics for each of the three properties you mention. You point out that the three properties are logically independent, and empirically dissociable, and as far as the big picture goes, I agree. But then again, there may be many more specific, local generalizations available once we restrict ourselves to charting the evolution of the “modern” human mind, especially the potential co-evolutionary dynamics between the three properties. For instance, it might be nomologically impossible for creatures like us to score high only on two but not the third of those dimensions. Theories that emphasize the co-evolution of brain, language, and symbolic thought, and how they “drag along” higher levels of meta-cognition, meta-cognitive awareness, and voluntary cognitive control would – if true – presumably support such generalizations.
[Ok, I gotta include one minor comment on your discussion of music cognition. You claim that it’s odd to think of the phone’s mic as being one of *my* sensory surfaces, while you take it as unproblematic to speak of “my own transducers” (or any other sub-personal perceptual/cognitive mechanisms inside the brain.) There are a few things an externalist might say in response to this. First, based on evidence about children’s naïve reasoning about property ownership, I’m not so sure about your intuitions; I would think that it’s more intuitive for a kid who grows up wearing glasses to consider those as *my” sensory surfaces, as opposed to speaking of “my” V1 or whatever. Second, I guess the really controversial claim would be that they can become part of me (i.e., part of my extended perceptual/cognitive system) rather than the weaker claim that they can be mine. If (BIG ‘if’!) the stronger claim is adjudicated in terms of the first-person phenomenology of one’s sense of agency or sense of ownership, then I would also be fairly confident that the boundaries not only of one’s body, but also of one’s self, do not always coincide with the biological boundaries of the organism (i.e., they sometimes include extra-corporeal structures, and sometimes exclude intracorporeal structures). However, I take it that the points you make about the relational/locational individuation of cognitive and perceptual system do not hinge on first-person intuitions about agency/ownership, so the above point may be moot; at least, it is not necessarily something an internalist has to deny. Anyway, let’s not get into what seems to be a tangential issue at best…)]
Best,
-Georg