Showing posts with label natural kinds. Show all posts
Showing posts with label natural kinds. Show all posts

Tuesday, 6 November 2012

Functionalism reconsidered

I've long considered myself to be a functionalist about mental states such as belief and pain. Functionalism is the theory that mental states should be identified not by their physical instantiation but by their functional role, i.e. the role that they play within a given system. The classic example is pain, which is said to be defined by behaviours such as flinch responses, yelling out, and crying (and perhaps a particular kind of first-person experience). One of the main motivations for functionalism is the "Martian intuition" - the intuition that were a silicon-based Martian to exhibit pain-behaviour, we would want to say that it is in pain, despite it lacking a carbon-based nervous system like our own. A less exotic intuition is that an octopus or capuchin monkey can probably feel pain, despite the exact physical instantiation of this pain differing from our own.


Martian octopus, perhaps in pain? 
(with permission from Ninalyn @ http://studiodecoco.tumblr.com/)

However I'm now beginning to suspect that there might be more than a few problems with functionalism. For starters, functional states are often defined as being those that are "relevantly similar" to an imagined paradigm case - thus, a Martian who screamed and recoiled when we punched might be said to be in pain, but one that laughed and clapped its hands (tentacles?) probably wouldn't. This is fine up to a point, especially in seemingly clear-cut cases like the above, but what should we say when we're faced with the inevitable borderline case?

Whether or not fish can feel pain seems to be a case like this. Research into fish pain behaviour is contentious - whilst fish exhibit apparent pain behaviour, they have only recently been shown to exhibit more complex pain avoidance behaviour that might be thought essential to pain. The problem is not just a lack of evidence either, there's a more fundamental lack of clarity about how exactly we should define the functional role of pain, or indeed any other mental state.

Having said that, the problem isn't limited to the functionalist account of mental states. Biological species appear to form vague natural kinds, a problem which has motivated the idea of homeostatic property cluster kinds, categories of kinds that share some, but not all, of their properties. So we maybe we could say that functional kinds, such as pain, are a category of HPC kinds. That still wouldn't necessarily give us a straight answer in genuine border-line cases, but at least we'd have good reason to think functional roles might sometimes pick out genuine kinds (albeit perhaps not natural kinds)

The problems don't stop there though. By arguing that it entails a radical form of cognitive extension, Mark Sprevak has pushed functionalism to its logical extreme. If he is correct then being a functionalist would commit you to apparently absurd conclusions,1 such as that the entire contents of the Dictionary of Philosophy sitting on my desk form part of my cognitive system, or that my capacity for mental arithmetic is bounded only by my access to electronic computing power. I think there might be a way for functionalism to avoid the full force of this argument, but it comes with its own problems and costs.

Essentially what the functionalist needs to do is to stop talking about cognition and mental states as though they were one kind of thing. They're not, and rather than lumping memory, personality, beliefs  and so on into one unitary framework, we need to look at giving finer-grained functional descriptions in each case. This might even mean getting rid of some mental states, such as belief, or at least admitting that they're more complex than we first thought. This approach will still entail some degree of cognitive extension, but hopefully in a more subtle and intuitive way. So whilst it might not be true that the contents of the Dictionary are part of my 'cognitive system', they may nonetheless form part of a belief-like system, albeit one that functions differently to my regular belief system. 

Would this still be functionalism? In a sense yes, because it would maintain a degree of multiple realisability, only at a more fine-grained level. So a Martian with a silicon brain might have beliefs, but equally they might have something more akin to the belief-like system that is constituted by me-and-the-Dictionary. The problem with functionalism is that it tends to reify our folk intuitions about mental states, and we need to remember that these might not be entirely accurate. I suppose I'm beginning to lean towards a form of eliminativism, although I still think that there's room for an instrumentalist account of functional roles. 


1. I say "apparently" because I'm not entirely convinced that one shouldn't just bite the bullet and accept these conclusions. That's probably a post for another day though.

Thursday, 4 October 2012

HPCK and Modal Representations

It's been a busy month, moving into a new flat and starting an MSc. I'm studying full time again, which in terms of blogging is a mixed blessing - lots of material, but very little time.

It's nice when you can combine two previously isolated ideas, and that's what I'm going to try and do today. One comes from the philosophy of science, Boyd's "homeostatic property cluster" theory of natural kinds, and the other is an idea from the philosophy of mind, that our mental images might not be entirely separate from our sensory perception.

I'll start with modal representations, because they're probably simpler. A mental representation is basically a mental state that stands for some part of the external world (Clark 1997: 463), whatever we take that to mean. Mental representation is a thorny topic, but all I'm interested in here is one aspect of the issue: whether such representations are composed of sensory information (modal) or are purely abstract (amodal). For example, does our representation of a sunny day call to mind the pleasant feeling of the sun on our skin, or do we somehow comprehend it in isolation from any sensation? For the previous century (analytic) philosophers tended to pick the latter option, no doubt influenced by classical logic, but some (relatively) recent experiments have questioned that assumption. It seems that there is a systematic connection between representations and the sensory qualities of what they represent, as demonstrated by experiments such as those conducted by Zwan, Stanfield & Yaxley (2002) and by Glenberg & Kaschak (2002). The implications of these experiments are still being debated, but one interpretation is that our representations (and by extension, our concepts) are composed of bundles of modal (sensory) data, rather than discrete, amodal definitions.

This is where Boyd comes in. His theory is a form of realism about natural kinds, but I think that it shares some interesting similarities with the idea of modal properties. Motivated by the messiness of biological kinds, Boyd characterises a natural kind as sharing a cluster of properties, none of which are necessary or sufficient. These kinds are rooted in the causal structure of the world, and are thus real, but they allow for the flexibility that is necessary when it comes to biological kinds. Given that our access to kinds is mediated by our senses, I think it might make sense to identify the modal bundles that I described above with Boyd's property clusters. Our concept or representation of a cow might consist of the vague appearance of a cow, the smell of cow dung, and the monotonous sound they make - and this will in some sense correspond with (at least some of) the properties in the natural kind cluster "cow". Boyd's point is that there doesn't have to be an exact matching for every instantiation of a natural kind, so everyone's perception of cow's can (and will) be subtly different.

I still haven't quite got to grips with Boyd's theory, and I'm not sure how much he would support this idea, but I think it could allow for an evolutionary justification of how classify natural kinds. This would be similar to Quine's empiricist position (see his 1969), and might not be as realist as Boyd would like.

  • Clark, A. 1997. "The Dynamical Challenge." Cognitive Science 21(4): 461-81.
  • Glenberg, A. & Kaschak, M. 2002. "Grounding language in action." Psychonomic Bulletin and Review 9: 558-65.
  • Quine, W.V.O. 1969."Natural Kinds." In Ontological Relativity and Other Essays. New York and London: Columbia University Press.
  • Zwan, R., Stanfield, R. & Yaxley, R. 2002. "Language comprehenders routinely represent the shape of objects?" Psychological Science 13: 168-71.