Wikia

Psychology Wiki

Modularity of mind

Talk0
34,141pages on
this wiki
Revision as of 21:53, July 26, 2013 by Dr Joe Kiff (Talk | contribs)

(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)

Assessment | Biopsychology | Comparative | Cognitive | Developmental | Language | Individual differences | Personality | Philosophy | Social |
Methods | Statistics | Clinical | Educational | Industrial | Professional items | World psychology |

Biological: Behavioural genetics · Evolutionary psychology · Neuroanatomy · Neurochemistry · Neuroendocrinology · Neuroscience · Psychoneuroimmunology · Physiological Psychology · Psychopharmacology (Index, Outline)


Modularity of mind is the notion that a mind may, at least in part, be composed of innate neural structures or modules which have distinct established evolutionarily developed functions. Somewhat different definitions of "module" have been proposed by different authorities.

Early investigationsEdit

Historically, questions regarding the functional architecture of the mind have been divided into two different theories of the nature of the faculties. The first can be characterized as a horizontal view because it refers to mental processes as if they are interactions between faculties such as memory, imagination, judgement, and perception, which are not domain specific (e.g., a judgement remains a judgement whether it refers to a perceptual experience or to the conceptualization/comprehension process)). The second can be characterized as a vertical view because it claims that the mental faculties are differentiated on the basis of domain specificity, are genetically determined, are associated with distinct neurological structures, and are computationally autonomous.

The vertical vision goes back to the 19th century movement called phrenology and its founder Franz Joseph Gall, who claimed that the individual mental faculties could be associated precisely, in a sort of one to one correspondence, with specific physical areas of the brain. Hence, someone's level of intelligence, for example, could be literally "read off" from the size of a particular bump on his posterior parietal lobe. This simplistic view of modularity has been disproven over the course of the last century.

Fodor's Modularity of MindEdit

In the 1980s, however, Jerry Fodor revived the idea of the modularity of mind, although without the notion of precise physical localizability. Drawing from Noam Chomsky's idea of the language acquisition device and other work in linguistics as well as from the philosophy of mind and the implications of optical illusions, he became one of its most articulate proponents with the 1983 publication of Modularity of Mind.[1]

According to Fodor, a module falls somewhere between the behaviorist and cognitivist views of lower-level processes.

Behaviorists tried to replace the mind with reflexes which Fodor describes as encapsulated (cognitively impenetrable or unaffected by other cognitive domains) and non-inferential (straight pathways with no information added). Low level processes are unlike reflexes in that they are inferential. This can be demonstrated by poverty of the stimulus arguments in which the proximate stimulus, that which is initially received by the brain (such as the 2D image received by the retina), cannot account for the resulting output (for example, our 3D perception of the world), thus necessitating some form of computation.

In contrast, cognitivists saw lower level processes as continuous with higher level processes, being inferential and cognitively penetrable (influenced by other cognitive domains, such as beliefs). The latter has been shown to be untrue in some cases, such as with many visual illusions (ex. Müller-Lyer illusion), which can persist despite a person's awareness of their existence. This is taken to indicate that other domains, including one's beliefs, cannot influence such processes.

Fodor arrives at the conclusion that such processes are inferential like higher order processes and encapsulated in the same sense as reflexes.

Although he argued for the modularity of "lower level" cognitive processes in Modularity of Mind he also argued that higher level cognitive processes are not modular since they have dissimilar properties. The Mind Doesn't Work That Way, a reaction to Steven Pinker's How the Mind Works, is devoted to this subject.

Fodor (1983) states that modular systems must—at least to "some interesting extent"—fulfill certain properties:

  1. Domain specificity, modules only operate on certain kinds of inputs—they are specialised
  2. Informational encapsulation, modules need not refer to other psychological systems in order to operate
  3. Obligatory firing, modules process in a mandatory manner
  4. Fast speed, probably due to the fact that they are encapsulated (thereby needing only to consult a restricted database) and mandatory (time need not be wasted in determining whether or not to process incoming input)
  5. Shallow outputs, the output of modules is very simple
  6. Limited accessibility
  7. Characteristic ontogeny, there is a regularity of development
  8. Fixed neural architecture.

Pylyshyn (1999) has argued that while these properties tend to occur with modules, one stands out as being the real signature of a module; that is the encapsulation of the processes inside the module from both cognitive influence and from cognitive access.[2] This is referred to as "information encapsulation". One example is that conscious awareness of the Müller-Lyer illusion being an illusion does not correct the visual processing.[3]

Evolutionary psychology and massive modularityEdit

Other perspectives on modularity come from evolutionary psychology, particularly from the work of Leda Cosmides and John Tooby. This perspective suggests that modules are units of mental processing that evolved in response to selection pressures. On this view, much modern human psychological activity is rooted in adaptations that occurred earlier in human evolution, when natural selection was forming the modern human species.

Evolutionary psychologists propose that the mind is made up of genetically influenced and domain-specific[4] mental algorithms or computational modules, designed to solve specific evolutionary problems of the past.[5] Cosmides and Tooby also state in a brief "primer" on their website,[6] that "…the brain is a physical system. It functions like a computer," "…the brain’s function is to process information," "different neural circuits are specialized for solving different adaptive problems," and "our modern skulls house a stone age mind."

The definition of module has caused confusion and dispute. J. A. Fodor initially defined module as "functionally specialized cognitive systems" that have nine features but not necessarily all at the same time. In his views modules can be found in peripheral processing such as low-level visual processing but not in central processing. Later he narrowed the two essential features to domain-specificity and information encapsulation. Frankenhuis and Ploeger[3] write that domain-specificity means that "a given cognitive mechanism accepts, or is specialized to operate on, only a specific class of information". Information encapsulation means that information processing in the module cannot be affected by information in the rest of the brain. One example being awareness that certain optical illusion, caused by low level processing, are false not preventing the illusions from persisting.[3]

Evolutionary psychologists instead usually define modules as functionally specialized cognitive systems that are domain-specific and may also contain innate knowledge about the class of information processed. Modules can be found also for central processing. This theory is sometimes referred to as massive modularity.[3]

A 2010 review by evolutionary psychologists Confer et al. suggested that domain general theories, such as for "rationality," has several problems: 1. Evolutionary theories using the idea of numerous domain-specific adaptions have produced testable predictions that have been empirically confirmed; the theory of domain-general rational thought has produced no such predictions or confirmations. 2. The rapidity of responses such as jealousy due to infidelity indicates a domain-specific dedicated module rather than a general, deliberate, rational calculation of consequences. 3. Reactions may occur instinctively (consistent with innate knowledge) even if a person have not learned such knowledge. One example being that in the ancestral environment it is unlikely that males during development learn that infidelity (usually secret) may cause paternal uncertainty (from observing the phenotypes of children born many months later and making a statistical conclusion from the phenotype dissimilarity to the cuckolded fathers).[7] With respect to general purpose problem solvers, Barkow, Cosmides, and Tooby (1992) have suggested in The Adapted Mind: Evolutionary Psychology and The Generation of Culture that a purely general problem solving mechanism is impossible to build due to the frame problem. Clune et al. (2013) have argued that computer simulations of the evolution of neural nets suggest that modularity evolves because, compared to non-modular networks, connection costs are lower.[8]

Several groups of critics, including psychologists working within evolutionary frameworks,[9] argue that the massively modular theory of mind does little to explain adaptive psychological traits. Proponents of other models of the mind argue that the computational theory of mind is no better at explaining human behavior than a theory with mind entirely a product of the environment. Even within evolutionary psychology there is discussion about the degree of modularity, either as a few generalist modules or as many highly specific modules.[9][10] Other critics suggest that there is little empirical support in favor of the domain-specific theory beyond performance on the Wason selection task, a task critics state is too limited in scope to test all relevant aspects of reasoning.[11][12] Moreover, critics argue that Cosmides and Tooby's conclusions contain several inferential errors and that the authors use untested evolutionary assumptions to eliminate rival reasoning theories.[13][11]

Wallace (2010) observes that the evolutionary psychologists' definition of "mind" have been heavily influenced by cognitivism and/or information processing definitions of the mind.[14] Critics point out that these assumptions underlying evolutionary psychologists' hypotheses are controversial and have been contested by some psychologists, philosophers, and neuroscientists. For example, Jaak Panksepp, an affective neuroscientist, point to the "remarkable degree of neocortical plasticity within the human brain, especially during development" and states that "the developmental interactions among ancient special-purpose circuits and more recent general-purpose brain mechanisms can generate many of the "modularized" human abilities that evolutionary psychology has entertained."[9]

Philosopher David Buller agrees with the general argument that the human mind has evolved over time but disagrees with the specific claims evolutionary psychologists make. He has argued that the contention that the mind consists of thousands of modules, including sexually dimorphic jealousy and parental investment modules, are unsupported by the available empirical evidence.[15] He has suggested that the "modules" result from the brain's developmental plasticity and that they are adaptive responses to local conditions, not past evolutionary environments.[16] However, Buller has also stated that even if massive modularity is false this does not necessarily have broad implications for evolutionary psychology. Evolution may create innate motives even without innate knowledge.[17]

In contrast to modular mental structure, some theories posit domain-general processing, in which mental activity is distributed across the brain and cannot be decomposed, even abstractly, into independent units. A staunch defender of this view is William Uttal, who argues in The New Phrenology (2003) that there are serious philosophical, theoretical, and methodological problems with the entire enterprise of trying to localise cognitive processes in the brain.[18] Part of this argument is that a successful taxonomy of mental processes has yet to be developed.

Merlin Donald argues that over evolutionary time the mind has gained adaptive advantage from being a general problem solver.[19] The mind, as described by Donald, includes module-like "central" mechanisms, in addition to more recently evolved "domain-general" mechanisms.

Arguments Against ModularityEdit

In contrast to modular mental structure, some theories posit domain-general processing, in which mental activity is distributed across the brain and cannot be decomposed, even abstractly, into independent units. A staunch defender of this view is William Uttal, who argues in The New Phrenology (2003) that there are serious philosophical, theoretical, and methodological problems with the entire enterprise of trying to localize cognitive processes in the brain. Part of this argument is that a successful taxonomy of mental processes has yet to be developed.

See alsoEdit

ReferencesEdit

  1. Fodor, Jerry A. (1983). Modularity of Mind: An Essay on Faculty Psychology. Cambridge, Mass.: MIT Press. ISBN 0-262-56025-9
  2. Pylyshyn, Z.W. (1999). Is vision continuous with cognition? The case for cognitive impenetrability of visual perception. Behavioral and Brain Sciences, 22(3), 341-423. Full text
  3. 3.0 3.1 3.2 3.3 DOI:10.1080/09515080701665904
    This citation will be automatically completed in the next few minutes. You can jump the queue or expand by hand
  4. Cosmides, L. & Tooby, J. (1994). Origins of Domain Specificity: The Evolution of Functional Organization. In L.A. Hirschfeld and S.A. Gelmen, eds., Mapping the Mind: Domain Specificity in Cognition and Culture. Cambridge: Cambridge University Press. Reprinted in R. Cummins and D.D. Cummins, eds., Minds, Brains, and Computers. Oxford: Blackwell, 2000, 523-543.
  5. Cosmides, L., & Tooby, J. (1992). Cognitive Adaptations for Social Exchange. In Barkow, Cosmides, and Tooby 1992, 163-228.
  6. Evolutionary Psychology Primer by Leda Cosmides and John Tooby. Psych.ucsb.edu. URL accessed on 2012-07-23.
  7. DOI:10.1037/a0018413
    This citation will be automatically completed in the next few minutes. You can jump the queue or expand by hand
  8. (2013). The evolutionary origins of modularity. Proceedings of the Royal Society 280 (1755).
  9. 9.0 9.1 9.2 Panksepp, J. & Panksepp, J. (2000). The Seven Sins of Evolutionary Psychology. Evolution and Cognition, 6:2, 108-131.
  10. Buller, David J. and Valerie Gray Hardcastle (2005) Chapter 4. "Modularity", in Buller, David J. The Adapted Mind: Evolutionary Psychology. The MIT Press. pp. 127 - 201
  11. 11.0 11.1 (1995). Logical reasoning and domain specificity. Biology and Philosophy 10 (1): 1–37.
  12. (2010) "The Mental Logic Theory of Conditional Propositions" Cognition and Conditionals: Probability and Logic in Human Thinking, 39–54, New York: Oxford University Press.
  13. (1999). Evolutionary Psychology: The Burdens of Proof. Biology and Philosophy 19 (2): 211–233.
  14. Wallace, B. (2010). Getting Darwin Wrong: Why Evolutionary Psychology Won’t Work. Exeter, UK: Imprint Academic.
  15. (2005). Evolutionary psychology: the emperor's new paradigm. Trends in Cognitive Sciences 9 (6): 277–283.
  16. (2000). Evolutionary Psychology, Meet Developmental Neurobiology: Against Promiscuous Modularity. Brain and Mind 1 (3): 307–325.
  17. (2005). Get Over: Massive Modularity. Biology & Philosophy 20 (4): 881–891.
  18. Uttal, William R. (2003). The New Phrenology: The Limits of Localizing Cognitive Processes in the Brain. Cambridge, Mass.: MIT Press.
  19. Donald, A Mind So Rare: The Evolution of Human Consciousness [1].

Further readingEdit

Key TextsEdit

BooksEdit

  • Fodor, Jerry A. (1983). Modularity of Mind: An Essay on Faculty Psychology. Cambridge, Mass.: MIT Press. ISBN 0-262-56025-9
  • Uttal, William R. (2003). The New Phrenology: The Limits of Localizing Cognitive Processes in the Brain. Cambridge, Mass.: MIT Press.

PapersEdit

Additional materialEdit

BooksEdit

PapersEdit

External linksEdit

This page uses Creative Commons Licensed content from Wikipedia (view authors).

Around Wikia's network

Random Wiki