Psychology Wiki
 
(9 intermediate revisions by the same user not shown)
Line 1: Line 1:
 
{{SocialPsy}}
 
{{SocialPsy}}
  +
{{expert}}
   
A '''cognitive bias''' is any of a wide range of [[observer effect]]s identified in [[cognitive science]] and [[social psychology]] including very basic [[Statistics|statistical]], social [[attribution]], and memory errors that are common to all [[human]] beings. Biases drastically skew the reliability of [[anecdotal evidence|anecdotal]] and [[Evidence (law)|legal evidence]]. Social biases, usually called [[attributional bias]]es affect our everyday social interactions. And biases related to [[probability]] and decision making significantly affect the [[scientific method]] which is deliberately designed to minimize such [[bias]] from any one observer.
+
A '''cognitive bias''' is a person's tendency to make errors in judgment based on [[cognitive]] factors, and is a phenomenon studied in [[cognitive science]] and [[social psychology]]. Forms of cognitive bias include errors in [[Statistics|statistical]] judgment, [[social attribution]], and memory that are common to all [[human]] beings. Such [[bias]]es drastically skew the reliability of [[anecdotal evidence|anecdotal]] and legal evidence. These are thought to be based upon [[heuristic]]s, or rules of thumb, which people employ out of habit or evolutionary necessity.
   
 
==Overview==
 
==Overview==
Bias arises from various life, loyalty and local risk and attention concerns that are difficult to separate or codify. They were first identified by [[Amos Tversky]] and [[Daniel Kahneman]] as a foundation of [[behavioral economics]]. Tversky and Kahneman claim that they are at least partially the result of problem-solving using [[heuristic]]s, including the [[availability heuristic]] and the [[representativeness heuristic]].
+
Bias arises from various life, loyalty and local risk and attention concerns that are difficult to separate or codify. Much of the present scientific understanding of biases stems from the work of [[Amos Tversky]] and [[Daniel Kahneman]] and their colleagues,{{Fact|date=April 2008}} whose experiments demonstrated distinct and replicable ways in which human judgment and decision-making differ from [[rational choice theory]]. This led to Tversky and Kahneman developing [[prospect theory]] as an alternative.{{Fact|date=April 2008}} Tversky and Kahneman claim that the biases they identified are at least partially the result of problem-solving using mental short-cuts or "heuristics", for instance using how readily or vividly something comes to mind as an indication of how often or how recently it was encountered (the [[availability heuristic]]). Other biases have been demonstrated in separate experiments, such as the [[confirmation bias]] demonstrated by Peter C. Wason{{Fact|date=April 2008}}
   
Recently, some scientists ([[David Funder]] and [[Joachim Krueger]]) have raised doubt as to whether all of the 'biases' are in fact errors. Their theories hold that some so called 'biases' may in fact be 'approximation shortcuts', that aid humans in making predictions when information is in short supply. For example, the [[false consensus effect]] may be viewed as a reasonable estimation based on a single known data point, your own opinion, instead of a false belief that other people agree with you.
+
Not all of 'biases' are necessarily errors. David Funder and Joachim Krueger have argued that some so called 'biases' may in fact be 'approximation shortcuts', which aid humans in making predictions when information is in short supply.{{Fact|date=April 2008}} For example, the [[false consensus effect]] may be viewed as a reasonable estimation based on a single known data point, your own opinion, instead of a false belief that other people agree with you.
   
 
==Types of cognitive biases==
 
==Types of cognitive biases==
  +
Biases can be distinguished on a number of dimensions. For example, there are biases specific to ''groups'' (such as the [[risky shift]]) as well as biases at the individual level.
The following is a list of the more commonly studied cognitive biases
 
* [[Hindsight bias]] sometimes called the "I-knew-it-all-along" effect, is the inclination to see past events as being predictable
 
* [[Fundamental attribution error]] the tendency for people to over-emphasize personality-based explanations for behaviors observed in others while under-emphasizing the role and power of situational influences on the same behavior.
 
* [[Confirmation bias]] the tendency to search for or interpret information in a way that confirms one's preconceptions.
 
* [[Self-serving bias]] the tendency to claim more responsibility for successes than failures. It may also manifest itself as a tendency for people to evaluate ambiguous information in a way beneficial to their interests.
 
   
  +
Some biases affect decision-making, where the ''desirability'' of options has to be considered (e.g. [[Sunk cost|Sunk Cost fallacy]]). Others such as [[Illusory correlation]] affect judgment of how likely something is, or of whether one thing is the cause of another. A distinctive class of biases affect memory,<ref name=Schacter1999>{{citation
There are of course several other cognitive biases, as people keep imagining more all the time. In fact, one might imagine that there is a cognitive bias that is a tendency to invent more cognitive biases. See the [[list of cognitive biases]].
 
  +
| author = Schacter, D. L.
  +
| year = 1999
  +
| title =The Seven Sins of Memory: Insights From Psychology and Cognitive Neuroscience
  +
| journal = American Psychologist
  +
| volume = 54
  +
| issue = 3
  +
| pages = 182-203}}</ref> such as consistency bias (remembering one's past attitudes and behaviour as more similar to one's present attitudes).
   
  +
Some biases reflect a subject's ''motivation'', <ref name=Kunda1990>{{citation
==Papers==
 
  +
| author= Kunda, Z.
  +
| year = 1990
  +
| title =The Case for Motivated Reasoning
  +
| journal =Psychological Bulletin
  +
| volume = 108
  +
| issue= 3
  +
| pages = 480-498
  +
}}</ref> for example the desire for a positive self-image leading to [[Egocentric bias]]<ref name =Hoorens1993>{{citation
  +
|author = Hoorens, V.
  +
| year = 1993
  +
| contribution = Self-enhancement and Superiority Biases in Social Comparison
  +
| title = European Review of Social Psychology 4
  +
| editor = Stroebe, W. and Hewstone, Miles
  +
| publisher = Wiley }}</ref> and the avoidance of unpleasant [[cognitive dissonance]]. Other biases are due to the particular way the brain perceives, forms memories and makes judgments. This distinction is sometimes described as "[[Hot cognition]]" versus "Cold Cognition", as motivated cognition can involve a state of [[arousal]].
   
  +
Among the "cold" biases, some are due to ''ignoring relevant information'' (e.g. [[Neglect of probability]]), whereas some involve a decision or judgement being ''affected by irrelevant information'' (for example the [[Framing (social sciences)|Framing effect]] where the exact same problem receives different responses depending on how it is described) or giving ''excessive weight'' to an unimportant but salient feature of the problem (e.g. [[Anchoring]]).
* Haselton, M. G. & Buss, D. M. (2003). Biases in social judgment: Design flaws or design features? In J. Forgas, K. Williams, & B. von Hippel (Eds.) ''Responding to the Social World: Implicit and Explicit Processes in Social Judgments and Decisions'', (pp. 23-43). Cambridge: Cambridge University Press. [http://homepage.psy.utexas.edu/homepage/Group/BussLAB/pdffiles/emt-sydney_chapter.pdf Full text]
 
   
  +
The fact that some biases reflect motivation, and in particular the motivation to have positive attitudes to oneself<ref name="Hoorens1993">Booba vs. Kiki!</ref> accounts for the fact that many biases are self-serving or self-directed (e.g. [[Illusion of asymmetric insight]], [[Self-serving bias]], [[Projection bias]]). There are also biases in how subjects evaluate in-groups or out-groups; evaluating in-groups as more diverse and "better" in many respects, even when those groups are arbitrarily-defined ([[Ingroup bias]], [[Outgroup homogeneity bias]]).
==External links==
 
  +
* [http://www.williamjames.com/Science/ERR.htm The Roots of Consciousness: To Err is human]
 
 
The following is a list of the more commonly studied cognitive biases.
* [http://www.cycleback.com/conceits.html Conceits and How Humans Interpret Their World]
 
 
{{For|other noted biases|list of cognitive biases}}
  +
* [[Anchoring]] on a past reference.
  +
* [[Framing (social sciences)|Framing]] by using a too narrow approach and description of the situation or issue.
 
* [[Hindsight bias]], sometimes called the "I-knew-it-all-along" effect, is the inclination to see past events as being predictable.
 
* [[Fundamental attribution error]] is the tendency for people to over-emphasize personality-based explanations for behaviors observed in others while under-emphasizing the role and power of situational influences on the same behavior.
 
* [[Confirmation bias]] is the tendency to search for or interpret information in a way that confirms one's preconceptions; this is related to the concept of [[cognitive dissonance]].
 
* [[Self-serving bias]] is the tendency to claim more responsibility for successes than failures. It may also manifest itself as a tendency for people to evaluate ambiguous information in a way beneficial to their interests.
  +
  +
==Practical Significance==
  +
Many social institutions rely on individuals to make rational judgments . A fair [[jury trial]], for example, requires that the jury ignore irrelevant features of the case (such as the attractiveness of the defendant), weigh the relevant features appropriately, consider different possibilities open-mindedly and resist [[fallacies]] such as [[appeal to emotion]]. The various biases demonstrated in these psychological experiments suggest that people will fail to do all these things. However, they fail to do so in systematic, directional ways that are predictable.
  +
  +
A&H Method decision maps illustrate the combination of reasons-claim argument strands as well as the influences of cognitive heuristics and psychological dominance structuring which emerge from those data. Researchers can compare decision maps illustrating how many different people have made a decision about the same question (e.g. "Should I have a doctor look at this troubling breast cancer symptom I've discovered." "Why did I ignore the evidence that the project was going over budget?") and then craft potential cognitive interventions aimed at improving decision making outcomes.
   
 
==See also==
 
==See also==
  +
{{Col-begin}}
  +
{{Col-1-of-3}}
  +
* [[Critical thinking]]
  +
* [[Cognitive bias modification]]
  +
* [[Cognitive distortion]]
 
* [[Cognitive psychology]]
 
* [[Cognitive psychology]]
  +
{{Col-2-of-3}}
* [[list of cognitive biases]]
 
  +
* [[Emotional bias]]
  +
* [[Evolutionary psychology]]
  +
* [[Fallacy]]
  +
* [[Filtering]]
  +
{{Col-3-of-3}}
  +
* [[Inductive bias]]
  +
* [[List of cognitive biases]]
  +
* [[Meta-model (NLP)]]
  +
* [[Mind reading]]
  +
* [[Over generalization]]
  +
* [[Source criticism]]
  +
{{col-end}}
   
  +
==References==
  +
{{reflist | 2}}
  +
  +
==Further reading==
  +
* Haselton, M. G., Nettle, D. & Andrews, P.W. (2005). The evolution of cognitive bias. In [[David Buss|D. M. Buss]] (Ed.), ''Handbook of Evolutionary Psychology,'' (pp. 724-746). Hoboken: Wiley. [http://www.sscnet.ucla.edu/comm/haselton/webdocs/handbookevpsych.pdf Full text]
  +
* Eiser, J. R. and Joop van der Pligt (1988) ''Attitudes and Decisions'' London: Routledge. ISBN 978-0415011129
  +
* Fine, Cordelia (2006) ''A Mind of its Own: How your brain distorts and deceives'' Cambridge, UK: Icon Books. ISBN 1-84046-678-2
  +
* Heuer, Richards J. Jr. (1999) Psychology of Intelligence Analysis. Central Intelligence Agency. http://www.au.af.mil/au/awc/awcgate/psych-intel/art5.html
  +
* Kahneman D., Slovic P., and Tversky, A. (Eds.) (1982) ''Judgment Under Uncertainty: Heuristics and Biases''. New York: Cambridge University Press ISBN 978-0521284141
  +
* Massimo Piatelli-Palmarini (1994). ''Inevitable Illusions: How Mistakes of Reason Rule Our Minds'' New York: John Wiley & Sons. ISBN 0-471-15962-X
  +
* Nisbett, R., and Ross, L. (1980) ''Human Inference: Strategies and shortcomings of human judgement''. Englewood Cliffs, NJ: Prentice-Hall ISBN 978-0134451305
  +
* Stuart Sutherland (2007). ''Irrationality: The Enemy Within'' Second Edition (First Edition 1994) Pinter & Martin. ISBN 978-1905177073
  +
* Tavris, Carol and Elliot Aronson (2007). ''Mistakes Were Made (But Not by Me): Why We Justify Foolish Beliefs, Bad Decisions and Hurtful Acts'' Orlando, Florida: Harcourt Books. ISBN 978-0-15-101098-1
  +
 
==External links==
 
* [http://www.williamjames.com/Science/ERR.htm The Roots of Consciousness: To Err Is human]
  +
* [http://www.cxoadvisory.com/gurus/Fisher/article/ Cognitive bias in the financial arena]
  +
  +
[[Category:Cognition]]
 
[[Category:Cognitive biases| ]]
 
[[Category:Cognitive biases| ]]
  +
[[Category:Decision theory]]
  +
[[Category:Intelligence analysis]]
  +
  +
<!--
  +
[[ca:Biaix cognitiu]]
  +
[[de:Fangschluss]]
  +
[[es:Sesgo cognitivo]]
  +
[[fr:Biais cognitif]]
  +
[[he:הטיה קוגניטיבית]]
  +
[[nl:Denkfout]]
  +
[[ja:認知バイアス]]
  +
[[pl:Błąd poznawczy]]
  +
[[ro:Distorsiune cognitivă]]
  +
-->
  +
   
 
{{enWP|Cognitive bias}}
 
{{enWP|Cognitive bias}}

Latest revision as of 00:33, 12 July 2013

Assessment | Biopsychology | Comparative | Cognitive | Developmental | Language | Individual differences | Personality | Philosophy | Social |
Methods | Statistics | Clinical | Educational | Industrial | Professional items | World psychology |

Social psychology: Altruism · Attribution · Attitudes · Conformity · Discrimination · Groups · Interpersonal relations · Obedience · Prejudice · Norms · Perception · Index · Outline


This article is in need of attention from a psychologist/academic expert on the subject.
Please help recruit one, or improve this page yourself if you are qualified.
This banner appears on articles that are weak and whose contents should be approached with academic caution.

A cognitive bias is a person's tendency to make errors in judgment based on cognitive factors, and is a phenomenon studied in cognitive science and social psychology. Forms of cognitive bias include errors in statistical judgment, social attribution, and memory that are common to all human beings. Such biases drastically skew the reliability of anecdotal and legal evidence. These are thought to be based upon heuristics, or rules of thumb, which people employ out of habit or evolutionary necessity.

Overview

Bias arises from various life, loyalty and local risk and attention concerns that are difficult to separate or codify. Much of the present scientific understanding of biases stems from the work of Amos Tversky and Daniel Kahneman and their colleagues,[How to reference and link to summary or text] whose experiments demonstrated distinct and replicable ways in which human judgment and decision-making differ from rational choice theory. This led to Tversky and Kahneman developing prospect theory as an alternative.[How to reference and link to summary or text] Tversky and Kahneman claim that the biases they identified are at least partially the result of problem-solving using mental short-cuts or "heuristics", for instance using how readily or vividly something comes to mind as an indication of how often or how recently it was encountered (the availability heuristic). Other biases have been demonstrated in separate experiments, such as the confirmation bias demonstrated by Peter C. Wason[How to reference and link to summary or text]

Not all of 'biases' are necessarily errors. David Funder and Joachim Krueger have argued that some so called 'biases' may in fact be 'approximation shortcuts', which aid humans in making predictions when information is in short supply.[How to reference and link to summary or text] For example, the false consensus effect may be viewed as a reasonable estimation based on a single known data point, your own opinion, instead of a false belief that other people agree with you.

Types of cognitive biases

Biases can be distinguished on a number of dimensions. For example, there are biases specific to groups (such as the risky shift) as well as biases at the individual level.

Some biases affect decision-making, where the desirability of options has to be considered (e.g. Sunk Cost fallacy). Others such as Illusory correlation affect judgment of how likely something is, or of whether one thing is the cause of another. A distinctive class of biases affect memory,[1] such as consistency bias (remembering one's past attitudes and behaviour as more similar to one's present attitudes).

Some biases reflect a subject's motivation, [2] for example the desire for a positive self-image leading to Egocentric bias[3] and the avoidance of unpleasant cognitive dissonance. Other biases are due to the particular way the brain perceives, forms memories and makes judgments. This distinction is sometimes described as "Hot cognition" versus "Cold Cognition", as motivated cognition can involve a state of arousal.

Among the "cold" biases, some are due to ignoring relevant information (e.g. Neglect of probability), whereas some involve a decision or judgement being affected by irrelevant information (for example the Framing effect where the exact same problem receives different responses depending on how it is described) or giving excessive weight to an unimportant but salient feature of the problem (e.g. Anchoring).

The fact that some biases reflect motivation, and in particular the motivation to have positive attitudes to oneself[3] accounts for the fact that many biases are self-serving or self-directed (e.g. Illusion of asymmetric insight, Self-serving bias, Projection bias). There are also biases in how subjects evaluate in-groups or out-groups; evaluating in-groups as more diverse and "better" in many respects, even when those groups are arbitrarily-defined (Ingroup bias, Outgroup homogeneity bias).

The following is a list of the more commonly studied cognitive biases.

For other noted biases, see list of cognitive biases.
  • Anchoring on a past reference.
  • Framing by using a too narrow approach and description of the situation or issue.
  • Hindsight bias, sometimes called the "I-knew-it-all-along" effect, is the inclination to see past events as being predictable.
  • Fundamental attribution error is the tendency for people to over-emphasize personality-based explanations for behaviors observed in others while under-emphasizing the role and power of situational influences on the same behavior.
  • Confirmation bias is the tendency to search for or interpret information in a way that confirms one's preconceptions; this is related to the concept of cognitive dissonance.
  • Self-serving bias is the tendency to claim more responsibility for successes than failures. It may also manifest itself as a tendency for people to evaluate ambiguous information in a way beneficial to their interests.

Practical Significance

Many social institutions rely on individuals to make rational judgments . A fair jury trial, for example, requires that the jury ignore irrelevant features of the case (such as the attractiveness of the defendant), weigh the relevant features appropriately, consider different possibilities open-mindedly and resist fallacies such as appeal to emotion. The various biases demonstrated in these psychological experiments suggest that people will fail to do all these things. However, they fail to do so in systematic, directional ways that are predictable.

A&H Method decision maps illustrate the combination of reasons-claim argument strands as well as the influences of cognitive heuristics and psychological dominance structuring which emerge from those data. Researchers can compare decision maps illustrating how many different people have made a decision about the same question (e.g. "Should I have a doctor look at this troubling breast cancer symptom I've discovered." "Why did I ignore the evidence that the project was going over budget?") and then craft potential cognitive interventions aimed at improving decision making outcomes.

See also

References

  1. Schacter, D. L. (1999), "The Seven Sins of Memory: Insights From Psychology and Cognitive Neuroscience", American Psychologist 54 (3): 182-203 
  2. Kunda, Z. (1990), "The Case for Motivated Reasoning", Psychological Bulletin 108 (3): 480-498 
  3. 3.0 3.1 Hoorens, V. (1993), "Self-enhancement and Superiority Biases in Social Comparison", in Stroebe, W. and Hewstone, Miles, European Review of Social Psychology 4, Wiley  Cite error: Invalid <ref> tag; name "Hoorens1993" defined multiple times with different content

Further reading

  • Haselton, M. G., Nettle, D. & Andrews, P.W. (2005). The evolution of cognitive bias. In D. M. Buss (Ed.), Handbook of Evolutionary Psychology, (pp. 724-746). Hoboken: Wiley. Full text
  • Eiser, J. R. and Joop van der Pligt (1988) Attitudes and Decisions London: Routledge. ISBN 978-0415011129
  • Fine, Cordelia (2006) A Mind of its Own: How your brain distorts and deceives Cambridge, UK: Icon Books. ISBN 1-84046-678-2
  • Heuer, Richards J. Jr. (1999) Psychology of Intelligence Analysis. Central Intelligence Agency. http://www.au.af.mil/au/awc/awcgate/psych-intel/art5.html
  • Kahneman D., Slovic P., and Tversky, A. (Eds.) (1982) Judgment Under Uncertainty: Heuristics and Biases. New York: Cambridge University Press ISBN 978-0521284141
  • Massimo Piatelli-Palmarini (1994). Inevitable Illusions: How Mistakes of Reason Rule Our Minds New York: John Wiley & Sons. ISBN 0-471-15962-X
  • Nisbett, R., and Ross, L. (1980) Human Inference: Strategies and shortcomings of human judgement. Englewood Cliffs, NJ: Prentice-Hall ISBN 978-0134451305
  • Stuart Sutherland (2007). Irrationality: The Enemy Within Second Edition (First Edition 1994) Pinter & Martin. ISBN 978-1905177073
  • Tavris, Carol and Elliot Aronson (2007). Mistakes Were Made (But Not by Me): Why We Justify Foolish Beliefs, Bad Decisions and Hurtful Acts Orlando, Florida: Harcourt Books. ISBN 978-0-15-101098-1

External links


This page uses Creative Commons Licensed content from Wikipedia (view authors).