Heuristic

From Wikipedia, the free encyclopedia
Jump to navigation Jump to search

A heuristic technique (/hjʊəˈrɪstɪk/; Ancient Greek: εὑρίσκω, "find" or "discover"), or a heuristic, is any approach to problem solving or self-discovery that employs a practical method that is not guaranteed to be optimal, perfect or rational, but which is nevertheless sufficient for reaching an immediate, short-term goal. Where finding an optimal solution is impossible or impractical, heuristic methods can be used to speed up the process of finding a satisfactory solution. Heuristics can be mental shortcuts that ease the cognitive load of making a decision.[1]:94[2] Examples that employ heuristics include using trial and error, a rule of thumb, an educated guess, an intuitive judgment, a guesstimate, profiling, or common sense.

Overview[edit]

Heuristics are the strategies derived from previous experiences with similar problems. These strategies depend on using readily accessible, though loosely applicable, information to control problem solving in human beings, machines and abstract issues.[3][4]

The most fundamental heuristic is trial and error, which can be used in everything from matching nuts and bolts to finding the values of variables in algebra problems. In mathematics, some common heuristics involve the use of visual representations, additional assumptions, forward/backward reasoning and simplification.[5] Here are a few commonly used heuristics from George Pólya's 1945 book, How to Solve It:[6]

  • If you are having difficulty understanding a problem, try drawing a picture.
  • If you can't find a solution, try assuming that you have a solution and seeing what you can derive from that ("working backward").
  • If the problem is abstract, try examining a concrete example.
  • Try solving a more general problem first (the "inventor's paradox": the more ambitious plan may have more chances of success).

In psychology, heuristics are simple, efficient rules, learned or inculcated by evolutionary processes, that have been proposed to explain how people make decisions, come to judgments, and solve problems typically when facing complex problems or incomplete information. Researchers test if people use those rules with various methods. These rules work well under most circumstances, but in certain cases can lead to systematic errors or cognitive biases.[7]

History[edit]

The study of heuristics in human decision-making was developed in the 1970s and 80s by psychologists Amos Tversky and Daniel Kahneman,[8] although the concept was originally introduced by Nobel laureate Herbert A. Simon. Simon's original, primary object of research was problem solving which showed that we operate within what he calls bounded rationality. He coined the term "satisficing", which denotes the situation where people seek solutions or accept choices or judgments that are "good enough" for their purposes, but which could be optimized.[9]

Rudolf Groner analyzed the history of heuristics from its roots in ancient Greece up to contemporary work in cognitive psychology and artificial intelligence,[10] and proposed a cognitive style "heuristic versus algorithmic thinking" which can be assessed by means of a validated questionnaire.[11]

Gerd Gigerenzer and his research group argued that models of heuristics need to be formal to allow for predictions of behavior that can be tested.[12] They study the fast and frugal heuristics in the "adaptive toolbox" of individuals or institutions, and the ecological rationality of these heuristics, that is, the conditions under which a given heuristic is likely to be successful.[13] The descriptive study of the "adaptive toolbox" is done by observation and experiment, the prescriptive study of the ecological rationality requires mathematical analysis and computer simulation. Heuristics – such as the recognition heuristic, the take-the-best heuristic, and fast-and-frugal trees – have been shown to be effective in predictions, particularly in situations of uncertainty. It is often said that heuristics trade accuracy for effort but this is only the case in situations of risk. Risk refers to situations where all possible actions, their outcomes and probabilities are known. In the absence of this information, that is under uncertainty, heuristics can achieve higher accuracy with lower effort.[14] This finding, known as a less-is-more effect, would not have been found without formal models. The valuable insight of this program is that heuristics are effective not despite of their simplicity — but because of it. Furthermore, Gigerenzer and Wolfgang Gaissmaier found that both individuals and organizations rely on heuristics in an adaptive way.[15]

Heuristics, through greater refinement and research, have begun to be applied to other theories, or be explained by them. For example: the cognitive-experiential self-theory (CEST) also is an adaptive view of heuristic processing. CEST breaks down two systems that process information. At some times, roughly speaking, individuals consider issues rationally, systematically, logically, deliberately, effortfully, and verbally. On other occasions, individuals consider issues intuitively, effortlessly, globally, and emotionally.[16] From this perspective, heuristics are part of a larger experiential processing system that is often adaptive, but vulnerable to error in situations that require logical analysis.[17]

In 2002, Daniel Kahneman and Shane Frederick proposed that cognitive heuristics work by a process called attribute substitution, which happens without conscious awareness.[18] According to this theory, when somebody makes a judgment (of a "target attribute") that is computationally complex, a rather easier calculated "heuristic attribute" is substituted. In effect, a cognitively difficult problem is dealt with by answering a rather simpler problem, without being aware of this happening.[18] This theory explains cases where judgments fail to show regression toward the mean.[19] Heuristics can be considered to reduce the complexity of clinical judgments in health care.[20]

Psychology[edit]

Informal models of heuristics[edit]

  • Affect heuristic
  • Anchoring and adjustment – Describes the common human tendency to rely more heavily on the first piece of information offered (the "anchor") when making decisions. For example, in a study done with children, the children were told to estimate the number of jellybeans in a jar. Groups of children were given either a high or low "base" number (anchor). Children estimated the number of jellybeans to be closer to the anchor number that they were given.[21]
  • Availability heuristic – A mental shortcut that occurs when people make judgments about the probability of events by the ease with which examples come to mind. For example, in a 1973 Tversky & Kahneman experiment, the majority of participants reported that there were more words in the English language that start with the letter K than for which K was the third letter. There are actually twice as many words in the English Language that have K as the third letter as those that start with K, but words that start with K are much easier to recall and bring to mind.[22]
  • Contagion heuristic
  • Effort heuristic
  • Escalation of commitment – Describes the phenomenon where people justify increased investment in a decision, based on the cumulative prior investment, despite new evidence suggesting that the cost, starting today, of continuing the decision outweighs the expected benefit. This is related to the sunk cost fallacy.
  • Familiarity heuristic – A mental shortcut applied to various situations in which individuals assume that the circumstances underlying the past behavior still hold true for the present situation and that the past behavior thus can be correctly applied to the new situation. Especially prevalent when the individual experiences a high cognitive load.
  • Naïve diversification – When asked to make several choices at once, people tend to diversify more than when making the same type of decision sequentially.
  • Peak–end rule
  • Representativeness heuristic – A mental shortcut used when making judgments about the probability of an event under uncertainty. Or, judging a situation based on how similar the prospects are to the prototypes the person holds in his or her mind. For example, in a 1982 Tversky and Kahneman experiment, participants were given a description of a woman named Linda. Based on the description, it was likely that Linda was a feminist. Eighty to ninety percent of participants, choosing from two options, chose that it was more likely for Linda to be a feminist and a bank teller than only a bank teller. The likelihood of two events cannot be greater than that of either of the two events individually. For this reason, the representativeness heuristic is exemplary of the conjunction fallacy.[22]
  • Scarcity heuristic
  • Simulation heuristic
  • Social proof

Formal models of heuristics[edit]

Cognitive maps[edit]

Heuristics were also found to be used in the manipulation and creation of cognitive maps. Cognitive maps are internal representations of our physical environment, particularly associated with spatial relationships. These internal representations of our environment are used by our memory as a guide in our external environment. It was found that when questioned about maps imaging, distancing, etc., people commonly made distortions to images. These distortions took shape in the regularization of images (i.e., images are represented as more like pure abstract geometric images, though they are irregular in shape).

There are several ways that humans form and use cognitive maps. Visual intake is a key part of mapping. The first is by using landmarks. This is where a person uses a mental image to estimate a relationship, usually distance, between two objects. The second is route-road knowledge, and is generally developed after a person has performed a task and is relaying the information of that task to another person. The third is a survey. A person estimates a distance based on a mental image that, to them, might appear like an actual map. This image is generally created when a person's brain begins making image corrections. These are presented in five ways:

  1. Right-angle bias: when a person straightens out an image, like mapping an intersection, and begins to give everything 90-degree angles, when in reality it may not be that way.
  2. Symmetry heuristic: when people tend to think of shapes, or buildings, as being more symmetrical than they really are.
  3. Rotation heuristic: when a person takes a naturally (realistically) distorted image and straightens it out for their mental image.
  4. Alignment heuristic: similar to the previous, where people align objects mentally to make them straighter than they really are.
  5. Relative-position heuristic: people do not accurately distance landmarks in their mental image based on how well they remember that particular item.

Another method of creating cognitive maps is by means of auditory intake based on verbal descriptions. Using the mapping based from a person's visual intake, another person can create a mental image, such as directions to a certain location.[23]

Philosophy[edit]

"Heuristic device" is used when an entity X exists to enable understanding of, or knowledge concerning, some other entity Y. A good example is a model that, as it is never identical with what it models, is a heuristic device to enable understanding of what it models. Stories, metaphors, etc., can also be termed heuristic in that sense. A classic example is the notion of utopia as described in Plato's best-known work, The Republic. This means that the "ideal city" as depicted in The Republic is not given as something to be pursued, or to present an orientation-point for development; rather, it shows how things would have to be connected, and how one thing would lead to another (often with highly problematic results), if one opted for certain principles and carried them through rigorously.

"Heuristic" is also often used as a noun to describe a rule-of-thumb, procedure, or method.[24] Philosophers of science have emphasized the importance of heuristics in creative thought and the construction of scientific theories.[25] (See The Logic of Scientific Discovery by Karl Popper; and philosophers such as Imre Lakatos,[26] Lindley Darden, William C. Wimsatt, and others.)

Law[edit]

In legal theory, especially in the theory of law and economics, heuristics are used in the law when case-by-case analysis would be impractical, insofar as "practicality" is defined by the interests of a governing body.[27]

The present securities regulation regime largely assumes that all investors act as perfectly rational persons. In truth, actual investors face cognitive limitations from biases, heuristics, and framing effects.

For instance, in all states in the United States the legal drinking age for unsupervised persons is 21 years, because it is argued that people need to be mature enough to make decisions involving the risks of alcohol consumption. However, assuming people mature at different rates, the specific age of 21 would be too late for some and too early for others. In this case, the somewhat arbitrary deadline is used because it is impossible or impractical to tell whether an individual is sufficiently mature for society to trust them with that kind of responsibility. Some proposed changes, however, have included the completion of an alcohol education course rather than the attainment of 21 years of age as the criterion for legal alcohol possession. This would put youth alcohol policy more on a case-by-case basis and less on a heuristic one, since the completion of such a course would presumably be voluntary and not uniform across the population.

The same reasoning applies to patent law. Patents are justified on the grounds that inventors must be protected so they have incentive to invent. It is therefore argued that it is in society's best interest that inventors receive a temporary government-granted monopoly on their idea, so that they can recoup investment costs and make economic profit for a limited period. In the United States, the length of this temporary monopoly is 20 years from the date the patent application was filed, though the monopoly does not actually begin until the application has matured into a patent. However, like the drinking-age problem above, the specific length of time would need to be different for every product to be efficient. A 20-year term is used because it is difficult to tell what the number should be for any individual patent. More recently, some, including University of North Dakota law professor Eric E. Johnson, have argued that patents in different kinds of industries – such as software patents – should be protected for different lengths of time.[28]

Stereotyping[edit]

Stereotyping is a type of heuristic that people use to form opinions or make judgments about things they have never seen or experienced.[29] They work as a mental shortcut to assess everything from the social status of a person (based on their actions)[2], to whether a plant is a tree based on the assumption that it is tall, has a trunk, and has leaves (even though the person making the evaluation might never have seen that particular type of tree before).

Stereotypes, as first described by journalist Walter Lippmann in his book Public Opinion (1922), are the pictures we have in our heads that are built around experiences as well as what we are told about the world.[30][31]

Artificial intelligence[edit]

A heuristic can be used in artificial intelligence systems while searching a solution space. The heuristic is derived by using some function that is put into the system by the designer, or by adjusting the weight of branches based on how likely each branch is to lead to a goal node.

Critiques and controversies[edit]

The concept of heuristics has critiques and controversies. The "We Cannot Be That Dumb" critique argues that the average person has low ability to make sound and effective judgments.[32]

See also[edit]

References[edit]

  1. ^ Myers, David G. (2010). Social psychology (Tenth ed.). New York, NY. ISBN 9780073370668. OCLC 667213323.
  2. ^ a b "Heuristics - Explanation and examples". Conceptually. Retrieved 2019-10-23.
  3. ^ Pearl, Judea (1983). Heuristics: Intelligent Search Strategies for Computer Problem Solving. New York, Addison-Wesley, p. vii. ISBN 978-0-201-05594-8
  4. ^ Emiliano, Ippoliti (2015). Heuristic Reasoning: Studies in Applied Philosophy, Epistemology and Rational Ethics. Switzerland: Springer International Publishing. pp. 1–2. ISBN 978-3-319-09159-4.
  5. ^ "The Definitive Glossary of Higher Mathematical Jargon — Heuristics". Math Vault. 2019-08-01. Retrieved 2019-10-23.
  6. ^ Pólya, George (1945) How to Solve It: A New Aspect of Mathematical Method, Princeton, NJ: Princeton University Press. ISBN 0-691-02356-5   ISBN 0-691-08097-6
  7. ^ Gigerenzer, Gerd (1991). "How to Make Cognitive Illusions Disappear: Beyond "Heuristics and Biases"" (PDF). European Review of Social Psychology. 2: 83–115. CiteSeerX 10.1.1.336.9826. doi:10.1080/14792779143000033. Retrieved 14 October 2012.
  8. ^ Daniel Kahneman, Amos Tversky, and Paul Slovic, eds. (1982) Judgment under Uncertainty: Heuristics & Biases. Cambridge, UK, Cambridge University Press ISBN 0-521-28414-7
  9. ^ Heuristics and heuristic evaluation. Interaction-design.org. Retrieved 2013-09-01.
  10. ^ Rudolf Groner, Marina Groner & Walter F. Bischof (1983). Methods of heuristics. Hillsdale N.J., Lawrence Erlbaum.
  11. ^ Rudolf Groner & Marina Groner (1991). Heuristische versus algorithmische Orientierung als Dimension des individuellen kognitiven Stils. In K. Grawe, N. Semmer, R. Hänni (Hrsg.), Über die richtige Art, Psychologie zu betreiben . Göttingen, Hogrefe.
  12. ^ Gerd Gigerenzer, Peter M. Todd, and the ABC Research Group (1999). Simple Heuristics That Make Us Smart. Oxford, UK, Oxford University Press. ISBN 0-19-514381-7
  13. ^ Gigerenzer, Gerd, and Reinhard Selten, eds. (2002) Bounded rationality: The adaptive toolbox. Cambridge, Massachusetts, MIT press. ISBN 978-0262571647
  14. ^ Gigerenzer, Gerd; Hertwig, Ralph; Pachur, Thorsten (2011-04-15). Heuristics: The Foundations of Adaptive Behavior. Oxford University Press. doi:10.1093/acprof:oso/9780199744282.001.0001. hdl:11858/00-001M-0000-0024-F172-8. ISBN 9780199894727.
  15. ^ Gigerenzer, Gerd; Gaissmaier, Wolfgang (January 2011). "Heuristic Decision Making". Annual Review of Psychology. 62: 451–482. doi:10.1146/annurev-psych-120709-145346. hdl:11858/00-001M-0000-0024-F16D-5. PMID 21126183. SSRN 1722019.
  16. ^ De Neys, Wim (2008-10-18). "Cognitive experiential self theory - Psychlopedia". Perspectives on Psychological Science. 7 (1): 28–38. doi:10.1177/1745691611429354. PMID 26168420. Archived from the original on 2013-07-31. Retrieved 2013-09-01.
  17. ^ Epstein, S.; Pacini, R.; Denes-Raj, V.; Heier, H. (1996). "Individual differences in intuitive-experiential and analytical-rational thinking styles". Journal of Personality and Social Psychology. 71 (2): 390–405. doi:10.1037/0022-3514.71.2.390. PMID 8765488.
  18. ^ a b Kahneman, Daniel; Shane Frederick (2002). "Representativeness Revisited: Attribute Substitution in Intuitive Judgment". In Thomas Gilovich; Dale Griffin; Daniel Kahneman (eds.). Heuristics and Biases: The Psychology of Intuitive Judgment. Cambridge: Cambridge University Press. pp. 49–81. ISBN 978-0-521-79679-8. OCLC 47364085.
  19. ^ Kahneman, Daniel (December 2003). "Maps of Bounded Rationality: Psychology for Behavioral Economics" (PDF). American Economic Review. 93 (5): 1449–1475. CiteSeerX 10.1.1.194.6554. doi:10.1257/000282803322655392. ISSN 0002-8282. Archived from the original (PDF) on 2018-02-19. Retrieved 2017-11-01.
  20. ^ Cioffi, Jane (1997). "Heuristics, servants to intuition, in clinical decision making". Journal of Advanced Nursing. 26: 203–208. doi:10.1046/j.1365-2648.1997.1997026203.x. PMID 9231296.
  21. ^ Smith, H. (1999). "Use of the anchoring and adjustment heuristic by children". Current Psychology. 18 (3): 294–300. doi:10.1007/s12144-999-1004-4.
  22. ^ a b Harvey, N (2007). "Use of heuristics: Insights from forecasting research". Thinking & Reasoning. 13 (1): 5–24. doi:10.1080/13546780600872502.
  23. ^ Sternberg, Robert J.; Karin Sternberg (2012). Cognitive Psychology (6th ed.). Belmont, CA: Wadsworth, Cengage Learning. pp. 310–1315. ISBN 978-1-111-34476-4.
  24. ^ K. M. Jaszczolt (2006). "Defaults in Semantics and Pragmatics", The Stanford Encyclopedia of Philosophy, ISSN 1095-5054
  25. ^ Roman Frigg and Stephan Hartmann (2006). "Models in Science", The Stanford Encyclopedia of Philosophy, ISSN 1095-5054
  26. ^ Kiss, Olga (2006). "Heuristic, Methodology or Logic of Discovery? Lakatos on Patterns of Thinking". Perspectives on Science. 14 (3): 302–317. doi:10.1162/posc.2006.14.3.302.
  27. ^ Gerd Gigerenzer and Christoph Engel, eds. (2007). Heuristics and the Law, Cambridge, The MIT Press, ISBN 978-0-262-07275-5
  28. ^ Johnson, Eric E. (2006). "Calibrating Patent Lifetimes" (PDF). Santa Clara Computer & High Technology Law Journal. 22: 269–314.
  29. ^ Bodenhausen, Galen V.; et al. (1999). "On the Dialectics of Discrimination: Dual Processes in Social Stereotyping", in Dual-process Theories in Social Psychology edited by Shelly Chaiken and Yaacov Trope. NY: Guilford Press. pp. 271–92. ISBN 978-1572304215. Retrieved 24 March 2015.
  30. ^ Kleg, Milton (1883). Hate Prejudice and Racism. Albany: State University of New York Press. p. 135. ISBN 978-0791415368. Retrieved 24 March 2015.
  31. ^ Gökçen, Sinan. "Pictures in Our Heads". European Roma Rights Centre. Retrieved 24 March 2015.
  32. ^ Gilovich, Thomas; Griffin, Dale; Kahnemann, Daniel, eds. (2002). Heuristics and Biases: The Psychology of Intuitive Judgment. Cambridge University Press. pp. 8–9.

Further reading[edit]