Content deleted Content added
No edit summary |
m Open access bot: url-access updated in citation with #oabot. |
||
(48 intermediate revisions by 26 users not shown) | |||
Line 1:
{{Short description|Theory of human visual attention}}
==Stages==
According to Treisman,
The second stage of feature integration theory is the focused attention stage, where a subject combines individual features of an object to perceive the whole object. Combining individual features of an object requires attention, and selecting that object occurs within a "master map" of locations. The master map of locations contains all the locations in which features have been detected, with each ___location in the master map having access to the multiple feature maps. These multiple feature maps, or sub-maps, contain a large storage base of features. Features such as color, shape, orientation, sound, and movement are stored in these sub-maps <ref>{{Cite journal |last1=Kristjánsson |first1=Árni |last2=Egeth |first2=Howard |date=2020-01-01 |title=How feature integration theory integrated cognitive psychology, neurophysiology, and psychophysics |journal=Attention, Perception, & Psychophysics |language=en |volume=82 |issue=1 |pages=7–23 |doi=10.3758/s13414-019-01803-7 |issn=1943-393X|doi-access=free |pmid=31290134 }}</ref><ref>{{Cite journal |last1=Chan |first1=Louis K. H. |last2=Hayward |first2=William G. |date=2009 |title=Feature integration theory revisited: Dissociating feature detection and attentional guidance in visual search. |url=http://doi.apa.org/getdoi.cfm?doi=10.1037/0096-1523.35.1.119 |journal=Journal of Experimental Psychology: Human Perception and Performance |language=en |volume=35 |issue=1 |pages=119–132 |doi=10.1037/0096-1523.35.1.119 |pmid=19170475 |issn=1939-1277|url-access=subscription }}</ref>.When attention is focused at a particular ___location on the map, the features currently in that position are attended to and are stored in "object files". If the object is familiar, associations are made between the object and prior knowledge, which results in identification of that object. This top-down process, using prior knowledge to inform a current situation or decision, is paramount in either identifying or recognizing objects.<ref>{{Cite book |last1=Nobre |first1=Kia |url=https://books.google.com/books?id=mtXQAgAAQBAJ |title=The Oxford Handbook of Attention |last2=Kastner |first2=Sabine |date=2014 |publisher=OUP Oxford |isbn=978-0-19-967511-1 |language=en}}</ref><ref>{{Cite journal |last1=Chan |first1=Louis K. H. |last2=Hayward |first2=William G. |date=2009 |title=Feature integration theory revisited: Dissociating feature detection and attentional guidance in visual search. |url=http://doi.apa.org/getdoi.cfm?doi=10.1037/0096-1523.35.1.119 |journal=Journal of Experimental Psychology: Human Perception and Performance |language=en |volume=35 |issue=1 |pages=119–132 |doi=10.1037/0096-1523.35.1.119 |pmid=19170475 |issn=1939-1277|url-access=subscription }}</ref> In support of this stage, researchers often refer to patients with [[Balint's syndrome]]. Due to damage in the parietal lobe, these people are unable to focus attention on individual objects. Given a stimulus that requires combining features, people with Balint's syndrome are unable to focus attention long enough to combine the features, providing support for this stage of the theory.<ref>{{Cite journal|last1=Cohen|first1=Asher|last2=Rafal|first2=Robert D.|date=1991|title=Attention and Feature Integration: Illusory Conjunctions in a Patient with a Parietal Lobe Lesion|url=http://www.jstor.org/stable/40062648|journal=Psychological Science|volume=2|issue=2|pages=106–110|doi=10.1111/j.1467-9280.1991.tb00109.x |jstor=40062648 |s2cid=145171384 |issn=0956-7976|url-access=subscription}}</ref>
[[File:FITstages.png|alt=The stages of
Treisman distinguishes between two kinds of visual search tasks, "feature search" and "conjunction search". Feature searches can be performed fast and pre-attentively for targets defined by only one
As a reaction to the feature integration theory, Wolfe (1994) proposed the Guided Search Model 2.0. According to this model, attention is directed to an object or ___location through a preattentive process. The preattentive process, as Wolfe explains, directs attention in both a bottom-up and top-down way. Information acquired through both bottom-up and top-down processing is ranked according to priority. The priority ranking ''guides'' visual search and makes the search more efficient. Whether the Guided Search Model 2.0 or the feature integration theory are "correct" theories of visual search is still a hotly debated topic.
▲Treisman distinguishes between two kinds of visual search tasks, "feature search" and "conjunction search". Feature searches can be performed fast and pre-attentively for targets defined by only one primitive feature, such as color, shape, movement or orientation. Conjunction searches proceed serially for targets defined by a conjunction (two or more) of primitive features - it is much slower and requires conscious attention. She concluded from many experiments that [[color]], [[Orientation (geometry)|orientation]], and [[intensity (disambiguation)|intensity]] are primitive features, for which feature searches may be performed.
==Experiments==
<!-- Deleted image removed: [[File:fourshapesexp.png|thumb|alt=An example of four colored shapes and two black letters.|An example of the stimuli found in Treisman et al. (1982).]] -->
Research participant R.M., a [[Bálint's syndrome]] sufferer who is unable to focus attention on individual objects, experiences illusory conjunctions when presented with simple stimuli such as a "blue O" or a "red T." For 23% of trials, even when able to view the stimulus for as long as 10 seconds, R.M. reported seeing a "red O" or a "blue T" <ref> Friedman-Hill et al., 1995; Robertson et al., 1997. </ref>. This finding is in accordance with feature integration theory's prediction of how one with a lack of focused attention would erroneously combine features.▼
▲As previously mentioned, Balint's syndrome patients have provided support for the feature integration theory. Particularly, Research participant R.M.,
[[File:treismanshapes.png|thumb|alt=The stimuli resembling a carrot, lake and tire, respectively.|The stimuli resembling a carrot, lake and tire, respectively. Treisman et al.(1986).]]However, if people use their knowledge or experience to perceive an object, it is less likely to make mistakes. Treisman did another experiment to explain this phenomenon – she presented three shapes to participants and illusory conjunctions persisted. Surprisingly, when she told participants that they were being shown a carrot, lake and tire (in place of the orange triangle, blue oval, and black circle, respectively), illusory conjunctions disappeared <ref>Illusory words: The roles of attention and of top–down constraints in conjoining letters to form words.▼
By Treisman, Anne; Souther, Janet. Journal of Experimental Psychology: Human Perception and Performance, Vol 12(1), Feb 1986, 3-17.</ref>. Treisman maintained that prior-knowledge played an important role in proper perception. Normally, bottom-up processing is used for identifying novel objects; but, once we recall prior knowledge, top-down processing is used. This explains why people are good at identifying familiar objects rather than unfamiliar.▼
[[File:treismanshapes.png|thumb|alt=The stimuli resembling a carrot, lake and tire, respectively.|The stimuli resembling a carrot, lake and tire, respectively. Treisman et al.(1986).]]
==Reading==▼
When identifying letters while reading, not only are there shapes picked up but also other features like their colors and surrounding elements. Individual letters are processed serially when spatially conjoined with another letter. The locations of each feature of a letter are not known in advance, even while the letter is in front of the reader. Since the ___location of the letter's features and/or the ___location of the letter is unknown, feature interchanges can occur if one is not attentively focused. This is known as [[lateral masking]], which in this case, refers to a difficulty in separating a letter from the background <ref>Anne Treisman and Garry Gelade (1980). "A feature-integration theory of attention." Cognitive Psychology, Vol. 12, No. 1, pp. 97-136.</ref>.▼
▲
==References==▼
▲By Treisman, Anne; Souther, Janet. Journal of Experimental Psychology: Human Perception and Performance, Vol 12(1), Feb 1986, 3-17.</ref>
* Anne Treisman and [[Garry Gelade]] (1980). "A feature-integration theory of attention." ''Cognitive Psychology'', Vol. 12, No. 1, pp. 97-136.▼
* Anne Treisman (1988). "Features and objects: the fourteenth Bartlett Memorial Lecture." ''Quarterly Journal of Experimental Psychology'', 40A, pp. 201-236.▼
*Anne Treisman and [[Nancy Kanwisher]] (1998). "Perceiving visually presented objects: recognition, awareness, and modularity." ''Current Opinion in Neurobiology'', '''8''', pp. 218-226.▼
▲==Reading==
; Notes▼
▲When identifying letters while reading, not only are
<references/>▼
==See also==
* [[Attention]]▼
* [[Binding problem]]
* [[Visual search]]
== External links and references ==▼
▲<references/>
▲==References==
▲* Anne Treisman and
* Anne Treisman and Hilary Schmidt (1982). "Illusory conjunctions in the perception of objects." ''Cognitive Psychology'', '''14''', pp. 107–141.
* Anne Treisman and Janet Souther (1986). "Illusory words: The roles of attention and of top–down constraints in conjoining letters to form words." ''Journal of Experimental Psychology: Human Perception and Performance'', '''12''' (1), pp. 3–17
▲* Anne Treisman (1988). "Features and objects: the fourteenth Bartlett Memorial Lecture." ''Quarterly Journal of Experimental Psychology'', '''40A''', pp.
▲*Anne Treisman and [[Nancy Kanwisher]] (1998). "Perceiving visually presented objects: recognition, awareness, and modularity." ''Current Opinion in Neurobiology'', '''8''', pp.
* J. M. Wolfe (1994). "Guided Search 2.0: A revised model of visual search." ''Psychonomic Bulletin & Review'', '''1''', pp. 202–238
* [http://web.mit.edu/bcs/nklab/media/pdfs/TreismanKanwisherCurrOpBio98.pdf 1998 paper by Treisman and Kanwisher at web.mit.edu]
Line 42 ⟶ 48:
[[Category:Cognition]]
[[Category:Human–computer interaction]]
|