Jump to content

Visual short-term memory

From Wikipedia, the free encyclopedia

In the study of vision, visual short-term memory (VSTM) is one of three broad memory systems including iconic memory and long-term memory. VSTM is a type of short-term memory, but one limited to information within the visual domain.

The term VSTM refers in a theory-neutral manner to the non-permanent storage of visual information over an extended period of time.[1] The visuospatial sketchpad is a VSTM subcomponent within the theoretical model of working memory proposed by Alan Baddeley; in which it is argued that a working memory aids in mental tasks like planning and comparison.[2][3] Whereas iconic memories are fragile, decay rapidly, and are unable to be actively maintained, visual short-term memories are robust to subsequent stimuli and last over many seconds. VSTM is distinguished from long-term memory, on the other hand, primarily by its very limited capacity.[4][5]

Overview

[edit]

The introduction of stimuli which were hard to verbalize, and unlikely to be held in long-term memory, revolutionized the study of VSTM in the early 1970s.[6][7][8] The basic experimental technique used required observers to indicate whether two matrices,[7][8] or figures,[6] separated by a short temporal interval, were the same. The finding that observers were able to report that a change had occurred, at levels significantly above chance, indicated that they were able to encode aspect of the first stimulus in a purely visual store, at least for the period until the presentation of the second stimulus. However, as the stimuli used were complex, and the nature of the change relatively uncontrolled, these experiments left open various questions, such as:

  1. whether only a subset of the perceptual dimensions comprising a visual stimulus are stored (e.g., spatial frequency, luminance, or contrast)
  2. whether perceptual dimensions are maintained in VSTM with greater fidelity than others
  3. the nature by which these dimensions are encoded (i.e., are perceptual dimensions encoded within separate, parallel channels, or are all perceptual dimensions stored as a single bound entity within VSTM?).

Set-size effects

[edit]

Much effort has been dedicated to investigating the capacity limits of VSTM. In a typical change-detection task, observers are presented with two arrays, composed of a number of stimuli. The two arrays are separated by a short temporal interval, and the task of observers is to decide if the first and second arrays are identical, or whether one item differs across the two displays.[a] Performance is critically dependent on the number of items in the array. While performance is generally almost perfect for arrays of one or two items, correct responses invariably decline in a monotonic fashion as more items are added. Different theoretical models have been put forward to explain limits on VSTM storage, and distinguishing between them remains an active area of research.

Models of capacity limits

[edit]

Slot models

[edit]

A prominent class of model proposes that observers are limited by the total number of items which can be encoded, either because the capacity of VSTM itself is limited.[b] This type of model has obvious similarities to urn models used in probability theory.[c] In essence, an urn model assumes that VSTM is restricted in storage capacity to only a few items, k (often estimated to lie in the range of three-to-five in adults, though fewer in children[9]). The probability that a suprathreshold change will be detected is simply the probability that the change element is encoded in VSTM (i.e., k/N). This capacity limit has been linked to the posterior parietal cortex, the activity of which initially increases with the number of stimuli in the arrays, but saturates at higher set-sizes.[10] Although urn models are used commonly to describe performance limitations in VSTM,[d] it is only recently that the actual structure of items stored has been considered. Luck and colleagues have reported a series of experiments designed specifically to elucidate the structure of information held in VSTM.[11] This work provides evidence that items stored in VSTM are coherent objects, and not the more elementary features of which those objects are composed.

Noise models

[edit]

An alternative framework has more been put forward by Wilken and Ma who suggest that apparent capacity limitations in VSTM are caused by a monotonic decline in the quality of the internal representations stored (i.e., monotonic increase in noise) as a function of set-size. In this conception capacity limitations in memory are not caused by a limit on the number of things that can be encoded, but by a decline in the quality of the representation of each thing as more things are added to memory. In their 2004 experiments, they varied color, spatial frequency, and orientation of objects stored in VSTM using a signal detection theory approach.[e] The participants were asked to report differences between the visual stimuli presented to them in consecutive order. The investigators found that different stimuli were encoded independently and in parallel, and that the major factor limiting report performance was neuronal noise (which is a function of visual set-size).[12]

Under this framework, the key limiting factor on working memory performance is the precision with which visual information can be stored, not the number of items that can be remembered.[12] Further evidence for this theory was obtained by Bays and Husain using a discrimination task. They showed that, unlike a "slot" model of VSTM, a signal-detection model could account both for discrimination performance in their study and previous results from change detection tasks.[f] These authors proposed that VSTM is a flexible resource, shared out between elements of a visual scene—items that receive more resource are stored with greater precision. In support of this, they showed that increasing the salience of one item in a memory array led to that item being recalled with increased resolution, but at the cost of reducing resolution of storage for the other items in the display.[13]

Psychophysical models

[edit]

Psychophysical experiments suggest that information is encoded in VSTM across multiple parallel channels, each channel associated with a particular perceptual attribute.[14] Within this framework, a decrease in an observer's ability to detect a change with increasing set-size can be attributed to two different processes:

  1. if decisions are made across different channels, decreases in performance are typically small, and consistent with decreases expected when making multiple independent decisions[15][16]
  2. if multiple decisions are made within the same channel, the decrease in performance is much greater than expected on the basis of increased decision-noise alone, and is attributed to interference caused by multiple decisions within the same perceptual channel.[17]

However, the Greenlee-Thomas model[15] suffers from two failings as a model for the effects of set-size in VSTM. First, it has only been empirically tested with displays composed of one or two elements. It has been shown repeatedly in various experimental paradigms that set-size effects differ for displays composed of a relatively small number of elements (i.e., 4 items or less), and those associated with larger displays (i.e., more than 4 items). The Greenlee-Thomas model offers no explanation for why this might be so. Second, while Magnussen, Greenlee, and Thomas[18][full citation needed] are able to use this model to predict that greater interference will be found when dual decisions are made within the same perceptual dimension, rather than across different perceptual dimensions, this prediction lacks quantitative rigor, and is unable to accurately anticipate the size of the threshold increase, or give a detailed explanation of its underlying causes.

In addition to the Greenlee-Thomas model, there are two other prominent approaches for describing set-size effects in VSTM. These two approaches can be referred to as sample size models,[19] and urn models.[g] They differ from the Greenlee-Thomas model by:

  1. ascribing the root cause of set-size effects to a stage prior to decision making
  2. making no theoretical distinction between decisions made in the same, or across different, perceptual dimensions.

Intermediate visual store

[edit]

There is some evidence of an intermediate visual store with characteristics of both iconic memory and VSTM.[20] This intermediate store is proposed to have high capacity (up to 15 items) and prolonged memory trace duration (up to 4 seconds). It coexists with VSTM but unlike it visual stimuli can overwrite the contents of its visual store.[21] Further studies suggests an involvement of visual area V4 in the retention of information about the color of the stimulus in visual working memory,[22][23] and the role of the VO1 area for retaining information about its shape.[23] It has been shown that in the VO2 region all characteristics of the stimulus retained in memory are combined into a holistic image.[23]

The function of visual short-term memory representations

[edit]

VSTM is thought[by whom?] to be the visual component of the working memory system, and as such it is used as a buffer for temporary information storage during the process of naturally occurring tasks. But what naturally occurring tasks actually require VSTM? Most work on this issue has focused on the role of VSTM in bridging the sensory gaps caused by saccadic eye movements. These sudden shift of gaze typically occur 2–4 times per second, and vision is briefly suppressed while the eyes are moving. Thus, the visual input consists of a series of spatially shifted snapshots of the overall scene, separated by brief gaps. Over time, a rich and detailed long-term memory representation is constructed from these brief glimpses of the input, and VSTM is thought[by whom?] to bridge the gaps between these glimpses and to allow the relevant portions of one glimpse to be aligned with the relevant portions of the next glimpse. Both spatial and object VSTM systems may play important roles in the integration of information across eye movements. Eye movements are also affected by VSTM representations. The constructed representations held in VSTM can affect eye movements even when the task does not explicitly require eye movements: the direction of small microsaccades point towards the location of objects in VSTM.[24]

See also

[edit]

Notes

[edit]
  1. ^ e.g., Luck & Vogel 1997.
  2. ^ e.g., Cowan 2001; Luck & Vogel 1997; Pashler 1988.
  3. ^ See, for example, Mendenhall 1967.[full citation needed]
  4. ^ e.g., Luck & Vogel 1997; Pashler 1988; Sperling 1960.
  5. ^ See also the closely related work by Palmer 1990.
  6. ^ e.g., Luck & Vogel 1997.
  7. ^ e.g., Pashler 1988.

References

[edit]
  1. ^ Buss, Aaron T.; Ross-Sheehy, Shannon; Reynolds, Greg D. (2018-10-01). "Visual working memory in early development: a developmental cognitive neuroscience perspective". Journal of Neurophysiology. 120 (4): 1472–1483. doi:10.1152/jn.00087.2018. ISSN 0022-3077. PMID 29897858. S2CID 49189631.
  2. ^ Buss, Aaron T.; Ross-Sheehy, Shannon; Reynolds, Greg D. (2018-10-01). "Visual working memory in early development: a developmental cognitive neuroscience perspective". Journal of Neurophysiology. 120 (4): 1472–1483. doi:10.1152/jn.00087.2018. ISSN 0022-3077. PMID 29897858. S2CID 49189631.
  3. ^ Logie, Robert (1988-04-01). "Working memory, Alan Baddeley, Oxford University Press, Oxford 1986. No. of pages: 289. Price £30.00 (Hardback), ISBN 0 19 852116 2". Applied Cognitive Psychology. 2 (2): 166–168. doi:10.1002/acp.2350020209.
  4. ^ Buss, Aaron T.; Ross-Sheehy, Shannon; Reynolds, Greg D. (2018-10-01). "Visual working memory in early development: a developmental cognitive neuroscience perspective". Journal of Neurophysiology. 120 (4): 1472–1483. doi:10.1152/jn.00087.2018. ISSN 0022-3077. PMID 29897858. S2CID 49189631.
  5. ^ Baddeley, Alan D.; Hitch, Graham (1974-01-01), Bower, Gordon H. (ed.), Working Memory, Psychology of Learning and Motivation, vol. 8, Academic Press, pp. 47–89, doi:10.1016/s0079-7421(08)60452-1, ISBN 9780125433082, retrieved 2022-05-16
  6. ^ a b Cermak, Gregory W. (1971). "Short-term recognition memory for complex free-form figures". Psychonomic Science. 25 (4): 209–211. doi:10.3758/BF03329095.
  7. ^ a b Phillips, W.A. (1974). "On the distinction between sensory storage and short-term visual memory". Perception & Psychophysics. 16 (2): 283–290. doi:10.3758/bf03203943.
  8. ^ a b Phillips, W.A.; Baddeley, A.D. (1971). "Reaction time and short-term visual memory". Psychonomic Science. 22 (2): 73–74. doi:10.3758/bf03332500.
  9. ^ Riggs, K.J.; McTaggart, J.; Simpson, A. (2006). "Changes in the capacity of visual working memory in 5- to 10-year-olds". Journal of Experimental Child Psychology. 95 (1): 18–26. doi:10.1016/j.jecp.2006.03.009. PMID 16678845.
  10. ^ Todd, J. Jay; Marois, René (2004). "Capacity limit of visual short-term memory in human posterior parietal cortex". Nature. 428 (6984): 751–754. Bibcode:2004Natur.428..751T. doi:10.1038/nature02466. PMID 15085133. S2CID 4415712.
  11. ^ Luck, S.J.; Vogel, E.K. (1997). "The capacity of visual working memory for features and conjunctions". Nature. 390 (6657): 279–281. Bibcode:1997Natur.390..279L. doi:10.1038/36846. PMID 9384378. S2CID 205025290.
  12. ^ a b Wilken, P.; Ma, W.J. (2004). "A detection theory account of change detection". J Vis. 4 (12): 1120–1135. doi:10.1167/4.12.11. PMID 15669916.
  13. ^ Bays, P.M.; Husain, M. (2008). "Dynamic shifts of limited working memory resources in human vision". Science. 321 (5890): 851–854. Bibcode:2008Sci...321..851B. doi:10.1126/science.1158023. PMC 2532743. PMID 18687968.
  14. ^ Magnussen, S (2000). "Low-level memory processes in vision". Trends in Neurosciences. 23 (6): 247–251. doi:10.1016/s0166-2236(00)01569-1. PMID 10838593. S2CID 16231057.
  15. ^ a b Greenlee, M.W.; Thomas, J.P. (1993). "Simultaneous discrimination of the spatial frequency and contrast of periodic stimuli". Journal of the Optical Society of America A. 10 (3): 395–404. Bibcode:1993JOSAA..10..395G. doi:10.1364/josaa.10.000395. PMID 8473947.
  16. ^ Vincent, A.; Regan, D. (1995). "Parallel independent encoding of orientation, spatial frequency, and contrast". Perception. 24 (5): 491–499. doi:10.1068/p240491. PMID 7567425. S2CID 25950156.
  17. ^ Magnussen, S.; Greenlee, M.W. (1997). "Competition and sharing of processing resources in visual discrimination". Journal of Experimental Psychology: Human Perception and Performance. 23 (6): 1603–1616. doi:10.1037/0096-1523.23.6.1603. PMID 9425670.
  18. ^ Magnussen, Greenlee & Thomas 1997
  19. ^ Palmer, J (1990). "Attentional limits on the perception and memory of visual information". Journal of Experimental Psychology: Human Perception and Performance. 16 (2): 332–350. doi:10.1037/0096-1523.16.2.332. PMID 2142203.
  20. ^ Sligte, Ilja G.; Scholte, H. Steven; Lamme, Victor A. F. (2008). "Are There Multiple Visual Short-Term Memory Stores?". PLOS ONE. 3 (2): e1699. Bibcode:2008PLoSO...3.1699S. doi:10.1371/journal.pone.0001699. PMC 2246033. PMID 18301775.
  21. ^ Pinto, Y.; Sligte, I.S.; Shapiro, K.L.; Lamme, V.A.F. (2013). "Fragile Visual Short-Term Memory is an object-based and location-specific storage". Psychonomic Bulletin & Review. 20 (4): 732–739. doi:10.3758/s13423-013-0393-4. PMID 23456410.
  22. ^ Sligte, I. G.; Scholte, H. S.; Lamme, V. A. F. (2009). "V4 Activity Predicts the Strength of Visual Short-Term Memory Representations". Journal of Neuroscience. 29 (23): 7432–7438. doi:10.1523/JNEUROSCI.0784-09.2009. PMC 6665414. PMID 19515911.
  23. ^ a b c Kozlovskiy, Stanislav; Rogachev, Anton (2021). "How Areas of Ventral Visual Stream Interact when We Memorize Color and Shape Information". Advances in Cognitive Research, Artificial Intelligence and Neuroinformatics. Advances in Intelligent Systems and Computing. Vol. 1358. Springer-Nature. pp. 95–100. doi:10.1007/978-3-030-71637-0_10. ISBN 978-3-030-71636-3. ISSN 2194-5357. S2CID 234902744.
  24. ^ Martinez-Conde, S; Alexander, R (2019). "A gaze bias in the mind's eye". Nature Human Behaviour. 3 (5): 424–425. doi:10.1038/s41562-019-0546-1. PMID 31089295. S2CID 71148025.

Sources

[edit]