History
The term "working memory" was coined by Miller, Galanter, and Pribram, and was used in the 1960s in the context of theories that likened the mind to a computer. In 1968, Atkinson and Shiffrin used the term to describe their "short-term store". What we now call working memory was formerly referred to variously as a "short-term store" or short-term memory, primary memory, immediate memory, operant memory, and provisional memory. Short-term memory is the ability to remember information over a brief period (in the order of seconds). Most theorists today use the concept of working memory to replace or include the older concept of short-term memory, marking a stronger emphasis on the notion of manipulating information rather than mere maintenance. The earliest mention of experiments on the neural basis of working memory can be traced back to more than 100 years ago, when Hitzig and Ferrier described ablation experiments of theTheories
Numerous models have been proposed for how working memory functions, both anatomically and cognitively. Of those, the two that have been most influential are summarized below.The multicomponent model
Working memory as part of long-term memory
Anders Ericsson and Walter Kintsch have introduced the notion of "long-term working memory", which they define as a set of "retrieval structures" in long-term memory that enable seamless access to the information relevant for everyday tasks. In this way, parts of long-term memory effectively function as working memory. In a similar vein, Cowan does not regard working memory as a separate system from long-term memory. Representations in working memory are a subset of representations in long-term memory. Working memory is organized into two embedded levels. The first consists of long-term memory representations that are activated. There can be many of these—there is theoretically no limit to the activation of representations in long-term memory. The second level is called the focus of attention. The focus is regarded as having a limited capacity and holds up to four of the activated representations. Oberauer has extended Cowan's model by adding a third component, a more narrow focus of attention that holds only one chunk at a time. The one-element focus is embedded in the four-element focus and serves to select a single chunk for processing. For example, four digits can be held in mind at the same time in Cowan's "focus of attention". When the individual wishes to perform a process on each of these digits—for example, adding the number two to each digit—separate processing is required for each digit since most individuals cannot perform several mathematical processes in parallel. Oberauer's attentional component selects one of the digits for processing and then shifts the attentional focus to the next digit, continuing until all digits have been processed.Capacity
Working memory is widely acknowledged as having limited capacity. An early quantification of the capacity limit associated with short-term memory was the " magical number seven" suggested by Miller in 1956. Republished: He claimed that the information-processing capacity of young adults is around seven elements, which he called "chunks", regardless of whether the elements are digits, letters, words, or other units. Later research revealed this number depends on the category of chunks used (e.g., span may be around seven for digits, six for letters, and five for words), and even on features of the chunks within a category. For instance, span is lower for long than short words. In general, memory span for verbal contents (digits, letters, words, etc.) depends on the phonological complexity of the content (i.e., the number of phonemes, the number of syllables), and on the lexical status of the contents (whether the contents are words known to the person or not). Several other factors affect a person's measured span, and therefore it is difficult to pin down the capacity of short-term or working memory to a number of chunks. Nonetheless, Cowan proposed that working memory has a capacity of about four chunks in young adults (and fewer in children and old adults). In the visual domain, some investigations report no fixed capacity limit with respect to the total number of items that can be held in working memory. Instead, the results argue for a limited resource that can be flexibly shared between items retained in memory (see below in Resource theories), with some items in the focus of attention being allocated more resource and recalled with greater precision. Whereas most adults can repeat about seven digits in correct order, some individuals have shown impressive enlargements of their digit span—up to 80 digits. This feat is possible by extensive training on an encoding strategy by which the digits in a list are grouped (usually in groups of three to five) and these groups are encoded as a single unit (a chunk). For this to succeed, participants must be able to recognize the groups as some known string of digits. One person studied by Ericsson and his colleagues, for example, used an extensive knowledge of racing times from the history of sports in the process of coding chunks: several such chunks could then be combined into a higher-order chunk, forming a hierarchy of chunks. In this way, only some chunks at the highest level of the hierarchy must be retained in working memory, and for retrieval the chunks are unpacked. That is, the chunks in working memory act as retrieval cues that point to the digits they contain. Practicing memory skills such as these does not expand working memory capacity proper: it is the capacity to transfer (and retrieve) information from long-term memory that is improved, according to Ericsson and Kintsch (1995; see also Gobet & Simon, 2000).Measures and correlates
Working memory capacity can be tested by a variety of tasks. A commonly used measure is a dual-task paradigm, combining a memory span measure with a concurrent processing task, sometimes referred to as "complex span". Daneman and Carpenter invented the first version of this kind of task, the " reading span", in 1980. Subjects read a number of sentences (usually between two and six) and tried to remember the last word of each sentence. At the end of the list of sentences, they repeated back the words in their correct order. Other tasks that do not have this dual-task nature have also been shown to be good measures of working memory capacity. Whereas Daneman and Carpenter believed that the combination of "storage" (maintenance) and processing is needed to measure working memory capacity, we know now that the capacity of working memory can be measured with short-term memory tasks that have no additional processing component. Conversely, working memory capacity can also be measured with certain processing tasks that don't involve maintenance of information. The question of what features a task must have to qualify as a good measure of working memory capacity is a topic of ongoing research. Recently, several studies of visual working memory have used delayed response tasks. These use analogue responses in a continuous space, rather than a binary (correct/incorrect) recall method, as often used in visual change detection tasks. Instead of asking participants to report whether a change occurred between the memory and probe array, delayed reproduction tasks require them to reproduce the precise quality of a visual feature, e.g. an object’s location, orientation or colour. Measures of working-memory capacity are strongly related to performance in other complex cognitive tasks, such as reading comprehension, problem solving, and with measures ofExperimental studies of working-memory capacity
There are several hypotheses about the nature of the capacity limit. One is that a limited pool of cognitive resources is needed to keep representations active and thereby available for processing, and for carrying out processes. Another hypothesis is that memory traces in working memory decay within a few seconds, unless refreshed through rehearsal, and because the speed of rehearsal is limited, we can maintain only a limited amount of information. Yet another idea is that representations held in working memory interfere with each other.Decay theories
The assumption that the contents of short-term or working memory decay over time, unless decay is prevented by rehearsal, goes back to the early days of experimental research on short-term memory. It is also an important assumption in the multi-component theory of working memory. The most elaborate decay-based theory of working memory to date is the "time-based resource sharing model". This theory assumes that representations in working memory decay unless they are refreshed. Refreshing them requires an attentional mechanism that is also needed for any concurrent processing task. When there are small time intervals in which the processing task does not require attention, this time can be used to refresh memory traces. The theory therefore predicts that the amount of forgetting depends on the temporal density of attentional demands of the processing task—this density is called "cognitive load". The cognitive load depends on two variables, the rate at which the processing task requires individual steps to be carried out, and the duration of each step. For example, if the processing task consists of adding digits, then having to add another digit every half second places a higher cognitive load on the system than having to add another digit every two seconds. In a series of experiments, Barrouillet and colleagues have shown that memory for lists of letters depends neither on the number of processing steps nor the total time of processing but on cognitive load.Resource theories
Resource theories assume that the capacity of working memory is a limited resource that must be shared between all representations that need to be maintained in working memory simultaneously. Some resource theorists also assume that maintenance and concurrent processing share the same resource; this can explain why maintenance is typically impaired by a concurrent processing demand. Resource theories have been very successful in explaining data from tests of working memory for simple visual features, such as colors or orientations of bars. An ongoing debate is whether the resource is a continuous quantity that can be subdivided among any number of items in working memory, or whether it consists of a small number of discrete "slots", each of which can be assigned to one memory item, so that only a limited number of about 3 items can be maintained in working memory at all.Interference theories
Several forms of interference have been discussed by theorists. One of the oldest ideas is that new items simply replace older ones in working memory. Another form of interference is retrieval competition. For example, when the task is to remember a list of 7 words in their order, we need to start recall with the first word. While trying to retrieve the first word, the second word, which is represented in proximity, is accidentally retrieved as well, and the two compete for being recalled. Errors in serial recall tasks are often confusions of neighboring items on a memory list (so-called transpositions), showing that retrieval competition plays a role in limiting our ability to recall lists in order, and probably also in other working memory tasks. A third form of interference is the distortion of representations by superposition: When multiple representations are added on top of each other, each of them is blurred by the presence of all the others. A fourth form of interference assumed by some authors is feature overwriting. The idea is that each word, digit, or other item in working memory is represented as a bundle of features, and when two items share some features, one of them steals the features from the other. The more items are held in working memory, and the more their features overlap, the more each of them will be degraded by the loss of some features.Limitations
None of these hypotheses can explain the experimental data entirely. The resource hypothesis, for example, was meant to explain the trade-off between maintenance and processing: The more information must be maintained in working memory, the slower and more error prone concurrent processes become, and with a higher demand on concurrent processing memory suffers. This trade-off has been investigated by tasks like the reading-span task described above. It has been found that the amount of trade-off depends on the similarity of the information to be remembered and the information to be processed. For example, remembering numbers while processing spatial information, or remembering spatial information while processing numbers, impair each other much less than when material of the same kind must be remembered and processed. Also, remembering words and processing digits, or remembering digits and processing words, is easier than remembering and processing materials of the same category. These findings are also difficult to explain for the decay hypothesis, because decay of memory representations should depend only on how long the processing task delays rehearsal or recall, not on the content of the processing task. A further problem for the decay hypothesis comes from experiments in which the recall of a list of letters was delayed, either by instructing participants to recall at a slower pace, or by instructing them to say an irrelevant word once or three times in between recall of each letter. Delaying recall had virtually no effect on recall accuracy. The interference theory seems to fare best with explaining why the similarity between memory contents and the contents of concurrent processing tasks affects how much they impair each other. More similar materials are more likely to be confused, leading to retrieval competition.Development
The capacity of working memory increases gradually over childhood and declines gradually in old age.Childhood
Measures of performance on tests of working memory increase continuously between early childhood and adolescence, while the structure of correlations between different tests remains largely constant. Starting with work in the Neo-Piagetian tradition, theorists have argued that the growth of working-memory capacity is a major driving force of cognitive development. This hypothesis has received substantial empirical support from studies showing that the capacity of working memory is a strong predictor of cognitive abilities in childhood. Particularly strong evidence for a role of working memory for development comes from a longitudinal study showing that working-memory capacity at one age predicts reasoning ability at a later age. Studies in the Neo-Piagetian tradition have added to this picture by analyzing the complexity of cognitive tasks in terms of the number of items or relations that have to be considered simultaneously for a solution. Across a broad range of tasks, children manage task versions of the same level of complexity at about the same age, consistent with the view that working memory capacity limits the complexity they can handle at a given age. Although neuroscience studies support the notion that children rely on prefrontal cortex for performing various working memory tasks, an fMRI meta-analysis on children compared to adults performing the n back task revealed lack of consistent prefrontal cortex activation in children, while posterior regions including theAging
Working memory is among the cognitive functions most sensitive to decline in old age. Several explanations for this decline have been offered. One is the processing speed theory of cognitive aging by Tim Salthouse. Drawing on the finding that cognitive processes generally slow as people grow older, Salthouse argues that slower processing leaves more time for working memory content to decay, thus reducing effective capacity. However, the decline of working memory capacity cannot be entirely attributed to slowing because capacity declines more in old age than speed. Another proposal is the inhibition hypothesis advanced by Lynn Hasher and Rose Zacks. This theory assumes a general deficit in old age in the ability to inhibit irrelevant information. Thus, working memory should tend to be cluttered with irrelevant content that reduces effective capacity for relevant content. The assumption of an inhibition deficit in old age has received much empirical support but, so far, it is not clear whether the decline in inhibitory ability fully explains the decline of working memory capacity. An explanation on the neural level of the decline of working memory and other cognitive functions in old age has been proposed by West. She argues that working memory depends to a large degree on theTraining
Some studies in the effects of training on working memory, including the first by Torkel Klingberg, suggest that working memory in those with ADHD can improve by training. This study found that a period ofIn the brain
Neural mechanisms of maintaining information
The first insights into the neuronal and neurotransmitter basis of working memory came from animal research. The work of Jacobsen and Fulton in the 1930s first showed that lesions to the PFC impaired spatial working memory performance in monkeys. The later work of Joaquin Fuster recorded the electrical activity of neurons in the PFC of monkeys while they were doing a delayed matching task. In that task, the monkey sees how the experimenter places a bit of food under one of two identical-looking cups. A shutter is then lowered for a variable delay period, screening off the cups from the monkey's view. After the delay, the shutter opens and the monkey is allowed to retrieve the food from under the cups. Successful retrieval in the first attempt – something the animal can achieve after some training on the task – requires holding the location of the food in memory over the delay period. Fuster found neurons in the PFC that fired mostly during the delay period, suggesting that they were involved in representing the food location while it was invisible. Later research has shown similar delay-active neurons also in the posterior parietal cortex, theLocalization in the brain
Localization of brain functions in humans has become much easier with the advent of brain imaging methods ( PET and fMRI). This research has confirmed that areas in the PFC are involved in working memory functions. During the 1990s much debate has centered on the different functions of the ventrolateral (i.e., lower areas) and the dorsolateral (higher) areas of the PFC. A human lesion study provides additional evidence for the role of the dorsolateral prefrontal cortex in working memory. One view was that the dorsolateral areas are responsible for spatial working memory and the ventrolateral areas for non-spatial working memory. Another view proposed a functional distinction, arguing that ventrolateral areas are mostly involved in pure maintenance of information, whereas dorsolateral areas are more involved in tasks requiring some processing of the memorized material. The debate is not entirely resolved but most of the evidence supports the functional distinction. Brain imaging has revealed that working memory functions are not limited to the PFC. A review of numerous studies shows areas of activation during working memory tasks scattered over a large part of the cortex. There is a tendency for spatial tasks to recruit more right-hemisphere areas, and for verbal and object working memory to recruit more left-hemisphere areas. The activation during verbal working memory tasks can be broken down into one component reflecting maintenance, in the left posterior parietal cortex, and a component reflecting subvocal rehearsal, in the left frontal cortex (Broca's area, known to be involved in speech production). There is an emerging consensus that most working memory tasks recruit a network of PFC and parietal areas. A study has shown that during a working memory task the connectivity between these areas increases. Another study has demonstrated that these areas are necessary for working memory, and not simply activated accidentally during working memory tasks, by temporarily blocking them through transcranial magnetic stimulation (TMS), thereby producing an impairment in task performance. A current debate concerns the function of these brain areas. The PFC has been found to be active in a variety of tasks that require executive functions. This has led some researchers to argue that the role of PFC in working memory is in controlling attention, selecting strategies, and manipulating information in working memory, but not in maintenance of information. The maintenance function is attributed to more posterior areas of the brain, including the parietal cortex. Other authors interpret the activity in parietal cortex as reflecting executive functions, because the same area is also activated in other tasks requiring attention but not memory. A 2003 meta-analysis of 60 neuroimaging studies found left frontal cortex was involved in low-task demand verbal working memory and right frontal cortex for spatial working memory. Brodmann's areas (BAs) 6, 8, and 9, in the superior frontal cortex was involved when working memory must be continuously updated and when memory for temporal order had to be maintained. Right Brodmann 10 and 47 in the ventral frontal cortex were involved more frequently with demand for manipulation such as dual-task requirements or mental operations, and Brodmann 7 in the posterior parietal cortex was also involved in all types of executive function. Working memory has been suggested to involve two processes with different neuroanatomical locations in the frontal and parietal lobes. First, a selection operation that retrieves the most relevant item, and second an updating operation that changes the focus of attention made upon it. Updating the attentional focus has been found to involve the transient activation in the caudal superior frontal sulcus and posterior parietal cortex, while increasing demands on selection selectively changes activation in the rostral superior frontal sulcus and posterior cingulate/