MEMORY SPAN TASKS Measuring Working Memory Capacity with
Transcription
MEMORY SPAN TASKS Measuring Working Memory Capacity with
Running head: MEMORY SPAN TASKS Measuring Working Memory Capacity with a Simple Span Task: The Chunking Span and its Relationship to Fluid Intelligence Mustapha Chekaf a, Nicolas Gauvrit b, Alessandro Guida c & Fabien Mathy d a Département de Psychologie, Université de Franche-Comté b c d CHArt Lab (PARIS-reasoning) Département de Psychologie, Université Rennes II Département de Psychologie, Université Nice Sophia Antipolis Author note: Corresponding Author: Fabien Mathy, Département de Psychologie, Université Nice Sophia Antipolis, Laboratoire BCL: Bases, Corpus, langage - UMR 7320, Campus SJA3, 24 avenue des diables bleus, 06357 Nice CEDEX 4. Email: [email protected]. This research was supported in part by a grant from the Région de Franche-Comté AAP2013 awarded to Fabien Mathy and Mustapha Chekaf. We are grateful to Caroline Jacquin for her assistance in data collection in Exp. 2. We are also grateful to the Attention & Working Memory Lab at Georgia Tech for helpful discussion, and particularly to Tyler Harrisson who suggested two of the structural equation models. Authors’ contribution: FM initiated the study and formulated the hypotheses; MC and FM conceived and designed the experiments; MC run the experiments; MC, NG, and FM analyzed the data; NG computed the algorithmic complexity; MC, NG, AG, and FM wrote the paper. Work submitted in part to the Proceedings of the Cognitive Science Society, 2015. Body text word count: 20500 1 Abstract It has been assumed that short-term memory and working memory refer to the storage and the storage + processing of information respectively. The two separate concepts have been measured in simple and complex span tasks respectively, but here, a new span task was developed to study how storage and processing interact in working memory and how they relate to intelligence. This objective was reached by introducing a chunking factor in the new span task that allowed manipulating the quantity of processing that can be used for storage while using a procedure similar to simple span tasks. The main hypothesis was that the storage × processing interaction induced by the chunking factor is an excellent indicator to study the relationship between working memory capacity and intelligence, because both (working memory and intelligence) depend on optimizing storage capacity. Two experiments used an adaptation of the SIMON® game in which chunking opportunities were estimated using an algorithmic complexity metric. The results show that the metric can be used to predict memory performance and that intelligence is well predicted by the chunking span task. 2 Measuring Working Memory Capacity with a Simple Span Task: The Chunking Span and its Relationship to Intelligence The present study explores the limits of the short-term memory (STM) span, measured by the length of the longest sequence that can be recalled over brief periods of time. One crucial issue when measuring individuals’ memory spans is that they are inevitably related to other processes that might inflate their measures, such as information reorganization into chunks (e.g., Cowan, 2001; Cowan, Rouder, Blume, & Saults, 2012; Mathy & Feldman, 2012; Miller, 1956) and/or even long-term memory storage (e.g., Ericsson & Kintsch, 1995; Gobet & Simon, 1996; Guida, Gobet, Tardieu, & Nicolas, 2012). This study aims to investigate how information reorganization through chunking can be used to optimize immediate recall by introducing a new simple span measure: the chunking span. It is based on a measure of complexity that, we argue, captures the sum of information that can be grouped to form chunks. In the paper, we examine how this new span measure relates to intelligence and other immediate memory measures (used as an umbrella term for short-term memory and working memory [WM]). The first section deals with what we think represents the mainstream consensus in the literature regarding the conceptualization of span tasks, including a brief review of the respective advantages and disadvantages of STM spans and WM spans. We then describes the rather paradoxical relation between these two kinds of measures and their respective STM and WM concept targets. Finally we introduce the chunking span task that allows varying the degree of involvement of storage and processing using a simple span task, which we believe helps better characterize STM/WM capacity and the nature of its relationship with general fluid intelligence . 3 Span tasks taxonomy It has been commonly accepted that STM and WM refer to the (temporary) storage and (temporary) storage + manipulation of information respectively (e.g., Baddeley & Hitch, 1974; Colom, Rebollo, Abad, & Shih, 2006; A. R. Conway, Cowan, Bunting, Therriault, & Minkoff, 2002; Kane et al., 2004; Engle, Tuholski, Laughlin, & Conway, 1999; for an indepth analysis, see Aben, Stapert, Blokland, 2012; see Davelaar, 2013, for a short explanation), and one of the most popular ideas has been that STM and WM are best represented by simple and complex span tasks respectively (see Shipstead, Redick, & Engle, 2012). Simple span tasks traditionally require retaining a series of items (e.g., digits, words, pictures), whereas in complex span tasks, participants have to maintain the to-be-recalled material while continuously performing concurrent tasks (such as the famous articulatory suppression, e.g., Baddeley, 2000; Baddeley, Thomson, & Buchanan, 1974; Baddeley, Lewis, & Vallar, 1984) or while executing discontinuously interspersed elementary concurrent tasks (e.g., Barrouillet, Bernardin, & Camos, 2004; Barrouillet, Bernardin, Portrat, Vergauwe, & Camos, 2007). The reading span (Daneman & Carpenter, 1980) and the operation span (Turner & Engle, 1989) are two famous examples of this dual-task paradigm, while digit and letter spans (Miller, 1956) traditionally refer to simple span tasks. By comparison with simple span tasks, the complex ones have had an advantage because of the role they have regularly played in theoretical advances, especially in the role of attention in WM (A. R. Conway et al., 2005; Engle, 2002; Oberauer, 2002). For example, they can be used to determine how processing and storage compete simultaneously for attention (e.g., Barouillet & Camos, 2012; Oberauer, Lewandowsky, Farrell, Jarrold, & Greaves, 2012). Although maybe less challenging for the community of researchers, simple span tasks are, however, still used in several intelligence tests (such as the Weschler's) since 4 their use with patients in diverse medical contexts is easy and the instructions are simple. However, one main disadvantage of simple span tasks is that the STM capacity estimated from the tasks can be inflated by practiced skills and strategies such as rehearsal and chunking (Cowan, 2001; Ericsson, Chase, & Faloon, 1980; Miller, 1956). For the last forty years (more or less since Baddeley and Hitch’s seminal 1974 paper) there has been a consensus that WM is more important for complex activities than STM (e.g., Ehrlich & Delafoy, 1990; Klapp, Marshburn, & Lester, 1983; Perfetti & Lesgold, 1977), and indeed complex spans have been reported to be better predictors of complex activities and fluid intelligence than simple spans (Cantor, Engle, & Hamilton, 1991; A. R. Conway, Cowan, Bunting, Therriault, & Minkoff, 2002; Daneman & Carpenter, 1980; Daneman & Merikle, 1996; Dixon, LeFevre, & Twilley, 1988; Unsworth & Engle, 2007a, 2007b; but see Colom, Rebollo, Abad, & Shih, 2006), and particularly for Raven's Advanced Progressive Matrices (A. R. Conway et al., 2005). However, Unsworth and Engle (2007a) recently showed that the prediction of simple spans of fluid intelligence could be increased, increasing list-length. They observed that when list-length reached five, simple spans became as good as complex spans, in terms of predicting fluid intelligence2 (for a comparable result, see Bailey, Dunlosky, & Kane, 2011). One possible explanation is that long-length-lists need to be reorganized by individuals in order to be stored (Chase & Simon, 1973; de Groot & Gobet, 1996; Miller, 1956), because they exceed immediate memory capacity. As such, the 7+/-2 estimation first found by Miller (1956) can be thought to represent an overestimation due to chunking of the true capacity of immediate memory, now most often estimated as being about 2 According to Unsworth and Engle (2006), when the elements to store or process are larger than 3 to 4 elements, primary memory (similar to the kind of immediate memory proposed by James, 1890) is insufficient and elements are displaced in secondary memory (similar to episodic long-term memory), and the process of searching and retrieving from secondary memory would explain the commonalties among long-list-length simple spans, complex spans and fluid intelligence. Another possibility, which is developed in this paper, is to stress the fact that these commonalties are due to information reorganization through chunking. 5 3 or 4 (e.g., Cowan, 2001; Cowan, 2005). Effectively, Mathy and Feldman (2012) modeled the connection between the 7 and 4 magical numbers through chunking. From this standpoint, our hypothesis is that the capability of simple spans (when using long-length-lists) to predict fluid intelligence could be due to information reorganization into chunks. The idea that information reorganization into chunks is an important characteristic of simple spans is completely in accordance with a line of research developed by Bor and colleagues (Bor, Cumming, Scott, & Owen, 2004; Bor, Duncan, Wiseman, & Owen, 2003; Bor & Owen, 2007; Bor & Seth, 2012). Instead of increasing list-lengths to trigger information reorganization, Bor and colleagues introduced systematic regularities into the memoranda in an attempt to induce a chunking process. They showed than when individuals could chunk information thanks to structured material, their simple span increased as did activation in lateral prefrontal areas, which the authors took as the cerebral substratum of information reorganization. These results show that simple spans considered to be only “storage tasks” can be viewed as storage + manipulation tasks as well, when systematic regularities are provided. Even if this method has been used several times by Bor and colleagues (Bor et al., 2003, 2004; Bor & Owen, 2007), it was applied only recently using a metric of compression that allows precisely manipulating the probability of chunking via measures of complexity (Mathy & Feldman, 2012). This will be presented below when we present our chunking span task, which allows a simple span to be changed gradually from a storage task to a storage + manipulation task. This possibility to vary along this dimension (all other things being equal) is made crucial by what we call the STM/WM paradox. The STM/WM paradox The STM/WM paradox derives from the fact that even if there is a mainstream consensus that simple spans can be equated with storage and complex spans with storage + 6 manipulation of information, the picture is much more complex and reveals intertwined concepts. For example, this entanglement led Oberauer, Lange and Engle (2004, p. 94) to conclude that “it might be more fruitful to turn things around: simple span = complex span + specialized mechanisms or strategies”. In fact, notwithstanding the well-known advantages and disadvantages of both types of tasks, the distinction between the STM and WM constructs that these tasks putatively represent is conceptually vague (some attempts at clarification are still being provided, see Aben et al., 2012; Davelaar, 2013). Therefore, we believe that there is still room for developing a new approach that could help clarify STM and WM taxonomy. A prevailing idea in many studies is that the greatest benefit of complex span tasks is to provide a well-controlled refined measure of the span (in comparison to simple span tasks), because processing is directed away from the storage task (concurrently in dual-task set-ups or intermittently in classic complex span tasks), preventing or controlling rehearsal or elaboration of the memoranda. However, coincidently, and contrary to the fundamental analysis of the dual tasks by Baddeley & Hitch (1974), the focus is rarely on the processing component that is supposed to represent half of the core idea of the WM concept (Aben et al., 2012; A. R. Conway et al., 2005; Unsworth, Redick, Heitz, & Engle, 2009). For instance, in the dual-task paradigm, one could also measure performance on the concurrent task (which directs the processing component away from the maintained items) to provide an index of WM capacity, but by design (the concurrent task is not difficult per se), performance on the concurrent task is at the ceiling level and thus it is more rarely used as a predictor3 (but see Unsworth et al., 2009). 3 Another way to study the processing component as a predictor is to measure processing efficiency separately and to partial it out in correlational studies (Salthouse & Babcock, 1991), but in this case, where single-task (processing only) is measured to study its mediation between WM capacity and higher-order cognition, the storage × processing interaction that is targeted in the present study cannot be determined. 7 In contrast, processing the memoranda is permitted in simple span tasks (e.g., participants are free to group a few digits in a digit span task), but paradoxically the target construct of this task is only the storage component that is supposed to best represent the older and supposedly less elaborate STM concept. This is one of the most challenging theoretical frameworks to describe to psychology students: STM is thought to be primarily concerned with storage but the simple tasks involve processing dedicated to the maintained items while WM is primarily concerned with a storage + processing combination that subsequent analyses overlook to focus mainly on storage performance. One interesting task would be of course one that allows methodically sliding across the [(storage) ... (storage + processing)] dimension while maintaining all other things equal. So far, some tasks more directly focus on the storage + processing combination, such as the backward span, the n-back, the letter-number sequencing subtest of the WAIS-IV (Wechsler, 2008), or the running span. Even if in all these tasks, the processing component can be fully dedicated to the stored items, a characteristic that is targeted in the present study, two other characteristics appear more problematic and have led us to propose the chunking span. The first one is that the participants usually find these tasks very difficult, and we believe that the difficulty inherent in these tasks does not leave much room for manipulating the processing demand. Secondly and more importantly, in these tasks, the processing component cannot be studied independently from the storage component, because the number of items to be processed is linearly related to the number of items to be stored. Therefore there is no clear separation between processing and storage because storage and processing both depend on a similar number of items (i.e., the memorandum). This is also the case when one increases the list-length in a simple span (Unsworth and Engle, 2006). As we will see, the chunking span task allows increasing or decreasing the processing demand independently of the number of 8 elements to be remembered by varying the complexity of the lists. To anticipate and to give an example, although the sequence "1223334444" requires 10 items to be sequentially processed, only one chunk can be stored once one successfully recognizes the regular pattern that makes the sequence easy to retain. This is not the case for "8316495207" where several chunks must come into play. Hence, our idea was to develop a memory span task in which both storage and processing could be measured simultaneously and independently, and we argue that this can be done by manipulating complexity and by inducing a chunking process. Chunking span tasks Our new task is based on the framework of SIMON®, a classic memory game from the 80s that can be played on a device that has four colored buttons (red, green, yellow, blue). The game consists of reproducing a sequence of colors by pressing the corresponding buttons. The device lights up the colored buttons at random and increases the number of colors by adding a supplementary color at the end of the previous sequence whenever the reproduction by the player is correct. The game progresses until the player makes a mistake. Gendle and Ransom (2006) reported a procedure for measuring a WM span using SIMON® and they showed that the procedure is resistant to practice effects. Other studies have shown that this setting has many advantages for studying different populations with speech or hearing pathologies (C. M. Conway, Karpicke, & Pisoni, 2007; Humes & Floyd, 2005; Karpicke & Pisoni, 2000, 2004). There were two important differences between the original game and the present adaptation. First, a given chosen sequence was not presented progressively but entirely in a single presentation. For instance, instead of being presented with a “1) blue, 2) blue-red, 3) blue-red-red, etc.”, that is, three series of the same increasing sequence until a mistake was made, the participant in this case would be given a blue-red-red sequence from the outset. If correct, a new sequence was given, possibly using a different complete length, 9 so there was no sequence of increasing length that could have favored a long-term memory process. Second, no sounds were associated with any of the colors. The new task was developed on the bases of the rationale that sequences of colors contain regularities that can be mathematized to estimate a chunking or compression process. Individuals have a tendency to organize information into chunks in order to reduce the quantity of information to retain (Anderson, Bothell, Lebiere, & Matessa, 1998; Cowan, Chen, & Rouder, 2004; Ericsson et al., 1980; Logan, 2004; Miller, 1956, 1958; NavehBenjamin, Cowan, Kilb, & Chen, 2007; Perlman, Pothos, Edwards, & Tzelgov, 2010; Perruchet & Pacteau, 1990; Tulving & Patkau, 1962). Chunking can be separated into two very different processes, chunk creation and chunk retrieval (Guida, Gobet, & Nicolas, 2013; Guida et al., 2012). The first occurs when individuals do not have strong knowledge of the information they are processing. During chunk creation, individuals use the focus of attention (Oakes, Ross-Sheehy, & Luck, 2006; Wheeler & Treisman, 2002) to bind separate elements, and reorganize the information they are processing into groups of elements or chunks. This has been integrated in various models: Cowan and Chen (2009), for example, suggested that one crucial function of the focus of attention (e.g., Cowan, 2001, 2005) is effectively to allow binding; Oberauer’s region of direct access has (among other functions) the same aim (e.g., Oberauer, 2002; Oberauer & Lange, 2009); and Baddeley’s episodic buffer (Baddeley, 2000; Baddeley, 2001) was also put forward to allow an explanation of binding. But once individuals have the knowledge to recognize groups of elements (e.g., “f,” “b,” “i”), they do not need any more to create but only to retrieve chunks from LTM. Chunking retrieval has had considerable impact on the study of immediate recall (Boucher, 2006; Cowan et al., 2004; Gilbert, Boucher, & Jemel, 2014; Gilchrist, Cowan, & Naveh-Benjamin, 2008; Guida et al., 2012; Maybery, Parmentier, & Jones, 2002; Ng & 10 Maybery, 2002), but less is known about the role of immediate memory in the creation of chunks in novel situations (Feigenson & Halberda, 2008; Kibbe & Feigenson, 2014; Moher, Tuerk, & Feigenson, 2012; Solway et al. 2014) and about the reciprocity of WM and chunking (Rabinovich, Varona, Tristan, & Afraimovich, 2014). Because the presence of regularity in the to-be-recalled material can account for greater capacities than expected (for instance, this is the case in the visual WM domain, Brady, Konkle, & Alvarez, 2009, 2011; Brady & Tenenbaum, 2013; temporal clustering, Farrell, 2008, 2012; but also when social cues are used to expand memory, Stahl & Feigenson, 2014), many efforts have been made to hinder grouping in WM span tasks to obtain a rigorous estimation of the span (Cowan, 2001). Only a few studies have manipulated chunking, for instance, by using learned co-occurrences of words, sequences structured by artificial grammar (e.g., Chen & Cowan, 2005; Cowan, Chen, & Rouder, 2004; Gilchrist et al., 2008; Majerus, Perez, & Oberaurer, 2012; NavehBenjamin et al., 2007), or multi-word chunks (Cowan, Rouder, Blume, & Saults, 2012). However, this method is based on having participants acquire long-term representations (explicitly or implicitly) before studying their subsequent use in span tasks, and this method involves a recognition process of previously encountered repetitions of sequences (e.g., Botvinick & Plaut, 2006; Burgess & Hitch, 1999; Cumming, Page, & Norris, 2003; French, Addyman, & Mareschal, 2011; Robinet, Lemaire, & Gordon 2011; Szmalec, Page, & Duyck, 2012). An alternative to this method is to prompt the formation of chunks in immediate memory while avoiding long-term learning effects (Bor et al., 2003, 2004; Bor & Owen, 2007; Mathy & Feldman, 2012). Prompting the formation of chunks can be done by introducing systematic regularities in order to measure whether they can be encoded at once in a more compact way, without repeatedly testing the participants with the same sequences. 11 The present study continues this line of research by demonstrating that participants exposed to simple sequences of colors show higher recall for more regular sequences without any relation to particular prior knowledge in long-term memory (although this is true for the sequences themselves, not for the colors which constitute the sequences). We show that the compressibility of the sequences contributes to the grouping process of the sequences of colors. Complexity for short strings The present study aims to provide a precise estimate of the compressibility of the tobe-remembered sequences of colors when prompting the formation of chunks in immediate memory. Note that contrary to previous studies (e.g., Burgess and Hitch, 1999) that could identify grouping effects by the presence of mini serial position curves during recall (another possibility is to study transitional-error probabilities, Johnson, 1969), the present study only seeks to estimate chunking opportunities more globally (i.e, for the entire to-be-remembered sequence), and as such, it does not focus on how chunks can be built up sequentially. Chunking opportunity can be defined as the probability of a sequence to be re-encoded so it can be retained using a series of meaningful blocks of information instead of independent items (for instance 0-11-11-0 instead of 011110 using a simple linear separation, 0-2*11-0 by further grouping the two similar blocks, or 011-110 using symmetry). The compressibility estimate that we develop here is adequate to capture any kind of regularity within sequences, which can be used by participants to simplify the recall process. To estimate the chunking opportunities that the participants could be offered within thousands of different sequences, a compressibility metric was sought to provide an 12 estimation of any possible grouping process5. More complexity simply means less chunking opportunities, which also indicates that memorization has to be mostly based on storage capability. Less complexity means that a sequence can be re-encoded for optimizing storage and in this case, processing takes precedence over storage. A major difficulty one encounters in this type of study is due to the apparent lack of a normalized measure of compressibility— or complexity. Some formal measures such as entropy are actually widely used as proxy for complexity, but they have come under harsh criticism (Gauvrit, Zenil, Delahaye, & SolerToscano, 2014). We believe that the notion of compression is potentially helpful to relate chunking and intelligence (see Baum, 2004, or Hutter, 2005, who developed similar ideas in artificial intelligence), because many human complex mental activities still fit our quite low storage capabilities. The accepted notion of (algorithmic) complexity in computer science was developed by Kolmogorov (1965) and later refined by Chaitin (1966). It bridges compression and complexity in a single definition: The algorithmic complexity of a sequence is the length of the shortest program (run on a Universal Turing Machine), that will build the sequence and halts (Li & Vitányi, 2009). The algorithmic complexity of long strings can be estimated and this estimation has already been applied to different domains (e.g., in genetics, Ryabko, 5 There is still a lack of consensus on whether chunking is perceptual or conceptual (Gilbert, Boucher, & Jemel, 2014), and on wether is is different from grouping. Feigenson and Halberda (2008), for instance, distinguished a form of chunking that requires conceptual recoding (e.g., parsing PBSBBCCNN into PBS-BBC-CNN, based on existing concepts in long-term memory) from a second form of chunking that only requires a recoding process. They developed the idea that ‘‘eggplant, screwdriver, carrot, artichoke, hammer, pliers’’ is easier to remember than ‘‘eggplant, broccoli, carrot, artichoke, cucumber, zucchini’’ because it can be parsed into three units of two conceptual types without any need to refer to a pre-existing ‘‘eggplant–carrot–artichoke’’ concept. The recoding process here is based on semantics, but the authors developed a third definition of chunking that bases the parsing process on spatiotemporal information (a classic example is dividing phone numbers into groups by proximity). This last idea is similar to Mathy and Feldman’s (2012), who attempted to quantify an immediate chunking process by systematically introducing sequential patterns in the to-beremembered material that are unrelated to knowledge in long-term memory. In the present study as well, although the colors are long-term memory concepts, the sequence yellow-red-yellow-red can easily be encoded without relying on any specific conceptual knowledge to form a new more compact representation such as “Twice yellow-red”. 13 Reznikova, Druzyaka, & Panteleeva, 2013; Yagil, 2009; e.g., in neurology, Fernandez et al., 2011, 2012; Machado, Miranda, Morya, Amaro Jr, & Sameshima, 2010), but contrary to long strings, the algorithmic complexity of short strings (3-50 symbols or values) could not be estimated before recent developments in computer science. However, thanks to recent breakthroughs, it is now possible to obtain a reliable estimation of the algorithmic complexity of short strings (Delahaye & Zenil, 2012; Soler-Toscano, Zenil, Delahaye, & Gauvrit, 2013, 2014). The method has already been used in psychology (e.g., Kempe, Gauvrit & Forsyth, 2015; Dieguez, Wagner-Egger & Gauvrit, in press) and it is now implemented as an Rpackage named ACSS (Algorithmic Complexity for Short Strings; Gauvrit, Singmann, SolerToscano, & Zenil, 2015). Algorithmic complexity is correlated to the human perception of randomness (Gauvrit, Soler-Toscano, & Zenil, 2014), and in this paper we hypothesize that it is a simple proxy for chunking opportunities. The basic idea at the root of the algorithmic complexity for short strings (acss) algorithm is to take advantage of the link between algorithmic complexity and algorithmic probability. The algorithmic probability m(s) of a string s is defined as the probability that a randomly chosen deterministic program, running on a Universal Turing Machine produces s and halts. This probability is related to algorithmic complexity by way of the algorithmic coding theorem which states that K(s) ~ –log2(m(s)), where K(s) is the algorithmic complexity of s. Instead of choosing random programs on a fixed Turing machine, one can equivalently choose random Turing machines and have it run on a blank tape. This has been done on huge samples of Turing machines (more than 10 billion Turing Machines), and led to a distribution d of strings, approximating m. The algorithmic complexity for short strings of a string s, acss(s) is defined as –log2(d), an approximation of K(s) by use of the coding theorem (see Gauvrit, Singmann, Soler-Toscano, & Zenil, 2015). 14 Relationship between storage × processing and intelligence One last (but certainly not least) prediction regarding the benefit of studying how chunking processes affect capacity is related to the relationship between chunking and intelligence. Many studies using latent variables have suggested that WM capacity accounts for 30% to 50% of the variance in g (A. R. Conway, Kane, & Engle, 2003; A. R. Conway et al., 2005). In comparison, simple span tasks account for less covariance (Shipstead et al., 2012). Again, one interesting exception of crucial importance for the present study is that this discrepancy in terms of variance is no longer true when the simple span tasks make use of longer sequences, for instance above 5 items (Unsworth & Engle, 2006, 2007a; see also Bailey et al., 2011; Unsworth & Engle, 2007b). Therefore simple spans can potentially account for the same percentage of the variance in g as complex spans. We pose that the saturation of the storage component occurs for the most complex sequences, since they offer no possibility to process and reorganize information. However, less complex sequences are assumed to favor the occurrence of chunking via reorganization and should thus involve storage × processing7. The two respective kind of sequences mimic in a way complex span tasks and simple span tasks respectively, and we naturally expect smaller spans for the most complex sequences. This difference has already been put forward 7 In this context, the variation of complexity is important because it is assumed that the most complex sequences cannot be easily reorganized and as such they reduce processing opportunities and mainly involve storage. We say "mainly" instead of "totally", since algorithmic complexity is not computable, which means that there can always be a way to reduce the complexity of a sequence that is left unnoticed by a given metric. Consequently, there might always be a residual portion of processing involved because the participants are free to group some items using any kind of simplification method. One example would be that a participant notices that the "red-blue-yellow-green" sequence resembles the Mauritian national flag. Although there is a slight chance that non-compressible sequences can be simplified, there is a greater chance that the memorization of most compressible sequences can be optimized. The simpler sequences require memory optimization because of their regularity, which should further solicit the processing demand. 15 by Bor and colleagues (Bor et al., 2003, 2004; Bor & Owen, 2007). When comparing cerebral activity during a simple span with chunkable material (material with systematic regularities) versus less chunkable material, they observed an increase of activation in lateral prefrontal areas when some material could be chunked, which the authors interpreted as due to information reorganization. Moreover this activation pattern has been linked to intelligence, as a similar neural network has also been found activated in the n-back task, complex spans, and fluid reasoning (Bor & Owen, 2007; Colom, Jung, & Haier, 2007; Duncan, 2006; Gray, Chabris, & Braver, 2003). According to Duncan, Schramm, Thompson, and Dumontheil (2012, p. 868), this shows that fluid intelligence could play a role in “the organizational process of detecting and using novel chunks”. Capitalizing on this idea, we predicted that the mnemonic span for the less complex sequences should better correlate with intelligence because it is in this condition that participants are able to create novel chunks, which can eventually optimize their storage capacity. Another reason (however linked) for making the same prediction is that intelligence tests generally require processing and storing information in conjunction, therefore less complex sequences should better correlate with intelligence since it is in this condition that they allow an estimate of the storage × processing construct. One may note that the targeted interaction between storage and processing that results from the hypothesized optimization process conflicts with preceding efforts in the literature to separate the storage and processing mechanisms (Cowan, 2001), but we believe that in fine, higher-cognition depends on the storage × processing capacity rather than each one separately. Experiment 1 aimed at studying the storage × processing capacity and its relationships to other span tasks and IQ. Experiment 2 used two conditions to separate storage and storage × processing capacities, and investigate their relationships to other span tasks and IQ. Experiment 1 was very liberal in terms of randomly choosing the sequences of colors, which lead us to develop a specific 16 estimation of memory capacity, whereas Experiment 2 used similar sequences across participants that allowed us to use a more standard scoring method for computing a memory span. Although there are already reports studying the relationship between the components of the WM system and intelligence at the latent variable level (see Chuderski et al., 2012; Colom et al., 2004, 2008; Krumm et al., 2009; Martinez et al., 2011 for examples), the present study combines experimental and correlational approaches because of the exploratory nature of the new task. Experiment 1 Experiment 1 develops a chunking memory span task based on an algorithmic complexity metric for short strings and study how it related to intelligence. This metric enabled us to estimate the storage × processing capacity in STM, by measuring the most complex sequence that could be remembered by each participant. The sequences were drawn at random in order to study complexity effects independently from participant’s individual performance, but with the main goal of having a maximum number of different sequences to fit a global performance function. We predicted that performance would decrease with complexity because complexity hinders the capacity to compress information. It was also predicted that performance on the chunking span task would better correlate with other span tasks that solicit a combination of storage and processing, but particularly when processing is dedicated to the stored items such as a memory updating task (i.e., this is not the case in complex span tasks). Finally, it was predicted that the storage × processing component estimated by the chunking span task would predict IQ better than any other types of span tasks because this task involves a process of storage optimization that requires full function of the processing component. 17 One interesting battery developed by Lewandowsky, Oberauer, Yang and Ecker (2010), the Working Memory Capacity Battery (WMCB) allowed us to test which of the storage-processing combinations best predicts IQ, and whether proximity with our task could be predicted by the storage-processing combination specific to each task. For the two complex span tasks of the WMCB, processing is directed away from the stored items. For the spatial span task of the WMCB, the processing component is left uncontrolled and in this case, the processing component is not fully involved but only moderately dedicated to the stored items. The memory updating task of the WMCB can however be regarded as a difficult task in which processing is fully dedicated to the manipulation of the stored items. Again, the processing component is not fully involved in this task because the number of items updated is linearly dependent on the number of stored items, meaning that if the participant's storage capacity is low, processing of the items is also directly limited. Our hypothesis was that our task draws on a clearer storage × processing combination in which the processing component can be used to optimize storage capacity, and in this particular case, the number of items that are processed sometimes exceeds the number of groups that are stored. In this latter case, storage capability (in terms of the number of chunks that can be stored) can be limited, but the number of items that can be packed into the chunks is less constrained. Accordingly, the chunking span task is expected to better correlate with the simple span task and the memory updating task than with the complex span tasks. Method Participants One hundred and eighty-three students enrolled at the Université de Franche-Comté, France (Mage = 21; SD = 2.8) volunteered to participate in this experiment and received course credits in exchange for their participation. 18 Procedure Depending on their availability and their need for course credits, the volunteers took part in one, two or three tests of our computerized adaptation of the electronic game SIMON®, the Working Memory Capacity Battery (WMCB) and the Raven’s Advanced Progressive Matrices (APM) (Raven, 1962). All the volunteers were administered our Simon task, some were also administered the WMCB (27), the Raven (26), or both (85). In all of the cases, the administration respected a strict chronology: 1) Simon, 2) Raven, 3) WMCB. Chunking span task. Each trial began with a fixation cross in the center of the screen (1000ms). The to-be-memorized sequence, consisting of a series of colored squares appearing one after the other, was then displayed (see Figure 1). Then in the recall phase, four colored buttons were displayed and participants could click on them to recall the whole sequence they had memorized and then validate their answer. After each recall phase, feedback (‘perfect’ or ‘not exactly’) was displayed according to the accuracy of the response. We built two versions of the adapted Simon task, thus, participants were either administered a spatial version or a nonspatial one. In the first version (N = 106), the colored squares were spatially located on the screen and were briefly lit up to constitute the to-beremembered sequence, as in the original game. To discourage spatial encoding, the colors were randomly assigned to the locations for each new trial. In the second version (N = 77), the colors were displayed one after another in the center of the screen in order to avoid any visuo-spatial strategy of encoding. To further discourage spatial strategies in the two versions, the colors were randomly assigned to the buttons in the response screen for each trial, which resulted in the colors never being in the same locations between trials. Our results showed no significant difference between these two conditions, thus both data sets were compiled in the following. 19 Each session consisted of a single block of 50 sequences varying in length (from one to ten) and in the number of possible colors (from one to four). New sequences were generated for each participant, with random colors and orders, with the aim of avoiding ascending presentation of the length or the complexity of the items (two items, then three, then four, etc.) and to hinder opportunities to develop strategies based on expectation of complexity or the number of the to-be-remembered items, in order to avoid any learning or prediction effect. The choice was made to generate random sequences and to measure their complexity a posteriori (as described below). A total of 9150 sequences (183 subjects × 50 sequences) were presented (average length = 6.24), each session lasted 25 min on average. Figure 1. Example of a sequence of three colors for the chunking memory span task adapted from the SIMON® game. 20 Working Memory Capacity Battery (WMCB). Lewandowsky, Oberauer, Yang, & Ecker (2010) developed this battery for assessing WM capacity through four tasks: an updating task (memory updating, MU), two span tasks (operation and sentence span, OS and SS), and a spatial span task (spatial short-term memory, SSTM). This battery was developed using MATLAB (MathWorks Ltd.) and the Psychophysics Toolbox (Brainard, 1997; Pelli, 1997). On each trial of MU, the participants were required to store a series of numbers in memory (from 3 to 5), which appeared in respective frames one after another for 1 second each. Next, only the empty frames remained on the screen, and this was followed by a sequence of several cues corresponding to arithmetic operations such as “+ 3” or “- 5” that were displayed in the frames one at a time in a random frame. The participants had to apply these operations to update the memorized sequence of digits. In both OS and SS, a complex span task paradigm was used so that the participants saw an alternating sequence of to-beremembered consonants and to-be-judged propositions (the correctness of equations in the 'OS' task, or the meaningfulness of the sentences in the 'SS' task). The participants were required to memorize the whole sequence of consonants for immediate serial recall. In SSTM, the participants were required to remember the location of dots in a 10 × 10 grid. The dots were presented one by one in random cells. After the sequence was terminated, the participants were cued to reproduce the pattern of dots using a mouse. They were instructed that the exact position of the dots or the order was irrelevant; they mostly had to remember the pattern made by the spatial relations between the dots. Completing the four tasks required 45 minutes on average. Raven’s Advanced Progressive Matrices After a practice session using the 12 questions in set #1, the participants were asked to complete the matrices in set #2 which contained 36 matrices, during a timed session averaging 40 minutes. The participants were 21 expected to select the correct missing cell of a matrix from a set of eight propositions. Correct completion of a matrix was scored one point, so the range of possible raw scores was 0–36. Results Effects of the sequence lengths and the number of colors per sequence First, we conducted a repeated-measures ANOVA with the length of the sequence as a repeated factor, and using the mean proportion of perfectly recalled sequences for each participant as a dependent variable (i.e., proportion of trials in which all items in a sequence were correctly recall8). Performance varied across lengths, F(9, 177) = 234.2, p < .001, η2 = .9, with the proportion regularly decreasing as a function of length. Figure 2 shows the performance curve based on the aggregated data by participant, by length and also as a function of the number of different colors that appeared within the sequences. The three performance curves resemble an S-shaped function as in previous studies (e.g., Crannell & Parrish, 1957). We then focused on a subset of the data in order to study the interaction between Length and Number of colors. We selected trials for which Length exceeded three items and we conducted a 7 (4, 5, 6, 7, 8, 9 or 10 items) × 3 (2, 3 or 4 different colors within sequences) repeated-measures ANOVA. There was still a significant main effect of the Length factor (respectively, Ms = .91, .82, .64, .49, .32, .2 and .13; SEs = .01, .01, .02, .02, .02, .01 and 0.01), F(6,1092) = 560.8, p < .01, ηp2 = .75, and a significant main effect of the number of 8 Because repetitions occur within sequences, the proportion of correctly recalled items was not computed as it unfortunately involves more complex scoring methods based on alignment algorithms (Mathy & Varré, 2013) which are not yet considered as standardized scoring procedures. 22 colors per sequence, (Ms = .6, .47 and .42; SEs = .01, .01 and .01), F(2,364) = 132.4, p < .01, ηp2= .42. Posthoc analysis (Bonferroni corrections were made for the pairwise comparisons) showed a systematic significant decrease between each Length condition and between each Number of colors condition. Finally, there was a significant interaction between Length and Number of colors, F (12, 2184) = 5.8, p < .01, ηp2 = 0.03: the effect of Length increased with the number of colors. Although this is a coarse estimation of chunking opportunities, the results show that memorization was facilitated by using sequences made of a fewer number of colors, and particularly when the sequences were longer. 23 Figure 2. Proportion of perfectly recalled sequences of colors as a function of sequence length and as a function of the number of different colors within sequences. Note: Error bars are +/one standard error. Effect of complexity To confirm the reliability of our complexity measure, we split the data into two groups of sequences of equivalent complexity for each participant. This split-half method enabled simulating a situation in which the participants had taken two equivalent tests. We obtained an adequate evidence of reliability between the two groups of sequences (r = .63, p < .001; Spearman-Brown coefficient = .77). A simple first analysis targeting both the effect of the number of colors and the effect of complexity on accuracy (i.e., the sequence is perfectly recalled) showed that complexity (β = −.62) took precedence over the number of colors (β = −.04) in a multiple linear regression analysis, when all of the 9150 trials were considered. To investigate more precisely the combined effects of complexity and list-length on recall in more detail, we used a logistic regression approach. A stepwise forward model selection based on BIC criterion suggested dropping the interaction term (see Table 1). This model showed a significant negative effect of complexity (z(9147) = -23.84, p < .001, standardized coefficient = -5.70 [non standardized: -.69]) as shown in Figure 3, and a significant positive effect of length (z(9147) = 16.27, p < .001, standardized coefficient = 3.74 [non standardized: 1.46]). Although length had a detrimental effect on recall, this effect was more than compensated by the detrimental effect of complexity, meaning that long simple strings were easier to recall than shorter but more complex strings. In other words, the effect of complexity was stronger than the effect of length (Table 1). 24 Table 1 Stepwise forward selection of several models based on BIC criterion. At each stage, for a given model (model ~ v, meaning that the base model is computed as a function of x to predict performance correct), the variables are listed in increasing BIC order after other variables are introduced. The final model includes complexity and length, but no interaction term. For instance, in the last model computed as a function of Complexity and Length, adding the interaction term increases the BIC, so the simplest model is chosen. Base model Variable included Deviance BIC Intercept only Complexity 8093.7 8111.9 Length 8462.3 8480.5 None (intercept only) 12477.5 12486.6 correct ~ Complexity Length 7805.2 7832.6 None (Complexity only) 8093.7 8111.9 correct ~ Complexity and Length None (Complexity and Length only) 7805.2 7832.6 Interaction 7800.0 7836.5 Figure 4 showed how performance decreased as a function of complexity in each length condition. The decreasing linear trend was significant for lengths 4, 6, 7, 8, 9 and 10 (r = -.34, p < .001; r = -.39, p < .001; r = -.41, p < .001; r = -.35, p < .001; r = -.45, p < .001; r = -.43, p < .001, respectively), which shows that the memory process could effectively be optimized for the less complex sequences. 25 Figure 3. Proportion of perfectly recalled sequences of colors as a function of complexity. Note: Error bars are +/- one standard error. One interesting result of our complexity measure was based on the comparison between the complexity of the stimulus sequence and the response sequence. When the participants failed to recall the right sequence, they showed a tendency to produce a simpler string in terms of algorithmic complexity. The mean complexity of the stimulus sequence was 25.3, but the mean complexity of the responses was only 23.3 across all trials (t(9149) = 26.24, p < .001, Cohen’s paired d = .27). 26 Figure 4. Proportion correct as a function of complexity, by sequence length. Error bars are +/- one standard error. To test whether participants with a greater score on the Raven better optimized their encoding of the most simple sequences (when the length of the sequences remained constant), we further separated the sample of participants into two groups (high IQ vs. low IQ) using the median of the Raven's scores. We also separated all the sequences of length > 5 (keeping the shorter sequences in the data set produced floor effects) into two complexity levels (low vs. high complexity) by selecting the sequences below the tenth percentile and above the ninetieth percentile of the complexity metric. The repeated-measures ANOVA, with complexity level (high vs. low) as a within-subjects factor and IQ level (high vs. low) as a between-subjects factor showed a significant interaction, F(1,109) = 15, p < .01, ηp2 = .12, which was, however, mostly accounted for by the almost null performance in the high complexity condition in both groups (Figure 5). 27 0.9 IQ < 50th % IQ > 50th % 0.8 0.7 Prop. correct 0.6 0.5 0.4 0.3 0.2 0.1 0 < 10th % > 90th % Complexity Figure 5. Proportion of correct recalled sequences of colors as a function of complexity (below 10th percentile vs. above 90th percentile) and IQ (below 50th percentile vs. above 50th percentile). Error bars are +/- one standard error. To ensure that this result was not obscured by a floor effect on the most complex sequences, and to gain more power, a subsequent mixed-model analysis was done on accuracy per trial with complexity as a within-subjects factor (the complexity variable was split into four quartiles to obtain a smaller number of levels), IQ as a between-subject factor (22 levels, since there were only 22 types of IQ scores), and subject as a random factor, which showed significant effects of complexity (F(2,15183.4) = 2143, p < .001), IQ (F(5,20.9) = 31.6, p < .001), and interaction, F(42,12.9) = 15.1, p < .001 (Figure 6 shows the interaction between 28 complexity and IQ, as performance slightly increased for lower complexity sequences in the highest IQ group). &', 9 ;!!!" ;!!!# ;!!!$ ;!!!% &'+ &'* 78.0'9:.882:5 &') &'( &'% &'$ &'# &'" & 9 !" !# !$ -./0123456 !% Figure 6. Proportion of correct recalled sequences of colors as a function of complexity and IQ. Note: Both the complexity and IQ variables were split into quartiles before aggregating the data by participant. Error bars are +/- one standard error. Correlations and factor analyses Table 2 shows the correlations between measures aggregated by participants, including the different span measures of the WMCB (MU, OS, SS, and SSTM, and WM was the global score for the entire battery), the Raven (M = 23.6, slightly above the 50th percentile; SD = 4.1; which corresponds to an average IQ of 101.2, SD = 11.7; N = 113), and the average capacity for the Simon span task. Because the participants were not administered 29 the same sequences, we computed a logistic regression for each subject to find the critical decrease in performance that occurs half-way down the logistic curve (i.e., the inflection point). This inflection point is therefore based on complexity and simply means that the participants failed on sequences more than 50% of the time when complexity was above the inflection point. Table 2 Correlation matrix for Experiment 1. Note: Raven, Raven’s Advanced Progressive Matrices raw scores; SIMON, individual inflection points of performance on the chunking span task based on an adaptation of the SIMON game, using the individual logistic regression curve on the complexity axis; MU, memory updating; OS, operation span; SS, sentence span; SSTM, spatial short-term memory. MU, OS, SS, and SSTM are the four subtests of the Working Memory Capacity Battery (WM is the composite score obtained using the battery); ** p < .001 SIMON WM MU OS SS SSTM Raven .428** .437** .545** .297** .326** .406** SIMON _ .531** .572** .457** .376** .515** WM _ .630** .767** .824** .630** MU _ .499** .466** .506** OS _ .651** .374** SS _ .345** 30 A first glance at the correlation matrix shows that in terms of prediction of the Raven, the Simon is comparable to the composite WM score produced by the WMCB (respectively r = .428 and r = .437, and this range of correlations corresponds to that found in the literature). One important aspect to recall is that both MU and SSTM allow the stored items to be processed while both OS and SS are standard complex span tasks that separate processing and storage. Accordingly, performance on the Simon span task should better correlate with MU and also with SSTM. A second prediction was based on the idea that the storage × processing product (reflected by tasks in which the stored items are fully processed) would better predict the average score of the Raven. The correlation between MU and the Raven was effectively the highest (r = .572), and the Simon was the second task to better correlate with the Raven. The Simon also best correlated with both MU and SSTM. One difficulty is that the covariance between the tasks does not allow any task to correlate significantly better than another with the Raven, and the difference between this particular correlation and others was not found to be significant using Steiger's (1980) formula. We addressed this concern by conducting a principal component analysis to extract two factors (which were expected to separate a storage component from a processing component). Specifically, the factor model was rotated using varimax, after a preliminary verification of the Kaiser-Meyer-Olkin measure of sampling adequacy (.827 being a high value according to Kaiser) and Bartlett’s test of sphericity (χ2(15) = 180, p < .001), providing evidence that the patterns of correlations would yield reliable distinct factors. Factor loadings for Raven, Simon, MU, OS, SS, and SSTM were respectively .79, .73, .74, .28, .20, and .75 on the first component and .08, .35, .39, .85, .88, and .21 on the second component (see Figure 7A), and the two components accounted for 40% and 30% of the variance respectively (the respective sums of squared loadings being 2.4 and 1.8, instead of the 3.3 and .89 eigenvalues for the 31 unrotated initial solution). Oblique rotation of the factors (using the direct oblimin method), produced a greater separation of the tasks on the two factors on the component plot, with factor loadings for Raven, Simon, MU, OS, SS, and SSTM being respectively .86, .72, .72, .08, -.02, and .77 on the first component and -.15, .16, .21, .86, .91, and .00 on the second component (see Figure 7B). We interpreted the two factors as clearly separating the complex span tasks (in which processing is estimated alone, while processing is saturated) and the tasks in which processing was dedicated to storage, but it is still difficult to see how the processing and storage components are separated in these analyses by the respective factors. A B 32 Figure 7. Resulting component plot in rotated space for Exp. 1 from the exploratory factor analysis using PCA and varimax (subplot A) or Oblimin (subplot B). OS, operation span; SS, sentence span; SIM, chunking span task based an adapted version of the SIMON game; SSTM, spatial short-term memory; MU, memory updating; Raven, Raven’s APM. To better estimate the relationship between storage, processing, IQ and span tasks, the data were submitted to confirmatory factor analysis (CFA) using IBM SPSS AMOS 21. A latent variable representing a construct in which storage and processing are separated during the task and another latent variable representing a construct in which both processes interact (the processing component) were sufficient to accommodate performance. The fit of the model shown in Figure 8A was excellent (χ2(7) = 2.85, p = .90; CFI, comparative fit index = 1.0; RMSEA, root mean squared of approximation = 0.0; RMR, root-mean square residual = .002; AIC and BIC criterions were both the lowest in comparison to a saturated model with all the variables correlated with one another and an independence model with all the variables uncorrelated), with a caveat that the data fails to fit the recommended conditions for 33 computing a CFA (see Conway et al., 2002, Wang et al., 2015), particularly because one factor is defined by only two indicators and because the correlation between the two factors reveal collinearity issues. These results suggest that the Raven is better predicted by the construct in which storage and processing are combined (r = .64, corresponding to 41% of shared variance, instead of r = .36 when separated), a construct that can be reflected in the present study by our chunking span task, a memory updating task, and a simple span task. To make sure that the absence of SSTM would not weaken the predictions too much, we tested a similar model that did not integrate the SSTM task (Figure 8B). In this case, we observed increased regression weights towards the Raven (.85 instead of .80, and -.26 instead of -.22, and a percentage of shared variance with the Raven of 43% instead of 41%; (χ2(3) = 1.22, p = .75). Figure 8C shows that when a factor already reflects all of the span tasks, a second factor reflecting only the tasks for which Storage and Processing are supposed to be combined still account independently for the Raven. The substraction of the less restrictive model (Figure 8A) from the more restrictive model (Figure 8D) showed that constraining the two paths to the Raven to be equal significantly reduced the model's ability to fit the data (χ2 (9 - 7 = 2) = 54.95 - 2.85 = 52.1), meaning that the .80 loading in Model A can be considered significantly greater than -.22. 34 A 36< +8 ?! 345 3<> !"#$%&#'('% 367 !! +: 93:: 3>> 367 3<5 +7 )#*+$' ?@A+B 34; !D= 36< 355 +> 348 =C 3>7 355 ! #$, ./012$+, !!&= +< B #$% /6 !"#$%&#'('% !" #B& #&& "7-089-1:1 #$2 "" /) C#)+ )#*+$' #2+ #$& ,-./01 #B& #&) /% #$) "'( #+$ /2 !34/5 #B) "8-0;8< =>?@A0/; (* C )#*+$' 35 347 +C ;! 38: 367 !"#$ &#'(' 345 !! +B 354 !"#$%&#'('% 366 377 3:9 )#*+$' 366 +5 37: !@? 367 369 )#*+$' 3:6 +7 367 ?A 377 ! #$, ./012$+, !!&? +6 378 ;<=+> D 4:, +, =! 4,: 46; !"#$%&#'('% 476 !! +C 4;; 48 , 45< )#*+$' 4<8 +6 =@A+B 47, !>? 479 475 +< 456 ?D 4<6 478 ! #$- . /0123$+- !!&? +8 E A B C D χ2 2.85 (7), p = .9 1.24 (3) p = .75 2.72 (5), p = .74 54.95 (9), p = 0 CFI 1 1 1 .735 RMSEA 0 0 0 .245 RMR .002 .020 .020 .216 )#*+$' 36 Figure 8. Path models from confirmatory factor analysis from Exp. 1 with (A) and without (B) the SSTM task. For comparison, a third model (C) used a factor reflecting all of the span tasks versus another factor reflecting only the three tasks in which storage and processing were combined in the task. A fourth model (D) further constrained the parameters between the Raven and the latent variables to 1 (dotted lines), and a table (E) recapitulates the fit of each of the models. OS, operation span; SS, sentence span; SIM, chunking span task based on an adaptation of the SIMON game; SSTM, spatial short-term memory; MU, memory updating; S and P, storage and processing; Raven, Raven’s APM. The numbers above the arrows from a latent variable (oval) to an observed variable (rectangle) represent the standardized estimates of the loadings of each task onto each construct. The paths connecting the latent variables (ovals) to each other represent their correlation. Error variances were not correlated. The numbers above the observed variables are the R-squared values. Discussion The objective of the first experiment was to investigate a span task integrating a chunking factor in order to manipulate the storage × processing interaction process. Our experimental setup was developed to allow the participants to chunk the to-be-remembered material in order to optimize storage. The chunking factor was estimated using a compressibility metric, to follow up on a previous conceptualization that a chunking process can be viewed as producing a maximally compressed representation (Mathy & Feldman, 2012). Using an algorithmic complexity metric for short strings, we acknowledge the difficulty in representing the chunking process precisely sequence by sequence and we preferred to adopt a more general approach stipulating that any regularity should be found and recoded by participants, particularly those with the greatest processing capacity. The 37 hypothetical storage × processing construct in which processing is fully dedicated to storage, represents the largest difference with other constructs reflected by other span tasks since 1) processing is separated from storage in complex span tasks, and 2) processing is moderately involved in simple span tasks. One further hypothesis was that this compound variable should help characterize the nature of the relationship between WM capacity and general intelligence. Taken together, our results suggest that chunking opportunities (for instance, when fewer colors are used to build a sequence, resulting in low complexity) favors the recall process, and more interestingly the advantage of having more repetitions of colors interacting with sequence length. This interaction shows that the chunking factor best applies to longer sequences, a result that recalls the observation made in the Introduction section that longer sequences in simple span tasks (i.e, > 5 or 6 items) are more likely to reveal a higher processing demand. More precisely, our complexity metric plainly predicts a decline in performance as sequences become more complex, and a systematic decreasing trend across sequence lengths was observed. Although length had a detrimental effect on recall, the effect of complexity was found stronger than the effect of length. Capacity (i.e., how the participants could deal with complexity) was then estimated by an individual inflection point along the complexity axis of accuracy performance. This variable was found to correlate with the Raven, which tends to indicate that the participants who show higher storage × processing capacity tend to obtain higher scores on the Raven. The correlation is comparable to the one obtained from a composite measure of WM capacity (using the WMCB). However, an exploratory analysis showed that the chunking span task saturated two principal factors in a way very similar to the Raven and to other tasks in which processing is also dedicated to storage (involving updating or short-term memory), contrary to complex span tasks which are usually one of the most famous tasks to estimate fluid intelligence. A confirmatory factor 38 analysis confirmed the greater predictability of the Raven with a latent variable in which the storage and processing components were combined, in opposition to a latent variable which represented the complex span tasks (in which storage and processing were separated). The correlation between the “Storage and Processing Combined” latent variable shares 41% of the variance with the Raven, which seems quite satisfying compared to 50% obtained by Kane, Hambrick, and Conway (2005) from more than 3000 participants and 14 different data sets to estimate the shared variance between WM capacity and general intelligence constructs (see also Ackerman, Beier, & Boyle, 2005, for a contrasting point of view). Our results also confirm previous findings that memory updating can be a good predictor of Gf (e.g., Friedman, Miyake, Corley, Young, Defries, & Hewitt, 2006; Schmiedek, Hildebrandt, Lövdén, Wilhelm, & Lindenberger, 2009; Salthouse, 2014), but here we considered that the updating task could be associated with a chunking span task under the same construct. Effectively, even though they seem to have nothing much in common, in both of these tasks processing is dedicated to storage. Another result showing a higher recall from high-IQ participants for the less complex sequences further emphasizes the link between general intelligence and the optimization of storage. Our results showed less variation of performance for the more complex sequences for both IQ groups. Although this first experiment was helpful to decipher the utility of a chunking span in the prediction of general intelligence, there was, however, a practical difficulty to make comparisons between low-complexity conditions and high-complexity conditions since the participants were not administered the same sequences. We expected a clear demarcation of IQ groups within the low-complexity condition, but low vs. high IQ groups were not clearly distinguished enough on the basis of sequence complexity (although our results still indicate a significant interaction between complexity and IQ when both variables were split in 39 quartiles). The second experiment attempts a more direct comparison of performance when the sequences are simple vs. when the sequences are more complex. Another weakness of Experiment 1 is that a specific scoring method was used to estimate the different spans. The second experiment attempts a more direct comparison of the simple vs. complex sequences, by use of a common scoring method. Experiment 2 Experiment 2 was set up to better separate the storage and the storage × processing components by using complex versus simple sequences respectively. Two versions of the chunking span task were created, one based on the most complex sequences of Experiment 1, and the second based on the simplest sequences of Experiment 1. The complex sequences targeted the storage-only demand because such sequences allow little or no compression. On the contrary, the simplest sequences targeted the storage × processing demand since the regularities could be used by participants to optimize storage. Also, Experiment 1 estimated capacity from an inflexion point obtained across individuals who were not administered the same sequences of colors. Another goal of the second experiment was to provide estimates of the span based on a similar set of sequences given to the participants in progressive order (from sequences of minimal length to the longest length achieved) to ensure a more proper evaluation of capacity. Performance on our chunking span tasks was compared to other simple span tasks in a commercial test (WAISIV) and again to a measure of general intelligence (Raven’s Advanced Progressive Matrices). Experiment 1 showed that complex span tasks were less correlated to a storage × processing construct, so we chose to focus on simple span tasks for which processing is partly dedicated to the storage process. It was hypothesized that the spans obtained from the Simon chunking span tasks would better correlate with the Raven. Effectively, the spans of the WAIS either 40 moderately involve processing (forward condition) or simply require linearly processing the stored items (backward or sequencing conditions). In the case of the backward and the sequencing conditions, processing cannot be regarded as subserving storage, because processing intervenes only after the items are stored (see Thomas, Milner, & Haberlandt, 2003, who showed that the sequences are initially stored before being processed). The span estimated during a chunking process is the only one that truly involves an optimization process based on the processing component being at the service of the storage component. If the span results from the participant’s storage capacity which hierarchically depends on an optimization process (storage capacity is supposed to be fixed for a given individual, but any form of structuration of the stimulus sequence can contribute to extending the number of stored items), we hypothesize that chunking span tasks will best correlate with intelligence. The rationale is that during the Raven, participants also use storage and processing in conjunction to solve the problems. Method Participants A total of 107 undergraduate students volunteered to participate in this experiment and received course credits in exchange for their participation (M = 22.9, SD = 5.9). All participants were administered a chunking span task and three subtests measuring the span from the WAIS-IV; 95 of the students agreed to be tested with the Raven’s APM (which was optional for getting course credits). The average IQ for this sample was 100.2 (SD = 13.4). Procedure 41 The tests were administered in the following order: the Simon chunking span task, the WM subtests of the WAIS-IV and finally, and optionally, the Raven. Simon chunking span task The procedure was designed to better match a standard memory span task in which a progression of difficulty is involved. Thus, the length of the presented sequences progressively increased, starting with length two, then three, etc. Two different sequences with a given length were presented (to match the number of repetitions per length in the WM subtests of the WAIS), and the session automatically stopped after the participant failed to correctly recall the two sequences of a given length. Scoring of the span followed that of the WAIS, in which the longest span attained at least once was considered as the subject’s span. Each participant was administered two complexity conditions (counterbalanced between participants). The sequences in each condition were taken from the 9150 sequences that were generated in Exp. 1. Based on the distribution of the sequences’ given complexity, all the sequences were ranked according to their complexity (the 100th percentile corresponded to the most complex sequences), and a series of sequences was chosen at the 50th or the 100th percentile to constitute the two experimental conditions (the sequences below the 50th percentile were judged too easy to constitute an interesting experimental condition). By construction, the Simple condition was conducive to inducing a chunking process, while the Complex condition allowed less chunking opportunities and as such was considered as mostly soliciting the storage component. WAIS-IV working memory subtests. The participants were administered three WM subtests from the WAIS-IV (Wechsler, 2008): the Digit Span Forward (DSF) which requires recalling a series of digits in correct order, the Digit Span Backward (DSB) which requires 42 recalling a series of digits in reverse order, and the Digit Span Sequencing (DSS) which requires recalling a series of digits in ascending order. The two last subtests clearly require mental manipulation of the stored items, in comparison to the first subtest which is considered the simpler short-term memory test. The scoring procedure followed the recommendation of the WAIS, the longest span attained at least once was considered the subject’s span. Note that although these tasks are supposed to index a WM construct in the WAIS, these subtests are closer to simple span tasks and might better reflect a STM construct (See Unsworth & Engle, 2007). Raven’s Advanced Progressive Matrices. As in Exp. 1, the participants were asked to complete the matrices of set #2 which contained 36 matrices after doing set #1 which was used as a warm-up. The range of possible raw scores was 0–36. Results To estimate the reliability of our task, we split the data into two groups (odd vs. even trial numbers) within each sequence length. Participants showed equivalent performance at the even and odd trials for each condition (Simple condition: r = .62, p < .001; SpearmanBrown coef. = .77; Complex condition: r = .59, p < .001; Spearman-Brown coef. = .74) and adequate evidence of reliability was obtained given such a short procedure. For comparison, we applied the same method to the digit span tasks, and we obtained estimates of reliability within the same range (Forward condition: r = .45, p < .05; Spearman-Brown coef. = .63; Backward condition: r = .66, p <.001; Spearman-Brown coef. = .78). The mean complexity of the recalled sequences was again lower than that of the to-beremembered sequences (t(2653) = 13.20, p < .001, Cohen’s paired d = 0.26). Once again, we 43 investigated the effect of length and complexity on perfect recall by way of logistic regression. A stepwise forward logistic model selection based on a BIC criterion suggested dropping the interaction term, as shown in Table 3. Table 3. Stepwise forward selection model based on BIC criterion. At each stage (model), the variables are listed in increasing BIC order. The final model includes complexity and length, but no interaction term. Base model Variable included Deviance BIC Intercept only Complexity 1672.9 1688.6 Length 1737.3 1753.0 None 2957.1 2965.0 Correct ~ complexity Length 1632.0 1655.6 None 1672.9 1688.6 Correct ~ Complexity and Length None 1632 1655.6 Interaction 1632 1663.5 A logistic regression on the 2654 trials was done on the proportion correct of perfectly recalled sequences as a function of complexity and length (see Figure 9). It confirmed the detrimental effect of complexity on recall (z(2651) = 9.77, p < .001, standardized coefficient = -6.41 [non standardized: -.95]) and also an effect of length (z = 6.273, p < .001, standardized coefficient = 3.94 [non standardized: 2.01948]). Overall this first result indicates 44 that the memory process could effectively be optimized for less complex sequences. One interesting point is that the inflection curve is close to the one obtained in Experiment 1 (20.2 in comparison to 21.7 for Experiment 1), which means that allowing a greater amount of time to participants did not lead them to increase performance much. Still this difference was found statistically reliable (t(270) = 4.5, p < .001; Cohen’s d = .50), although this difference must be tempered by the fact that the sequences of Exp. 2 were in general shorter (4.7 colors per sequence instead of 6.2 colors in Experiment 1, t(5497) = 33, p < .001; Cohen’s d = .63). Also, the difference between the two inflection points only reflected a difference of .7 colors recalled (7 in Experiment 1 and 6.3 in Experiment 2). 45 Figure 9. Proportion correct as a function of complexity of the Simon span tasks (with the conditions Simple and Complex mixed) in Exp. 2. Error bars are +/- one standard error. The decreasing mean spans reported in Figure 10 for DSB, Simon Complex, DSF, DSS, and Simon Simple are respectively 5.1 (SD = 1.3), 5.6 (SD = 1.0), 6.5 (SD = 1.1), 6.7 (SD = 1.4) and 6.8 (SD = 1.0). A repeated-measures ANOVA on the mean span was conducted with the type of span task as a within-subjects factor (Simon Simple, Simon Complex, DSF, DSB, DSS) on the data collapsed by participants. The result of the ANOVA was significant (F (4,424) = 68.7, p < .001, ŋp² = .39), indicating that almost 40% of the variance was accounted for by the type of span task factor. 46 Figure 10. Mean span as a function of type of span task in Exp. 2. Note: Simple, Simon chunking span task, simple version; DSS, Digit Span Sequencing; DSF, Digit Span Forward; Complex, Simon chunking span task, complex version; DSB, Digit Span Backward. Posthoc analysis (Bonferroni corrections were made for the pairwise comparisons) showed significant systematic pairwise differences except between Simon Simple and DSS, and between DSS and DSF. A noteworthy difference was found between the two respective mean spans for the two conditions in the chunking span tasks (M = 1.25; SD = 1.1), but surprisingly, the small difference of 1.25 on average means that between one and two additional colors were grouped on average within the simpler sequences. Table 4 Correlation matrix for Experiment 2. Note: Raven, Raven’s Advanced Progressive Matrices raw scores; COMPL, mean span on the Simon chunking span task in the complex condition (with chunking opportunities reduced); SIMPL, mean span on the Simon chunking span task in the simple condition (with chunking opportunities induced); DSF, Digit Span Forward; DSB, Digit Span Backward; DSS, Digit Span Sequencing; ** p < .001 COMPL DSF DSB DSS Raven SIMPL .422** .294** .337** .157 .413** COMPL .229* .353** .310** .385** DSF .473** .273** .290** DSB .476** .446** DSS .297** 47 Despite the moderate difference between the two mean spans observed between the simple and complex conditions, Table 4 shows that these two conditions highly correlated (r = .42), in comparison with other variables. Similarly, DSF and DSB shared the greatest percentage of variance (r = .47), as well as DSB and DSS (r = .48). Thus the digit spans showed high mutual correlation, but none of the digit span tasks correlated more with either Simon simple or Simon Complex than the two together. One possibility is that the participants could still chunk many of the most complex sequences, making the two Simon conditions akin, and accounting for the slight difference of 1.25 colors reported above. The possibility that participants chunk the less compressible sequences does not contradict compressibility theories because the estimate of the compressibility of a string is averaged across possible Turing machines. For a given complex string, a particular machine can still be able to perform compression. Another possibility is that participants verbalized the tasks in both conditions, and pronunciation of the sequences of colors was limited by the duration of rehearsal. In that case, chunking would not have been as beneficial as expected if only the number of colors to be pronounced determined capacity. Regarding correlations with the Raven, the highest correlation was found with DSB, but the multicollinearity of the data makes interpretation of the pairwise correlations difficult. Principal component analysis was used to explore our data and to extract two factors (which were expected to separate the chunking span tasks and the WM span task), and the factor model was then rotated using varimax, after a preliminary verification of the Kaiser-MeyerOlkin measure of sampling adequacy (.796 being a high value according to Kaiser) and Bartlett’s test of sphericity (χ2(15) = 123, p < .001) to indicate if the patterns of correlations would yield reliable distinct factors. Factor loadings for COMPL, SIMPL, DSF, DSB, DSS, and Raven were respectively .21, .09, .78, .77, .71, and .33 on the first component and .75, 48 .83, .10, .32, .20, and .66 for the second component (see Figure 11A), and the two components accounted for 31% and 31% of the variance respectively (the respective sums of squared loadings being 1.9 and 1.8, instead of the 2.8 and .93 eigenvalues for the unrotated initial solution). Oblique rotation of the factors (using the direct oblimin method) produced a greater separation of the tasks on the two factors on the component plot. Factor loadings for COMPL, SIMPL, DSF, DSB, DSS, and Raven were respectively .04, -.11, .83, .75, -.72, and .19 on the first component and .76, .88, -.10, .15, .04, and .63 for the second component (see Figure 11B). The two factors clearly separated the digit span tasks and the chunking span tasks. It is worth noting that the Raven loaded with the second set of tasks. A B 49 Figure 11. Resulting component plot in rotated space for Exp. 2 from the exploratory factor analysis using PCA and Varimax (subplot A) or Oblimin (subplot B). Note: SIMPL, Simon chunking span task, version Simple; COMPL, Simon chunking span task, version Complex; DSF, DSB and DSS, Digit Span Forward, Backward and Sequencing (WAIS-IV); Raven, Raven’s Advanced Progressive Matrices. The data were submitted to a confirmatory factor analysis using IBM SPSS AMOS 21 in order to test the prediction that tasks allowing the processing and storage components to fully function together in association to optimize storage are better predictors of general intelligence than the STM span tasks of the WAIS. A latent variable representing a chunking construct (derived from the Simon span tasks) and another latent variable representing a simpler STM construct (derived from the digit span tasks of the WAIS9) were sufficient to 9 Again, we chose to label the latent variable STM instead of WM because the digit span tasks can be seen as simple span tasks (See Unsworth & Engle, 2007). 50 accommodate performance. The fit of the model shown in Figure 12 was excellent (χ2(7) = 3.2, p = .87; CFI, comparative fit index = 1.0; RMSEA, root mean squared of approximation = 0.0; RMR, root-mean square residual = .049; AIC and BIC criterions were both the lowest in comparison to a saturated model with all the variables correlated with one another and an independence model with all the variables uncorrelated), but again, as for Figure 8, with a caveat that the data fails to fit the recommended conditions for computing a CFA (because one factor is defined by only two indicators and because the correlation between the two factors reveal collinearity issues. These results suggests that the Raven is best predicted by the Chunking latent variable, a construct that can be reflected in the present study by the two chunking span tasks. Figure 12B shows that a factor reflecting only the tasks for which Storage and Processing are supposed to be combined still account independently for the Raven once a first factor reflects all of the span tasks. The substraction of the less restrictive model (Figure 8C) from the more restrictive model (Figure 8A) showed that constraining the two paths to the Raven to be equal significantly reduced the model's ability to fit the data (χ2 (9 - 7 = 2 ) = 16.65 – 3.24 = 13.41), meaning that the .42 loading in Model A can be considered significantly greater than -.27. A 51 (." !" ,3<=>! (6. (.6 /01&23&4 (67 /?<=>!@ !- (.()* #$%!&' (6- BC0!D ()) (-; +,A !) (87 (67 !. +,5 ()) !8 ,9: (7) (8; +,, B ,&. $. )2345$ ,-. ,'9 6#A=B2=C ,// 67345$8 $( ,'' ,'9 ,'1 ,&0 ,-0 $/ ?)D ,'' $& !"#$% ,&' ?)@ $' :;<$=> ,/& ,'' )*+ ,0( ,&0 ?)) C 52 )*+ !3 $%&'(! )+2 )*/ )+2 ,@C9D%9E ,-&'(!. !B )3+ );2 )+* 678!9: )"= )3+ 4$< !" )+" )+# !* 4$5 )"* !# >?@!A $01 )23 )#2 4$$ D A B C χ2 3.24 (7), p = .86 2.69 (6), p = .85 16.65 (9), p = .05 CFI 1 1 .93 RMSEA 0 0 .25 RMR .003 .003 1 Figure 12. Path models for confirmatory factor analysis from Exp. 2. An initial model (A) is compared with two other models: In (B), we used a factor reflecting all of the span tasks versus another factor reflecting only the three tasks in which storage and processing were combined in the task; In (C), we further constrained the parameters between the Raven and the latent variables to 1 (dotted lines). The table (D) recapitulates the fit of each of the models. All the correlations and loadings are standardized estimates. The numbers above the observed variables are the R-squared values. Simple, Simon Simple; Complex, Simon Complex; DSF, DSB and DSS, Digit Span Forward, Backward and Sequencing (WAIS-IV); Raven, Raven’s Advanced Progressive Matrices; WM, traditional working memory construct; Chunking, chunking construct 53 Discussion The main goal of the second experiment was to provide several estimates of the STM span using similar scoring method. Performance on our chunking span tasks was therefore compared to three other simple span tasks (Forward, Backward and Sequencing digit span tasks of the WAIS-IV) using the scoring method provided by the WAIS. The chunking span task was hypothesized to best correlate with the Raven because both require to make sense of complexity while storing information, a process that might require optimizing storage capacity to free up memory and leave room for more complex reasoning. Experiment 2 was also devised to better separate the storage and the storage × processing demands by allowing two degrees of involvement of the processing component: low for the less compressible sequences that constituted the first condition (Complex version of the SIMON chunking span task) and high for the second condition that included more compressible sequences (Simple version). The average span was found very close in the two conditions, with an average difference of only 1.2 additional colors recalled in the Simple condition of the chunking span task than in the Complex condition. Although this difference was low, the correlation of the Simple version still slightly better correlated with the Raven than the Complex version. The verbalization of the colors could account for the fact that the span revolved around 7 for both the less compressible and the most compressible sequences, as if the participants were rehearsing the sequences of colors verbally without taking profit of the mathematical regularities (this is acknowledged as a limitation of our study in the General Discussion). Nevertheless, when the two tasks were used to estimate a latent variable covering a chunking process, the prediction of the Raven was found larger than for the three digit span tasks. General Discussion Chunking and intelligence 54 Our goal was to develop a chunking span task in which storage and processing can fully to study its relation with intelligence. The quantity of chunking was estimated by computing the algorithmic complexity of the to-be-remembered series of colors appearing one after the other in a task inspired by the Simon game. The modification of this quantity allowed us to slide across the [(storage) ... (storage + processing)] dimension as we posited that more chunking opportunities should involve a greater amount of processing. The chunking process was also thought to avoid separating the memory contents from the material to be processed. We then addressed one primary question: Is chunking a reliable predictor of intelligence when chunking is implemented in a simple span task? To answer this question, we used confirmatory factor analysis and structural equation modeling with 290 young adults to test hypotheses about the nature of the WM and STM constructs and their relations to general intelligence. The nonconventional idea of the present study is that our chunking span tasks are simple span tasks that can target a WM construct. We hypothesized that chunking and intelligence both rely on individuals' storage capacity, which hierarchically depends on an optimization process (see also Duncan et al., 2012). Storage capacity is supposed to be fixed for a given individual, but any form of structuration of the stimulus sequence can contribute to extending the number of stored items. This optimization process could be helpful to make sense of the regularities that can be found in the items of the Raven or in the to-be-recalled sequences of repeated colors. The rationale was that participants use storage and processing in conjunction to solve the problems of the Raven and to chunk the colors of the to-be-remembered sequences of repeated colors, or, as put forward by Duncan et al. (2012, p. 868), fluid intelligence can be used in “the organizational process of detecting and using novel chunks”. 55 In two experiments, we found that the chunking spans were structurally closer to the performance on the Raven than any complex span task of the WMCB and any of the simple span tasks of the WAIS. The memory updating tasks and the spatial STM task of the WMCB had more shared variance with the chunking span tasks and the Raven, suggesting that the Raven is best predicted when processing and storage function together in a span task. This confirms a result by Mathy and Varré (2013) who showed that a span task using alphanumeric lists that included a few repetitions of items (thus inducing a chunking process) better correlated with the memory updating task of the WMCB than similar task that included no repetition of items. One important result is that our chunking span tasks compete with the well-known complex span tasks in which the to-be-recalled items are interspersed with other activities unrelated to the retention of the items (e.g., Kane, Hambrick, & Conway, 2005). This result was refined in Exp. 2 which showed that the simple span tasks better account for the Raven when they involve a chunking factor than when the simple span tasks do not integrate any possible chunking processes. The present study shows that simple span tasks can effectively compete with complex span tasks, and this was achieved here by prompting the creation of chunks in immediate memory while avoiding a long-term learning effect. These findings are broadly consistent with previous studies (see Unsworth & Engle, 2007a) suggesting that prediction of fluid intelligence by simple spans reached that of complex spans by increasing list-lengths. Another example where reorganization seems to explain the high correlation between simple span and intelligence is the conceptual span task (Haarmann, Davelaar, & Usher, 2003). In this task, words are presented successively as in a classic word span, except that the words are drawn from three semantic categories that are used to aid recall of the words. Haarmann et al. (2003) observed that the conceptual span correlated better with reasoning than the reading span. According to the authors, this was due 56 to the fact that this simple span was supposed to tap the construct of semantic STM (e.g., Hanten & Martin, 2000; Martin & Romani, 1994). However, Kane and Miyake (2007) later suggested that the capacity of the conceptual span (semantic or not) to correlate with intelligence or high cognitive processes depends on a clustering ability. Effectively, when the to-be-remembered words were presented in a clustered fashion, the conceptual span task lost its ability to predict intelligence compared to an unclustered version. Therefore, it seems likely that the span tasks better correlate with higher cognitive processes when they prompt reorganization of information. The present study concludes that processing and storage should be examined together when processing is fully dedicated to the stored items, and we believe that the interaction between storage and processing that best represents chunking provides a true index of WM capacity. Thus, among the span task taxonomy that was discussed in the Introduction section, the chunking span seems to be a good candidate for predicting intelligence and measuring WM. This is in line with Unsworth et al. (2009) who argue that processing and storage should be examined together because WM is capable of processing and storing information simultaneously. Effectively, in complex spans, the interaction between storage and processing has also been put forward, particularly by Towse, Cowan, Hitch and Horton (2008) in the recall reconstruction hypothesis. This hypothesis was based on observations made with the reading span task (e.g., Cowan et al., 2003; Hitch, Towse, & Hutton, 2001) which showed that the sentences (the processing part of the task) can be used to reconstruct the target words (the storage part of the task). This hypothesis supports the study by Cowan et al. (2003) in which long pauses were often needed to recall the words. Towse et al.’s (2008) hypothesis puts forward two important assertions: 1) processing and storage need not always be thought of as completely separate events, and 2) processing and storage 57 are not necessarily competitive components of WM tasks that, for instance, compete for cognitive resources (e.g., Barrouillet, Bernardin, & Camos, 2004; Barrouillet, Plancher, Guida, & Camos, 2012; Case, 1985) or interfere with each other (Oberauer, Farrell, Jarrold, Pasiecznik, & Greaves, 2012; Oberauer, Lewandowsky, Farrell, Jarrold, & Greaves, 2012; Saito & Miyake, 2004). As highlighted and synthesized by the recall reconstruction hypothesis (see also Towse, Hitch, Horton, & Harvey, 2010), processing and storage can also interact synergistically (e.g., Cowan et al., 2003; Hitch et al., 2001; Guida, Tardieu, & Nicolas, 2009; Osaka, Nishizaki, Komori, & Osaka, 2002; Schroeder, Copeland, & BiesHernandez, 2012). Interestingly, this has not been observed in all WM tasks. For instance, Cowan et al. (2003) failed to observe this interaction in a digit WM task. Such an interaction has only been observed so far in reading span tasks in which, as suggested by Cowan et al. (2003), participants can retrieve the semantic or linguistic structure and use it as a cue to recall the sentence-final words. Therefore, the linguistic domain seems more prone to induce an interaction between storage and processing. Our chunking span shows that it is also possible to induce this interaction outside of the linguistic realm even if it is done with a simple span as in our case. Uncontrolled chunking ? One risk of devising a task in which processing is fully dedicated to storage is putting aside uncontrolled specific strategies for simply remembering more information. One caveat has been that the elaboration of memoranda into meaningful chunks inflates capacity, and because this strategy can be more effective for some subjects, "allowing such strategies to determine variance in the capacity measure may obscure real relationships between storage capacity and other measures" (Fukuda, Vogel, Mayr, and Awh, 2010, p. 679). Effectively, 58 strategies might artificially improve a person’s measured span. However, the present study is based on the assumption that one way to control specific individual strategies is to allow people to develop the same strategies by introducing systematic regularities into the memoranda (without asking them to do something in particular). In the present study, these regularities were fit by an algorithmic complexity metric. Our results show that the average complexity reached by the participants between Exp. 1 and Exp. 2 is quite close when considering the two inflection points in performance as a function of complexity. This difference was still statistically significant, but corresponded to only .7 additional colors recalled in Experiment 1 (which lasted about 25 minutes) than in Experiment 2 (which lasted about 5 minutes). This confirms the observation of Gendle and Ransom (2006) that the Simon game is resistant to practice effects. This also means that the average spans between the two experiments were not influenced much by elaborated strategies that could have been developed better while allowing a greater amount of time to participants. One remaining risk is that intelligence might be best measured by chunking span tasks because these tasks require intelligence in the first place. We believe that the optimization process that goes with associating the colors one another is an elementary process that can only be a basis for developing higher order cognition, but intelligence cannot be reduced to such an elementary process. Intelligence could therefore be predicted by such tasks without totally underlying the perception of patterns and they reorganization in immediate memory. Capacity limit One last interesting result of the present study is that our chunking span tasks lead to greater spans than the 4±1 capacity usually associated with the WM construct, and do not conclude with a trade-off between processing and storage. One point of view is that any span task with extra processes (chunking in the present study) would necessarily lead to lower spans in the 59 case of a trade-off between processing and storage (see Logan, 2004. p. 219). Another point of view is that more processing is supposed to be associated with better compression and should result in a greater number of items stored here. This is only true if the recoding process does not take more space than necessary. Effectively, more encoding does not necessarily involve shorter representations. For instance, and to simplify the idea, blue-red-blue-blueblue-red (or brbbbr) can be elegantly recoded as xbbx using x=br as a new chunk, simply because br is redundant. However, this supposes that brbbbr is replaced with the longer "x=br;xbbx" because the recoding of br into x needs to be remembered anyhow. The resulting compressed representation xbbx cannot be dissociated from the compression process that includes x=br. The new description x=br;xbbx is now 10-symbols long instead of the original sequence brbbbr which is 7-symbols long. A compressed sequence should better optimize storage, but the method is only useful when the recoding scheme helps gain space. This is not the case for this example. The gain can be estimated using an information-theoretic criterion such as the Minimum-Description-Length (MDL) principle (Robinet, Lemaire, & Gordon, 2011). For instance, the method would work for a longer original sequence such as brbbbrbrbrbrbrbr (16 symbols) because x=br;xbbxxxxxx (14 symbols) is shorter. The new code "x=br;xbbxxxxxx" could be further maximally compressed using x=br;xbb6x but again, a new code needs to specify that a number symbol (e.g., 6) indicates a repetition. Therefore, short strings can still be recoded while there is a risk of increasing the original length. Whether it leads to a longer or a shorter description, the recoding process could have a direct relationship with the level of processing principle (Craik and Lockhart, 1972), which states that memory trace is a function of the depth to which processing is performed. In the present case, recoding a sequence with a longer description but with a more structured representation could still help the recall process. 60 Limitations We present here, one limitation concerning our study. In Experiment 2, the results showed an average difference of only 1.2 more colors recalled in the Simple condition of the chunking span task than in the Complex condition. When the two spans were divided by one another to estimate processing efficiency, we found a negative correlation between processing and storage. This means that the greater the capacity for the most complex sequences (e.g., 7) the less it is likely that the participants can increase their capacity for the simpler sequences (e.g., the span still revolves around 7). Notwithstanding the possibility that some participants have a great storage capacity but a null processing efficiency because they do not obtain better results in the Simple condition than in the Complex condition, a first more plausible interpretation is that some participants make more effort to organize the less compressible sequences, which does not leave much room for the most compressible sequences. Another possible interpretation is that the limitation is due to the use of phonological representations that limit the duration of rehearsal. No matter the compressibility of the sequences, the number of colors to be pronounced covertly would, for instance, be limited by the two-second limit on the phonological store, and indeed many participants declared that they spontaneously verbalized the to-be-recalled sequences. One last test was therefore conducted by dividing the two spans obtained for the simple and complex versions of the chunking span task for each participant. This aimed at estimating the number of items that could be packed by the participants, which reflected their processing capability (this directly measures an average chunk-creation ability). In this case, the 61 correlation between this estimate of the processing component and Raven was not found significant (r = -.04) while the correlation between processing was found negative with storage (r = -.59, p < .001) and positive with storage × processing (r = .46, p < .001). This means that the greater the capacity was for the most complex sequences (e.g., 7) the less likely it was that the participants could increase their capacity for the most simple sequences (e.g., the span still revolved around 7). Overall, our tasks do not seem to be able to express as simply as we first expected three clearly separated estimates of storage, storage × processing, and storage × processing/ storage = processing constructs. Conclusion The rationale of the present study was that sequences of colors of the Simon game contain regularities that can be mathematized to estimate a chunking process, and that the quantity of chunking induced in a to-be-recalled sequence can represent the processing demand. The chunking span task allows the processing and storage components to fully interact to optimize storage. Although it is not commonly accepted in the literature that span tasks can take benefit from favoring the processing of the stored items (which explains the plethora of complex span tasks in the literature), the chunking span task was found a reliable predictor of general intelligence in comparison to other simple or complex span tasks. Our conclusion is that chunking span tasks can predict IQ better than any other types of span tasks because this task involves a process of storage optimization that requires full function of the processing component. 62 63 References Aben, B., Stapert, S., & Blokland, A. (2012). About the distinction between working memory and short-term memory. Frontiers in Psychology, 3. Ackerman, P. L., Beier, M. E., & Boyle, M. O. (2005). Working memory and intelligence: The same or different constructs? Psychological Bulletin, 131(1), 30–60. Alvarez, G. A., & Cavanagh, P. (2004). The capacity of visual short-term memory is set both by visual information load and by number of objects. Psychological Science, 15(2), 106–111. Anderson, J. R., Bothell, D., Lebiere, C., & Matessa, M. (1998). An integrated theory of list memory. Journal of Memory and Language, 38(4), 341–380. Baddeley, A. D. (2000). The episodic buffer: a new component of working memory? Trends in Cognitive Sciences, 4(11), 417–423. Baddeley, A. D. (2001). Is working memory still working? American Psychologist, 56(11), 851. Baddeley, A. D., & Hitch, G. (1974). Working memory. Psychology of Learning and Motivation, 8, 47–89. Baddeley, A. D., Lewis, V., & Vallar, G. (1984). Exploring the articulatory loop. The Quarterly Journal of Experimental Psychology Section A, 36(2), 233‑252. Baddeley, A. D., Thomson, N., & Buchanan, M. (1975). Word length and the structure of short-term memory. Journal of Verbal Learning and Verbal Behavior, 14(6), 575– 589. Bailey, H., Dunlosky, J., & Kane, M. J. (2011). Contribution of strategy use to performance on complex and simple span tasks. Memory & Cognition, 39(3), 447–461. 64 Barrouillet, P., Bernardin, S., & Camos, V. (2004). Time constraints and resource sharing in adults’ working memory spans. Journal of Experimental Psychology: General, 133(1), 83. Barrouillet, P., Bernardin, S., Portrat, S., Vergauwe, E., & Camos, V. (2007). Time and cognitive load in working memory. Journal of Experimental Psychology: Learning, Memory, and Cognition, 33(3), 570. Barrouillet, P., & Camos, V. (2012). As time goes by temporal constraints in working memory. Current Directions in Psychological Science, 21(6), 413–419. Barrouillet, P., Plancher, G., Guida, A., & Camos, V. (2013). Forgetting at short term: When do event-based interference and temporal factors have an effect? Acta Psychologica, 142(2), 155–167. Baum, E. B. (2004). What is thought? Cambridge, MA: MIT press. Bor, D., Cumming, N., Scott, C. E., & Owen, A. M. (2004). Prefrontal cortical involvement in verbal encoding strategies. European Journal of Neuroscience, 19(12), 3365–3370. Bor, D., Duncan, J., Wiseman, R. J., & Owen, A. M. (2003). Encoding strategies dissociate prefrontal activity from working memory demand. Neuron, 37(2), 361–367. Bor, D., & Owen, A. M. (2007). A common prefrontal–parietal network for mnemonic and mathematical recoding strategies within working memory. Cerebral Cortex, 17(4), 778–786. Bor, D., & Seth, A. K. (2012). Consciousness and the prefrontal parietal network: insights from attention, working memory, and chunking. Frontiers in Psychology, 3, 63. Botvinick, M. M., & Plaut, D. C. (2006). Short-term memory for serial order: a recurrent neural network model. Psychological Review, 113(2), 201. Boucher, V. J. (2006). On the function of stress rhythms in speech: Evidence of a link with grouping effects on serial memory. Language and Speech, 49(4), 495–519. 65 Brady, T. F., Konkle, T., & Alvarez, G. A. (2009). Compression in visual working memory: using statistical regularities to form more efficient memory representations. Journal of Experimental Psychology: General, 138(4), 487. Brady, T. F., Konkle, T., & Alvarez, G. A. (2011). A review of visual memory capacity: Beyond individual items and toward structured representations. Journal of Vision, 11(5), 4. Brady, T. F., & Tenenbaum, J. B. (2013). A probabilistic model of visual working memory: Incorporating higher order regularities into working memory capacity estimates. Psychological Review, 120(1), 85. Brainard, D. H. (1997). The psychophysics toolbox. Spatial Vision, 10, 433–436. Burgess, N., & Hitch, G. J. (1999). Memory for serial order: A network model of the phonological loop and its timing. Psychological Review, 106(3), 551. Cantor, J., Engle, R. W., & Hamilton, G. (1991). Short-term memory, working memory, and verbal abilities: How do they relate? Intelligence, 15(2), 229–246. Case, R. (1985). Intellectual development: Birth to adulthood. New York, NY: Academic Press. Chaitin, G. J. (1966). On the length of programs for computing finite binary sequences. Journal of the ACM, 13(4), 547–569. Chase, W. G., & Simon, H. A. (1973). Perception in chess. Cognitive Psychology, 4(1), 55– 81. Chen, Z., & Cowan, N. (2005). Chunk limits and length limits in immediate recall: A reconciliation. Journal of Experimental Psychology. Learning, Memory, and Cognition, 31(6), 1235‑1249. Chuderski, A., Taraday, M., Nęcka, E., & Smoleń, T. (2012). Storage capacity explains fluid intelligence but executive control does not. Intelligence, 40(3), 278-295. 66 Colom, R., Abad, F. J., Quiroga, M. Á., Shih, P. C., & Flores-Mendoza, C. (2008). Working memory and intelligence are highly related constructs, but why?. Intelligence, 36(6), 584-606. Colom, R., Jung, R. E., & Haier, R. J. (2007). General intelligence and memory span: evidence for a common neuroanatomic framework. Cognitive Neuropsychology, 24(8), 867–878. Colom, R., Rebollo, I., Abad, F. J., & Shih, P. C. (2006). Complex span tasks, simple span tasks, and cognitive abilities: A reanalysis of key studies. Memory & Cognition, 34(1), 158–171. Colom, R., Rebollo, I., Palacios, A., Juan-Espinosa, M., & Kyllonen, P. C. (2004). Working memory is (almost) perfectly predicted by g. Intelligence, 32(3), 277-296. Conway, A. R., Cowan, N., Bunting, M. F., Therriault, D. J., & Minkoff, S. R. (2002). A latent variable analysis of working memory capacity, short-term memory capacity, processing speed, and general fluid intelligence. Intelligence, 30(2), 163–183. Conway, A. R., Kane, M. J., Bunting, M. F., Hambrick, D. Z., Wilhelm, O., & Engle, R. W. (2005). Working memory span tasks: A methodological review and user’s guide, Psychonomic Bulletin & Review, 12(5), 769–786. Conway, A. R., Kane, M. J., & Engle, R. W. (2003). Working memory capacity and its relation to general intelligence. Trends in cognitive sciences, 7(12), 547-552. Conway, C. M., Karpicke, J., & Pisoni, D. B. (2007). Contribution of implicit sequence learning to spoken language processing: Some preliminary findings with hearing adults. Journal of Deaf Studies and Deaf Education, 12(3), 317–334. Cowan, N. (2001). The magical number 4 in short-term memory: a reconsideration of mental storage capacity. The Behavioral and Brain Sciences, 24(1), 87‑114. Cowan, N. (2005). Working memory capacity. Hove, East Sussex, UK: Psychology Press. 67 Cowan, N., & Chen, Z. (2009). How chunks form in long-term memory and affect short-term memory limits. In A. Thorn & M. Page (Eds.), Interactions between short-term and long-term memory in the verbal domain (pp. 86-101). Hove, East Sussex, UK: Psychology Press. Cowan, N., Chen, Z., & Rouder, J. N. (2004). Constant capacity in an immediate serial-recall task: A logical sequel to Miller (1956). Psychological Science, 15(9), 634–640. Cowan, N., Elliott, E. M., Saults, J. S., Morey, C. C., Mattox, S., Hismjatullina, A., & Conway, A. R. A. (2005). On the capacity of attention: Its estimation and its role in working memory and cognitive aptitudes. Cognitive Psychology, 51(1), 42‑100. Cowan, N., Rouder, J. N., Blume, C. L., & Saults, J. S. (2012). Models of Verbal Working Memory Capacity: What Does It Take to Make Them Work? Psychological Review, 119(3), 480-499. Cowan, N., Towse, J. N., Hamilton, Z., Saults, J. S., Elliott, E. M., Lacey, J. F., … Hitch, G. J. (2003). Children’s working-memory processes: A response-timing analysis. Journal of Experimental Psychology: General, 132(1), 113. Craik, F. I., & Lockhart, R. S. (1972). Levels of processing: A framework for memory research. Journal of Verbal Learning and Verbal Behavior, 11(6), 671–684. Crannell, C. W., & Parrish, J. M. (1957). A comparison of immediate memory span for digits, letters, and words. The Journal of Psychology: Interdisciplinary and Applied, 44, 319‑327. Cumming, N., Page, M., & Norris, D. (2003). Testing a positional model of the Hebb effect. Memory, 11(1), 43–63. Daneman, M., & Carpenter, P. A. (1980). Individual differences in working memory and reading. Journal of Verbal Learning and Verbal Behavior, 19(4), 450–466. 68 Daneman, M., & Merikle, P. M. (1996). Working memory and language comprehension: A meta-analysis. Psychonomic Bulletin & Review, 3(4), 422–433. Davelaar, E. J. (2013). Short-term memory as a working memory control process. Frontiers in Psychology, 4,13. De Groot, A. D., & Gobet, F. (1996). Perception and memory in chess: Studies in the heuristics of the professional eye. Assen, Netherlands: Van Gorcum. Dieguez, S., Wagner-Egger, P., & Gauvrit, N. (in press). “Nothing happens by accident”, or does it? A low prior for randomness does not explain belief in conspiracy theories. Psychological Science. Delahaye, J.-P., & Zenil, H. (2012). Numerical evaluation of algorithmic complexity for short strings: A glance into the innermost structure of randomness. Applied Mathematics and Computation, 219(1), 63–77. Dixon, P., LeFevre, J.-A., & Twilley, L. C. (1988). Word knowledge and working memory as predictors of reading skill. Journal of Educational Psychology, 80(4), 465. Duncan, J. (2006). EPS Mid-Career Award 2004: Brain mechanisms of attention. The Quarterly Journal of Experimental Psychology, 59(1), 2‑27. Duncan, J., Schramm, M., Thompson, R., & Dumontheil, I. (2012). Task rules, working memory, and fluid intelligence. Psychonomic Bulletin & Review, 19(5), 864–870. Ehrlich, M.-F., & Delafoy, M. (1990). La mémoire de travail: structure, fonctionnement, capacité. L'Année Psychologique, 90(3), 403–427. Engle, R. W. (2002). Working memory capacity as executive attention. Current Directions in Psychological Science, 11(1), 19–23. 69 Engle, R. W., Tuholski, S. W., Laughlin, J. E., & Conway, A. R. (1999). Working memory, short-term memory, and general fluid intelligence: a latent-variable approach. Journal of Experimental Psychology: General, 128(3), 309. Ericsson, K. A., Chase, W. G., & Faloon, S. (1980). Acquisition of a memory skill. Science, 208(4448), 1181–1182. Ericsson, K. A., & Kintsch, W. (1995). Long-term working memory. Psychological Review, 102(2), 211. Farrell, S. (2008). Multiple roles for time in short-term memory: Evidence from serial recall of order and timing. Journal of Experimental Psychology: Learning, Memory, and Cognition, 34(1), 128. Farrell, S. (2012). Temporal clustering and sequencing in short-term memory and episodic memory. Psychological Review, 119(2), 223. Feigenson, L., & Halberda, J. (2008). Conceptual knowledge increases infants’ memory capacity. Proceedings of the National Academy of Sciences, 105(29), 9926–9930. Fernández, A., Ríos-Lago, M., Abásolo, D., Hornero, R., Álvarez-Linera, J., Paul, N., … Ortiz, T. (2011). The correlation between white-matter microstructure and the complexity of spontaneous brain activity: a diffusion tensor imaging-MEG study. Neuroimage, 57(4), 1300–1307. Fernández, A., Zuluaga, P., Abásolo, D., Gómez, C., Serra, A., Méndez, M. A., & Hornero, R. (2012). Brain oscillatory complexity across the life span. Clinical Neurophysiology, 123(11), 2154–2162. French, R. M., Addyman, C., & Mareschal, D. (2011). TRACX: a recognition-based connectionist framework for sequence segmentation and chunk extraction. Psychological Review, 118(4), 614. 70 Fukuda, K., Vogel, E., Mayr, U., & Awh, E. (2010). Quantity, not quality: The relationship between fluid intelligence and working memory capacity. Psychonomic Bulletin & Review, 17(5), 673–679. Gauvrit, N., Singmann, H., Soler-Toscano, F., & Zenil, H. (2015). Algorithmic complexity for psychology: A user-friendly implementation of the coding theorem method. Behavior Research Methods, 1-16. Gauvrit, N., Soler-Toscano, F., & Zenil, H. (2014). Natural scene statistics mediate the perception of image complexity. Visual Cognition, 22(8), 1084-1091. Gauvrit, N., Zenil, H., Delahaye, J.-P., & Soler-Toscano, F. (2014). Algorithmic complexity for short binary strings applied to psychology: a primer. Behavior Research Methods, 46(3), 732‑744. Gendle, M. H., & Ransom, M. R. (2006). Use of the electronic game SIMON® as a measure of working memory span in college age adults. Journal of Behavioral and Neuroscience Research, 4, 1‑7. Gilbert, A. C., Boucher, V. J., & Jemel, B. (2014). Perceptual chunking and its effect on memory in speech processing: ERP and behavioral evidence. Frontiers in Psychology, 5. Gilchrist, A. L., Cowan, N., & Naveh-Benjamin, M. (2008). Working memory capacity for spoken sentences decreases with adult ageing: Recall of fewer but not smaller chunks in older adults. Memory, 16(7), 773–787. Gobet, F., & Simon, H. A. (1996). Recall of rapidly presented random chess positions is a function of skill. Psychonomic Bulletin & Review, 3(2), 159–163. Gray, J. R., Chabris, C. F., & Braver, T. S. (2003). Neural mechanisms of general fluid intelligence. Nature Neuroscience, 6(3), 316–322. 71 Guida, A., Gobet, F., & Nicolas, S. (2013). Functional cerebral reorganization: a signature of expertise? Reexamining Guida, Gobet, Tardieu, and Nicolas’(2012) two-stage framework. Frontiers in Human Neuroscience, 7. Guida, A., Gobet, F., Tardieu, H., & Nicolas, S. (2012). How chunks, long-term working memory and templates offer a cognitive explanation for neuroimaging data on expertise acquisition: a two-stage framework. Brain and Cognition, 79(3), 221–244. Guida, A., Tardieu, H., & Nicolas, S. (2009). The personalisation method applied to a working memory task: Evidence of long-term working memory effects. European Journal of Cognitive Psychology, 21(6), 862–896. Haarmann, H. J., Davelaar, E. J., & Usher, M. (2003). Individual differences in semantic short-term memory capacity and reading comprehension. Journal of Memory and Language, 48(2), 320–345. Hanten, G., & Martin, R. C. (2000). Contributions of phonological and semantic short-term memory to sentence processing: Evidence from two cases of closed head injury in children. Journal of Memory and Language, 43(2), 335–361. Hitch, G. J., Towse, J. N., & Hutton, U. (2001). What limits children’s working memory span? Theoretical accounts and applications for scholastic development. Journal of Experimental Psychology: General, 130(2), 184. Humes, L. E., & Floyd, S. S. (2005). Measures of working memory, sequence learning, and speech recognition in the elderly. Journal of Speech, Language, and Hearing Research, 48(1), 224–235. Hutter, M. (2005). Universal artificial intelligence. Berlin, Germany: Springer-Verlag. James, W. (1890). The methods and snares of psychology. In The principles of psychology, Vol I (p. 183‑198). New York, NY: Henry Holt and Co. 72 Johnson, N. F. (1969). Chunking: Associative chaining versus coding. Journal of Verbal Learning and Verbal Behavior, 8(6), 725–731. Kane, M. J., Hambrick, D. Z., & Conway, A. R. (2005). Working memory capacity and fluid intelligence are strongly related constructs: comment on Ackerman, Beier, and Boyle (2005). Psychological Bulletin, 131(1), 66–71. Kane, M. J., Hambrick, D. Z., Tuholski, S. W., Wilhelm, O., Payne, T. W., & Engle, R. W. (2004). The generality of working memory capacity: a latent-variable approach to verbal and visuospatial memory span and reasoning. Journal of Experimental Psychology: General, 133(2), 189. Kane, M. J., & Miyake, T. M. (2007). The validity of « conceptual span » as a measure of working memory capacity. Memory & Cognition, 35(5), 1136–1150. Karpicke, J., & Pisoni, D. B. (2000). Memory span and sequence learning using multimodal stimulus patterns: Preliminary findings in normal-hearing adults. Research on Spoken Language Processing. Kempe, V., Gauvrit, N., & Forsyth, D. (2015). Structure emerges faster during cultural transmission in children than in adults. Cognition, 136, 247-254. Kibbe, M. M., & Feigenson, L. (2014). Developmental origins of recoding and decoding in memory. Cognitive Psychology, 75, 55–79. Klapp, S. T., Marshburn, E. A., & Lester, P. T. (1983). Short-term memory does not involve the « working memory » of information processing: The demise of a common assumption. Journal of Experimental Psychology: General, 112(2), 240. Kolmogorov, A. N. (1965). Three approaches to the quantitative definition of information. Problems of Information Transmission, 1(1), 1–7. Krumm, S., Schmidt-Atzert, L., Buehner, M., Ziegler, M., Michalczyk, K., & Arrow, K. (2009). Storage and non-storage components of working memory predicting 73 reasoning: A simultaneous examination of a wide range of ability factors. Intelligence, 37(4), 347-364. Lewandowsky, S., Oberauer, K., Yang, L.-X., & Ecker, U. K. (2010). A working memory test battery for MATLAB. Behavior Research Methods, 42(2), 571–585. Li, M., & Vitányi, P. M. (2009). An introduction to Kolmogorov complexity and its applications. New York, NY: Springer Verlag. Logan, G. D. (2004). Working memory, task switching, and executive control in the task span procedure. Journal of Experimental Psychology: General, 133(2), 218. Luck, S. J., & Vogel, E. K. (1997). The capacity of visual working memory for features and conjunctions. Nature, 390(6657), 279–281. Machado, B. S., Miranda, T. B., Morya, E., Amaro Jr, E., & Sameshima, K. (2010). Algorithmic complexity measure of EEG for staging brain state. Clinical Neurophysiology, 121, S249–S250. Majerus, S., Martinez Perez, T., & Oberauer, K. (2012). Two distinct origins of long-term learning effects in verbal short-term memory. Journal of Memory and Language, 66(1), 38–51. Martin, R. C., & Romani, C. (1994). Verbal working memory and sentence comprehension: A multiple-components view. Neuropsychology, 8(4), 506. Martínez, K., Burgaleta, M., Román, F. J., Escorial, S., Shih, P. C., Quiroga, M. Á., & Colom, R. (2011). Can fluid intelligence be reduced to ‘simple’short-term storage?. Intelligence, 39(6), 473-480. Mathy, F., & Feldman, J. (2012). What’s magic about magic numbers? Chunking and data compression in short-term memory. Cognition, 122(3), 346–362. Mathy, F., & Varré, J.-S. (2013). Retention-error patterns in complex alphanumeric serialrecall tasks. Memory, 21(8), 945–968. 74 Maybery, M. T., Parmentier, F. B., & Jones, D. M. (2002). Grouping of list items reflected in the timing of recall: Implications for models of serial verbal memory. Journal of Memory and Language, 47(3), 360–385. Miller, G. A. (1956). The magical number seven, plus or minus two: some limits on our capacity for processing information. Psychological Review, 63(2), 81. Moher, M., Tuerk, A. S., & Feigenson, L. (2012). Seven-month-old infants chunk items in memory. Journal of Experimental Child Psychology, 112(4), 361–377. Naveh-Benjamin, M., Cowan, N., Kilb, A., & Chen, Z. (2007). Age-related differences in immediate serial recall: Dissociating chunk formation and capacity. Memory & Cognition, 35(4), 724–737. Ng, H. L., & Maybery, M. T. (2002). Grouping in short-term verbal memory: Is position coded temporally? The Quarterly Journal of Experimental Psychology: Section A, 55(2), 391–424. Oakes, L. M., Ross-Sheehy, S., & Luck, S. J. (2006). Rapid development of feature binding in visual short-term memory. Psychological Science, 17(9), 781–787. Oberauer, K. (2002). Access to information in working memory: exploring the focus of attention. Journal of Experimental Psychology: Learning, Memory, and Cognition, 28(3), 411. Oberauer, K., Farrell, S., Jarrold, C., Pasiecznik, K., & Greaves, M. (2012). Interference between maintenance and processing in working memory: The effect of item– distractor similarity in complex span. Journal of Experimental Psychology: Learning, Memory, and Cognition, 38(3), 665. Oberauer, K., & Lange, E. B. (2009). Activation and binding in verbal working memory: A dual-process model for the recognition of nonwords. Cognitive Psychology, 58(1), 102–136. 75 Oberauer, K., Lange, E., & Engle, R. W. (2004). Working memory capacity and resistance to interference. Journal of Memory and Language, 51(1), 80–96. Oberauer, K., Lewandowsky, S., Farrell, S., Jarrold, C., & Greaves, M. (2012). Modeling working memory: an interference model of complex span. Psychonomic Bulletin & Review, 19(5), 779–819. Osaka, M., Nishizaki, Y., Komori, M., & Osaka, N. (2002). Effect of focus on verbal working memory: Critical role of the focus word in reading. Memory & Cognition, 30(4), 562– 571. Pelli, D. G. (1997). The VideoToolbox software for visual psychophysics: Transforming numbers into movies. Spatial Vision, 10(4), 437–442. Perfetti, C. A., & Lesgold, A. M. (1977). Discourse comprehension and sources of individual differences. In M.A. Just & P.A. Carpenter (Eds), Cognitive processes in comprehension. Hillsdale, NJ: Erlbaum. Perlman, A., Pothos, E. M., Edwards, D. J., & Tzelgov, J. (2010). Task-relevant chunking in sequence learning. Journal of Experimental Psychology: Human Perception and Performance, 36(3), 649. Perruchet, P., & Pacteau, C. (1990). Synthetic grammar learning: Implicit rule abstraction or explicit fragmentary knowledge? Journal of Experimental Psychology: General, 119(3), 264. Rabinovich, M. I., Varona, P., Tristan, I., & Afraimovich, V. S. (2014). Chunking dynamics: heteroclinics in mind. Frontiers in Computational Neuroscience, 8, 22. Raven, J. C. (1962). Advanced progressive matrices: Sets I and II. HK Lewis. Robinet, V., Lemaire, B., & Gordon, M. B. (2011). MDLChunker: a MDL-based cognitive model of inductive learning. Cognitive Science, 35(7), 1352–1389. 76 Ryabko, B., Reznikova, Z., Druzyaka, A., & Panteleeva, S. (2013). Using ideas of Kolmogorov complexity for studying biological texts. Theory of Computing Systems, 52(1), 133–147. Saito, S., & Miyake, A. (2004). On the nature of forgetting and the processing–storage relationship in reading span performance. Journal of Memory and Language, 50(4), 425–443. Salthouse, T. A., & Babcock, R. L. (1991). Decomposing adult age differences in working memory. Developmental Psychology, 27(5), 763. Schroeder, P. J., Copeland, D. E., & Bies-Hernandez, N. J. (2012). The influence of story context on a working memory span task. The Quarterly Journal of Experimental Psychology, 65(3), 488–500. Shipstead, Z., Redick, T. S., & Engle, R. W. (2012). Is working memory training effective? Psychological Bulletin, 138(4), 628. Soler-Toscano, F., Zenil, H., Delahaye, J.-P., & Gauvrit, N. (2013). Correspondence and independence of numerical evaluations of algorithmic information measures. Computability, 2(2), 125–140. Soler-Toscano, F., Zenil, H., Delahaye, J.-P., & Gauvrit, N. (2014). Calculating Kolmogorov complexity from the output frequency distributions of small Turing machines. PloS One, 9(5), e96223. Solway, A., Diuk, C., Córdova, N., Yee, D., Barto, A. G., Niv, Y., & Botvinick, M. M. (2014). Optimal hehavioral hierarchy. PLoS Computational Biology, 10(8), e1003779. Stahl, A. E. & Feigenson, L. (2014). Social knowledge facilitates chunking in infancy. Child Development, 85(4), 1477-1490. Steiger, J. H. (1980). Tests for comparing elements of a correlation matrix. Psychological Bulletin, 87(2), 245. 77 Szmalec, A., Page, M., & Duyck, W. (2012). The development of long-term lexical representations through Hebb repetition learning. Journal of Memory and Language, 67(3), 342–354. Thomas, J. G., Milner, H. R., & Haberlandt, K. F. (2003). Forward and Backward Recall Different Response Time Patterns, Same Retrieval Order. Psychological Science, 14(2), 169–174. Towse, J. N., Cowan, N., Hitch, G. J., & Horton, N. J. (2008). The recall of information from working memory: Insights from behavioral and chronometric perspectives. Experimental Psychology, 55(6), 371. Towse, J. N., Hitch, G. J., Horton, N., & Harvey, K. (2010). Synergies between processing and memory in children’s reading span. Developmental Science, 13(5), 779–789. Tulving, E., & Patkau, J. E. (1962). Concurrent effects of contextual constraint and word frequency on immediate recall and learning of verbal material. Canadian Journal of Psychology, 16(2), 83. Turner, M. L., & Engle, R. W. (1989). Is working memory capacity task dependent? Journal of Memory and Language, 28(2), 127–154. Unsworth, N., & Engle, R. W. (2006). Simple and complex memory spans and their relation to fluid abilities: Evidence from list-length effects. Journal of Memory and Language, 54(1), 68–80. Unsworth, N., & Engle, R. W. (2007a). On the division of short-term and working memory: an examination of simple and complex span and their relation to higher order abilities. Psychological Bulletin, 133(6), 1038. Unsworth, N., & Engle, R. W. (2007b). The nature of individual differences in working memory capacity: active maintenance in primary memory and controlled search from secondary memory. Psychological Review, 114(1), 104. 78 Unsworth, N., Redick, T. S., Heitz, R. P., Broadway, J. M., & Engle, R. W. (2009). Complex working memory span tasks and higher-order cognition: A latent-variable analysis of the relationship between processing and storage. Memory, 17(6), 635–654. Wang, T., Ren, X., Li, X., & Schweizer, K. (2015). The modeling of temporary storage and its effect on fluid intelligence: Evidence from both Brown–Peterson and complex span tasks. Intelligence, 49, 84-93. Wechsler, D. (2008). Wechsler adult intelligence scale–Fourth Edition (WAIS–IV). San Antonio, TX: NCS Pearson. Wheeler, M. E., & Treisman, A. M. (2002). Binding in short-term visual memory. Journal of Experimental Psychology: General, 131(1), 48. Yagil, G. (2009). The structural complexity of DNA templates—Implications on cellular complexity. Journal of Theoretical Biology, 259(3), 621–627. 79
Documents pareils
Short-term memory tasks coupled with online chunking: a straight
and memory are inextricably intertwined
► the extraction of regularities domain is twofold:
implicit learning (based on exploited statistics;
Saffran, Aslin & Newport, 1996), which forms long-term
...