Which of the following is true about the cognitive processes that produce behavior?

  • Journal List
  • HHS Author Manuscripts
  • PMC3065938

Learn Behav. Author manuscript; available in PMC 2011 Mar 29.

Published in final edited form as:

PMCID: PMC3065938

NIHMSID: NIHMS190745

Abstract

In his 1948 address to the Division of Theoretical-Experimental Psychology of the American Psychological Association, Kenneth W. Spence discussed six distinctions between cognitive and stimulus-response (S-R) theories of learning. In this article I first review these six distinctions, and then focus on two of them in the context of my own research. This research concerns the specification of stimulus-stimulus associations in associative learning, and the characterization of neural systems underlying those associations. In the course of describing Spence's views and my research, I hope to communicate some of the richness of Spence's S-R psychology, and its currency within modern scientific analyses of behavior.

Spence (1950, p. 161) characterized cognitive theories of learning as those that “emphasized the formation and modification of cognitive patterns representative of the relationships in the environment.” For the most part, within these theories, such as those of Koffka (1935), Kohler (1940), Lewin (1936), and Tolman (1932), learning was construed as part of a larger problem of perceptual organization and reorganization with experience. By contrast, stimulus-response (S-R) theories, such as those of Guthrie (1935), Hull (1943), Spence (1936), and Thorndike (1898) emphasized such constructs as habits and S-R bonds, which referred to hypothetical learning states or intervening variables. S-R theories provided rules relating stimulus factors such as reward magnitude, number and timing to the strengths of those intervening variables, and rules relating those variables to empirical response measures. On the whole, Spence saw few points of disagreement between these two theoretical positions, and attributed most of the dissension between the camps to misinterpretation of S-R theory by cognitive theorists. From my perspective, the tone conveyed in his paper was that of a patient teacher pointing out the mistakes of well-intentioned but misguided students of cognition, but without sparing the somewhat-less-pervasive (in his eyes) mistakes of S-R theorists.

I. Six distinctions between cognitive and S-R theories

Metaphors: map control rooms vs telephone switchboards

Perhaps because spatial-learning tasks provided an important test arena for early cognitive and S-R psychologists, cognitive theories of learning became associated with the metaphor of “map control rooms”, in which spatial representations and relations were acquired, computed, and exploited. By contrast, S-R theories became attached to the analogy of “telephone switchboards” by which stimulus inputs were, through learning, connected to new response outputs (Tolman, 1948). Spence (1950, p. 161) asserted that “no scientifically oriented person in psychology, however, would ever take such analogies, whether telephone switchboards or map control rooms, as serious attempts at theoretical representations of learning changes”. For example, the meaning of Hull's habit construct was “given by the mathematical function relating it to the antecedent experimental variables” and thus “any comparison of switchboards with map control rooms is entirely beside the point” (Spence, 1950, p. 163). More relevant was the question of whether learning in spatial tasks was more “map-like” or “habit-like”, that is, how much of the spatial information contained in those tasks was encoded in learning, and could be used to guide performance later, when, for example, the usual paths were blocked, or shortcuts opened (Tolman, Richie, & Kalish, 1946a, b). Interestingly, much of this debate was eventually distilled to the issue of place versus response, which could just as easily phrased entirely in S-R terms as “approach cue X” versus “turn right”.

Neurophysiological basis of learning: brain fields or receptor-effector connections

Whereas cognitive theorists referred to reorganization of “electrical brain fields” and “neurophysiological trace systems”, Hull (1943) related habit formation to the establishment of neural “receptor-effector” connections. Spence noted that these differences had little or no significance for learning theory because these neurophysiological models were nothing more than analogies and played no role in the deductions or inferences of either camp. Properties of “brain fields” were inferred from introspection, rather than physiological investigation, and with few exceptions, knowledge or even speculation about the physiological basis of reflexes did not inform the construction of S-R theories. Indeed, Spence (1950, p. 164) stated that “picturing neurophysiological processes without specifying the hypothetical relations that tie them up with the experimental variables and the response measure is almost a complete waste of time so far as furthering our understanding of learning phenomena is concerned.” In section 3 of this article, I will relate examples of progress in relating brain and behavior which have been more useful in characterizing the nature of learning.

Stimulus-stimulus (S-S) or stimulus-response (S-R) associations

Cognitive theorists of the time were clear that learning involved associations among, or reorganization of, sensory-perceptual processes. By contrast, Guthrie, Thorndike, Hull and others posited that learning involved S-R associations, between stimuli and “muscle contraction and glandular secretion” (Guthrie, 1946, p. 7). Spence (1950) pointed out however that Hull's emphasis on S-R associations followed from his neurophysiologizing, rather than from his mathematical definition of habit. From Spence's perspective, there was nothing intrinsically S-R or S-S about habit, as defined within the Hullian system, although, like Hull, he clearly believed that most learning involved formation of associations between stimuli and responses.

“I do not find it difficult to conceive of both types of organizations or associations being established in learning. Certainly simple types of perceptual learning would appear possibly to involve intersensory associations. I seriously doubt, however, whether learning is exclusively of this type, or even that the majority of it is. Indeed, ……evidence would appear to support more strongly the S-R conception than the S-S” (Spence, 1950, pp. 164-165.)

Most of this article (section 2) will concern this issue.

Contents vs conditions of learning

Spence noted that cognitive theorists tended to emphasize the “intrinsic” properties of their constructs, whereas S-R theorists tended to be concerned with the empirical relations among experimental variables that determine their constructs. In Rescorla's (1975) terminology, the cognitive theorists concentrated on the content of learning whereas S-R theorists focused on the conditions under which learning occurred. Although castigating cognitive theorists for relying too much on introspection to make inferences about the contents of learning, Spence recognized that such contents could still be rigorously defined in terms of environmental variables, lauding Tolman's sign-gestalt psychology. Thus, he concluded that these concerns reflected “a very real difference between the two theoretical camps, but… one of emphasis rather than of conflict” (Spence, 1950, p. 166).

Stimulus variables: intrinsic vs. extrinsic

At the time of Spence's writing, investigators from the cognitive tradition tended to examine the effects of variables that influenced the receipt of stimuli (e.g., orienting and attention) and perceptual organization (e.g., figure-ground relations, part-whole relations, belongingness, set). By contrast, most studies of behavior from the S-R tradition focused on temporal variables, and those that related to motivation, such as reward magnitude and deprivation state. But Spence's (1950, p. 167) attitude was that “such differences of emphasis…. do not necessarily involve conflict”, and noted that S-R theorists' preoccupation with time and motivation did not preclude an interest in stimulus reception and organizational variables. Indeed, in the past 25 years a great deal of learning research emerging from the S-R tradition has addressed these later issues, for example, interests in orienting and attention (Holland, 1997; Mackintosh, 1975; Pearce & Hall, 1980) and in perceptual organization (Cook, Riley & Brown, 1992; Rescorla, 1986).

Organism as active or passive processor of information

At the time of Spence's address, there was considerable discussion about the “role of the organism” in information processing, whether animals “merely passively receive and react to all the stimuli which are physically present” (Tolman, 1948, p. 189), or actively processed that information, selecting and assembling the elements of association prior to forming associations themselves. Again, Spence felt that cognitive theorists had misrepresented the S-R position:

“It is difficult to know for sure just what Tolman and others who have expressed this notion mean by this kind of statement. ….Much of the S-R theory is concerned with the classical conditioning situation in which the conditions of stimulation are extremely simple. No “active looking” for the cue… or special receptor orientation is necessary… But even in the case of this simple learning situation the S-R learning theorist has not assumed that organisms passively receive and react to all stimuli that are physically present” (Spence, 1950, p. 168.)

Thus, Spence set the stage for subsequent incorporation of variables that are now described as “attentional” into S-R theory. For example, reinforcement contingencies might reasonably be expected to extend to both explicit attentional processing such as eye movements and orienting behaviors (e.g., Siegel, 1967) and more implicit processing assumed by more recent learning theories, such as changes in learning rate parameters. All of these could influence the subsequent reception and use of particular stimuli in association formation later.

II. Is learning S-S or S-R?

As Spence noted, there is nothing intrinsically S-S or S-R about a habit. For example, Figures 1a and 1b reproduce Spence's (1950) and Holland's (1990) diagrams distinguishing potential S-S and S-R associations in a simple Pavlovian conditioning experiment. If the second term of an association is late in the processing stream, we describe the association as S-R and if it is early we describe it as S-S. From this perspective, how late is late and how early is early define the question of whether an association or habit is S-R or S-S.

Which of the following is true about the cognitive processes that produce behavior?

A: Spence's (1950) diagram contrasting stimulus-stimulus (S-S) and stimulus-response (S-R) associations. Sc and Su refer to the conditioned and unconditioned stimuli, respectively, A1 and A2 refer to afferent processes produced by Sc and Su, E1 and E2 refer to efferent processes produced by A1 and A2, and Ro and Ru to the orienting and unconditioned responses. B: Holland's (1990) diagram contrasting S-S and S-R associations. CS, US, and UR refer to conditioned stimulus, unconditioned stimulus, and orienting response, respectively. Letters a-d refer to hypothetical stages of processing.

Because in the simplest case, such views suppose a linear chain of event processing from stimulus reception to motor output, it seems unlikely that dichotomous categorization of learned behaviors could be realistically made within either psychological theories or neurophysiological theories, by pointing to differences in stimulus versus response processing or sensory versus motor brain regions. Nevertheless, there is likely to be considerable agreement that some events are “more stimulus” and others are “more response”. Perhaps the earliest test arenas for asking the S-S or S-R question were those in which either stimulus or response aspects of the second, “reinforcer” event in an association event were minimized.

Responses without stimuli

Loucks (1935) examined learning when foreleg flexion in the presence of a buzzer was induced by electrical stimulation of the appropriate region of motor cortex, rather than the usual foreleg shock US. In this procedure, the electrical stimulation was presumed to produce a response without preceding stimulus input. Thus, S-S learning should be impossible and S-R learning unaffected in this preparation. Indeed Loucks found no evidence for acquisition of the leg flexion response, suggesting that contiguity of stimulus and response was insufficient for learning. Spence (1950) conceded the point but then noted that in a second experiment, Loucks (1935) found that if the buzzer-motor cortex stimulation was also followed by food, leg flexion was acquired. Spence asserted that this pair of outcomes was easily interpretable within an S-R reward-based theory: Unlike leg shock, motor cortex stimulation produced a flexion response but no drive (pain) whose reduction could reinforce habit formation, whereas the addition of food added such a source of reward. Although others were to find successful limb flexion conditioning with motor cortex stimulation (e.g. Doty & Giurgea, 1961), later experiments (e.g., Thomas, 1971; Wagner, Thomas, & Norton, 1967) cast doubt on whether such studies revealed direct Pavlovian conditioning of leg flexion responses or the instrumental conditioning of responses that reduced the negative postural consequences of the sudden uncontrolled leg movements elicited by motor cortex stimulation. And apart from these concerns, one could easily question whether massive, unpatterned motor cortex stimulation would provide an appropriate basis for plasticity needed for S-R learning at that point in the system. Notably, in some systems, such as eyelid conditioning, both CS and US can be replaced by direct stimulation of neural pathways (e.g., Steinmetz, Lavond & Thompson, 1989; Thompson et al., 1998; see section 3 of this paper).

Stimuli without responses

By the same logic, blocking responses to USs or choosing USs with minimal responses should substantially impair learning, if learning is S-R, but have little effect if learning is S-S. Early studies with peripheral blockade of responses during training typically revealed substantial responding after the blockade was removed (e.g., Finch, 1938). However, it is again easily argued that no one really expects S-R learning to be localized at the neuromuscular junction or sympathetic neuron. Such studies proved more useful in addressing claims that adventitious instrumental reinforcement of motor responses was responsible for the emergence of putatively Pavlovian CRs.

A more fruitful line of research was that of sensory preconditioning, in which relatively neutral stimuli, such as lights and sounds, were paired. Although typically such CS1-CS2 pairings did not produce readily observable changes in behavior by themselves, subsequent pairing of one of them (CS2) with a more significant event such as food or shock typically produced learning responding to the other cue (CS1) as well (Figure 2a). Typically, CRs to the untrained CS1 cue were thought to be mediated by the chaining of two separately-acquired associations at the time of performance. In the initial phase, S-S associations were formed between events that were mostly sensory in nature. Then in the second phase, associations were formed between CS2 and US. Finally, when CS1 was presented in test, it would call up CS2, which in turn would elicit a CR (Figure 2b). Of course, imaginative S-R theorists quickly proposed alternative, S-R solutions to the sensory preconditioning problem. For example (Figure 2c), in the first phase S-R associations could be formed between the CS1 and orienting responses to CS2, and then associations between sensory feedback from those responses and the response to the US in the second phase. Finally, in test, S1 would elicit the conditioned OR originally elicited by CS2 and feedback from those conditioned ORs would in turn elicit the UR-like CR conditioned to those feedback stimuli in the second phase (Osgood, 1953, p. 461-462). Because such responses were often not easily observed, these accounts were often difficult to support or refute. Nevertheless, hundreds of sensory preconditioning experiments have yielded evidence for learning in the absence of obvious responses (e.g., Thompson, 1972). I suspect Spence would have been content to describe these results as examples of sensory-sensory habits.

Which of the following is true about the cognitive processes that produce behavior?

Procedures and outcomes of the sensory preconditioning experiment. A: Experimental procedures. B: Stimulus-stimulus mediated performance account. C: Stimulus-response mediated performance account. D: Stimulus-stimulus mediated learning account. CS1, CS2, and US refer to two conditioned stimuli and the unconditioned stimulus. OR2 and OR2-s refer to the orienting response originally elicited by CS2 and to the stimulus feedback from OR2. Dashes (---) signify arranged relations; arrows (→) signify learned associations; double-headed arrows (← →) signify bidirectional associations; lines (—) signify unconditioned relations.

Both the S-S and S-R accounts above assume that the performance of CRs to CS1 is mediated at the time of test presentations by chaining together separate associations previously established between CS1 and CS2 and between CS2 and the US. No direct associations are formed between CS1 and the US. An alternative account is that direct associations are formed between CS1 and the US in the second phase. By this “mediated learning” (Holland, 1981a) or “backward sensory preconditioning” (Ward-Robinson & Hall, 1996) account, in the second phase, CS2 associatively activates a representation of CS1, which is then followed by the US. Thus, associations can be formed directly between CS1 and the US, which are then revealed in the final test of CS1 alone (Figure 2d). It is usually difficult to distinguish between these alternatives, although it is notable that in an early human sensory preconditioning experiment, Brogden (1947) distinguished between subjects who attributed their response to CS1 in test to its reminding them of CS2 and subjects who that claimed that in Phase 2 CS1 had actually been paired with the US. Subjects can be encouraged to use a “mediated learning” strategy by using conditioning parameters that discourage use of the CS1-CS2 associations required by the mediated performance account and instead encourage the use of CS2-CS1 associations required by the mediated learning account. For example, Ward-Robinson and Hall (1996) presented the two neutral stimuli in the order CS2-CS1 in the first phase of a sensory preconditioning experiment, and found substantial responding to CS1 in test.

Post-training assessments of the nature of learning

In the experiments discussed above, the logic was to examine the consequences of altering the conditions of learning so that they would be conducive (or not) to either S-S or S-R learning. However, although such studies were useful for showing that learning could be S-S or S-R, they were not informative about whether learning in any typical conditioning procedure was S-S or S-R. The use of post-training tests of various sorts has often made it possible to collect such information.

Transfer tests

Perhaps the most common use of post-training transfer tests was to determine whether rats in simple mazes learned spatial attributes of the maze environment or turn responses. For example, in Tolman's “sunburst” experiment (Tolman et al., 1946a), rats were trained in an apparatus in which rats first ran straight down an alley, then had to make a left and two right turns before finding food on the right side of the room. In a transfer test, rats were presented with alleys leading in a fan-shaped array of angles from the start area. If rats had learned “turn left” or “turn right” responses, they might be expected to choose the left-pointing (0-deg) or right-pointing (180-deg) alleys, if they had learned to “go straight then make turns” they would select the straight-ahead (90-deg) alley, but if they had learned the approximate location of the food in the room, they might select an alley that pointed to that location (135-deg). Although Tolman et al. (1946a) found considerable individual variation in the rats' choice behavior, the bulk of the data supported the assertion that rats learned the approximate location of food (or that they integrated straight and right response vectors appropriately). A version of this experiment that has lasted longer follows T-maze training with a test in the maze when it has been rotated 180 degrees (Packard & McGaugh, 1996; Poldrack and Packard, 2003; Tolman, Richie, & Kalish, 1946b). Thus in the test, “response” and “place” strategies are placed in opposition. If the rat makes the same turn (e.g., right) as in training, response learning is identified, whereas if it approaches the same side of the room as in training, place learning is inferred.

Post-training changes in reward

Another way of phrasing the S-R vs S-S question was whether rewards facilitated the formation of associations or participated in them. That is, did reward simply stamp in (Thorndike, 1898) or catalyze (Colwill & Rescorla, 1986; Rescorla, 1988) stimulus-response associations or did animals form associations between stimuli and sensory and other properties of rewards? An initial approach to this question was to establish steady state performance with a particular reward, and then substitute another reward. If reinforcers merely stamp-in associations, then no disruptions in performance would be expected (presuming the substituted reinforcer was of comparable value), whereas if animals anticipated particular rewards, then aspects of performance might be sensitive to the change. Early delayed-response experiments showed that monkeys could maintain memory representations of particular food items over relatively long intervals. For example, working with Kohler, Tinkelpaugh (1928) trained monkeys in a task in which, in view of the subject, the experimenter placed food under one cup but not another, identical cup. After a delay in which the cups were screened from the monkey's view, the screen was removed and the monkey told to get the food. Several different foods were used, in different sessions. Normally, the monkey selected the correct cup and quickly consumed the food reward. However, if the experimenter hid banana, but then substituted lettuce under the cup while it was hidden from the monkey's view, the monkey examined but did not touch the lettuce, thoroughly investigated the cup, and on occasion shrieked at the experimenter. By contrast, on a subsequent trial, in which lettuce was hidden and maintained under the cup, the monkey consumed the lettuce rapidly and without incident. Thus, from these reactions to the substituted reward, monkeys must have represented specific aspects of reward in memory. Unfortunately, because the reward expectancy was produced by presentation of the reward itself rather than by an associate of the reward, such experiments shed little light on whether reward properties are represented in associative learning. Closer to the mark was an experiment performed in Tolman's laboratory by Elliot (1928), in which different groups of rats were trained to run in a complex maze for either sunflower seeds or the preferred bran mash. After substantial training, in the bran group the bran reward was replaced by the less-preferred sunflower seeds. Although the rats readily consumed the sunflower seeds with no signs of disturbance, performance on subsequent trials was considerably disrupted, so as to be worse than the performance of rats that had been trained with sunflower seed reward from the start. This disruption suggested that the rats expected a particular reward. However, Elliot did not perform the analogous sunflower-to-bran substitution, and so the disruption may have reflected the effects of reduction in reward magnitude alone.

More recent experiments examined the effects of US substitution in Pavlovian conditioning, in a more symmetrical manner. For example, Holland and Forbes (1982) found that either of two USs (differently-flavored sucrose pellets) was better maintained in memory (assessed with a preparatory-releaser procedure, e.g. Terry & Wagner, 1976) if they were preceded by an incongruent Pavlovian CSl (i.e., for the other US) than by a congruent CS. Similarly, in my laboratory, Alexander Johnson (unpublished data) recently found that mice showed disruption of consumption of each of a pair of fluid USs (lecithin and sucrose) when each was accompanied by an auditory signal for the other fluid. Thus, at least some properties of the reinforcer seem to be coded by the CS after simple conditioning procedures.

Reinforcer revaluation

The post-training reinforcer revaluation procedure (described formally by Rozeboom, 1958) has become the gold standard of assaying the presence or absence of S-S (Pavlovian) or R-S (instrumental) associations (Pickens & Holland, 2004). Animals first receive Pavlovian or instrumental training with one or more reinforcers. Then, the value of a reinforcer is altered in the absence of the cues or responses of training. Finally, responding is assessed in the absence of any of the reinforcers. Often, the level of response observed in test spontaneously changes to reflect changes in value of the reinforcer with which it had been previously paired. Major advantages of this paradigm include its utility after nearly any sort of training procedure and the availability of a wide range of methods for altering the value of rewards either upward (“inflation”) or downward (“devaluation”). For example, motivational states relevant to the reinforcers can be induced chemically (e.g., increasing or decreasing sodium balance after training with salt solution rewards) or by altering exposure to them (e.g., selective satiation of one of two food rewards). Similarly, particular aspects of rewards may be altered associatively by pairing them with various agents that increase or decrease their attractiveness. Even nonassociative manipulations, such as simple exposure to the reinforcer itself, or weaker or stronger versions of the reinforcer, can be effective (e.g. Rescorla, 1973, 1974).

An early application of this method was a study by Zener and McCurdy (1939). Food-deprived dogs received pairings of two cues with different-flavored dog biscuits. Then the dogs were carefully sated on one of the flavors of dog biscuit. Finally, the dogs' reactions to each cue were noted. They responded much less energetically to the cue for the biscuit on which they were sated. Although by present standards the study was somewhat flawed because the biscuits continued to be presented in the test (and thus the reduced responding may have reflected the lower reinforcement value of the sated biscuit rather than a spontaneous adjustment of responding based on its altered value), these and other results encouraged Zener (1937, p. 386) to describe conditioning as “a reorganization into some kind of functional whole of the perceptual systems corresponding to the conditioned and unconditioned stimuli; and in the functional relation of this organized system to the urge or tension system originally excited by the unconditioned stimulus” rather than the learning of S-R associations.

A similar logic motivated early studies of discriminative maze or alley performance by Tolman (1933), Miller (1935), and Tolman and Gleitman (1949). Tolman (1933) initially trained rats in a two-choice discrimination box with two distinctive stimuli, black or white curtains. Passing through (say) the black curtains led to a food box whereas passing through the white curtains led to a blind alley with no food. After substantial training on this task, the rats were placed directly in the food box and given an electric shock. Finally the rats were returned to the start box and allowed to run. If rats had truly learned the response-food box associations or “sign-gestalt expectations”, then even on the first trial they should fail to run. However each rat “immediately dashed off gaily and just as usual through the whole discrimination apparatus and bang whack into the food compartment in which he had just been shocked” (Tolman, 1933, p. 250.)

In a reply to Tolman (1933), Miller (1935) went one better by describing an S-R account for how rats could exhibit flexible behavior in such a situation, and proceeded to demonstrate it. In his account, Miller made use of the concept of fractional anticipatory goal responses. Rats might make powerful goal responses (Rg, e.g., chewing), portions of which (rg) could be conditioned to alley cues that precede the goal box itself. When a rat was subsequently placed in the goal box and shocked, leaping or freezing responses to the shock might be conditioned to stimulus feedback from the previous food-based goal responses (e.g., Rg-Sg→Rshock). Then, when the rat was placed back in the alley, performance of the conditioned rg would yield the feedback cue sg, which in turn would elicit a conditioned, fractional part of the response to shock, interfering with running. Miller suggested that all that would be needed to generate Tolman's anticipated data would be to give the goal boxes distinctive goal responses. He did this by using two rewards (food and water) and two distinctive goal boxes, one requiring a climb and sharp right turn and one that required a straight entry followed by a sharp left turn. Unlike Tolman's apparatus, which required a simultaneous choice discrimination, Miller used a single alley with a single goal box, but exposed the rats to the other goal box in the absence of the alley. When rats were later placed directly into the entry of one of the goal boxes and shocked, the shock response would be conditioned to a particular Sg arising from the distinctive Rg in that box. If a rat was shocked in the same goal box as had always been found at the end of the alley, placing that rat back in the alley start box for test would elicit an rg whose sg feedback was associated with the fractional shock response, and hence produce interference with running. By contrast, if the rat was shocked in the other goal box, the alley-elicited rg's sg feedback would not be associated with the shock response, and no such interference would be anticipated. Miller (1935) indeed found this outcome. Thus, Miller's S-R account not only explained such devaluation performance, but suggested conditions under which it would and would not be observed.

Interestingly, in Tolman and Gleitman's (1949) classic T-maze study, in which Tolman finally succeeded in finding spontaneous devaluation performance, the two goal boxes were also constructed in a way that they required distinctive goal responses (although Tolman and Gleitman instead emphasized the perceptual differences between the boxes). Finally, in one of the more bizarre twists on the logic of rg-sg mediation of devaluation effects, using a variant of Miller's (1935) procedure, Fowler and Miller (1963) found that goal box shocks could either depress or enhance test performance, depending on whether the shocks were administered to the front or rear feet of the rats. If the shocks were administered to the front feet, the Rshock would be jumping back, but if they were administered to the back feet, the Rshock was leaping forward. When replaced in the alleys, rats that had received goal shocks to the front feet jumped back, slowing their performance, but those who had received goal shocks to the back feet leaped forward, enhancing their performance. Thus, once again, the rg-sg mediation system not only provided an S-R account for “devaluation” effects but also predicted the nature of the outcome based on characterizations of the conditioned fractional anticipatory response.

The modern devaluation experiment

The modern devaluation experiment is associated with Rescorla's use of the procedure to examine the content of Pavlovian defensive (Rescorla, 1973, 1974) and appetitive (Holland & Rescorla, 1975) conditioning, and then later use of the procedure by Dickinson, Rescorla and their colleagues to examine instrumental reward conditioning (e.g., Adams & Dickinson, 1981; Balleine & Dickinson, 1991; Colwill & Rescorla, 1985b, 1986; Rescorla, 1987). In each of those studies, a modest effect of US devaluation after CS-US pairings was observed. For example, Holland and Rescorla (1975) paired a tone with food pellets, and then devalued the food by either satiating the rats on those pellets or pairing them with high-speed rotation, inducing motion sickness. They found that rats that received such devaluation showed less responding than those that had not received satiation or who had received unpaired presentations of food and rotation. At the same time, rats that had received second-order conditioning of the same tone showed no such effects of either devaluation treatment on second-order CRs, thus ruling out a variety of simple performance-based accounts for reduced responding after devaluation.

Results such as these have encouraged the authors of texts of learning and introductory psychology to assert that conditioning is predominantly S-S. But such a conclusion is not necessarily well-supported. First, devaluation effects are typically small, and perhaps confined to a limited range of training parameters. I have received many communications requesting “how-to” information from researchers who have failed to find devaluation effects. Although generally I have been able to set them on the road to success in obtaining the phenomenon, it is notable how small many devaluation effects are in absolute terms. For example, Figure 3 shows devaluation effects in the last dozen experiments published by members of my laboratory. Most of these studies were done in intact animals although a few represent devaluation effects in rats with sham lesions of various brain regions known to be important for devaluation. The notable feature of these data is how small the differences between control and devaluation conditions are, and how much responding remains after devaluation. In most cases devaluation of the US itself was complete, that is, rats consumed nearly nothing of the food itself. Colwill and Rescorla (1990) have discussed this issue extensively in the context of instrumental devaluation, pointing to the importance of such variables as the similarity between patterns of US delivery in training and devaluation stages. Nevertheless, the residual is quite large in most circumstances (but see Colwill & Rescorla, 1990; Kerfoot et al., 2007).

Which of the following is true about the cognitive processes that produce behavior?

Test results of 12 devaluation experiments from my laboratory. In each experiment an auditory or visual conditioned stimulus (CS) was first paired with food, then the food was either Devalued by pairing it with lithium chloride (LiCl) or Maintained by presenting food and LiCl unpaired, and finally the CS was presented alone in test.

Second, the results of devaluation experiments can vary substantially as a function of the devaluation agent chosen. For example, Holland and Straub (1979) and Holland (1981b) examined the effects of devaluing food by satiation or by pairing it with lithium chloride or high speed rotation. Figure 4 shows consumption of the food and performance of anticipatory food cup entry responses to cues paired with food after devaluation by each of these three methods. In these studies it was encouraging that the amount of reduction in consumption was positively correlated with the amount of reduction in food cup responding during the CS. At the same time there are some conspicuous absences of such correlation as well. For example, Galarce, Crombag, and Holland (2007) trained rats with two auditory cues signaling two foods (sucrose and maltodextrin). One food was paired with LiCl while the other one was not. Finally, rats were tested in the presence of the two tones. The unusual feature of this experiment was that, in different test sessions, either the averted or the nonaverted US was presented in the food cup during the devaluation test. Interestingly, the continuous presence of either of those foods in the food cups had little effect on responding to the CSs. Although not surprisingly, the rats spent more time in the food cup when it contained the nonaverted food than when it contained the averted food, the CS for the nonaverted food enhanced food cup responding more than the cue for the averted food regardless of which food was physically present. Thus, responding based on goal expectancy may occur despite clearly disconfirmatory goal evidence.

Which of the following is true about the cognitive processes that produce behavior?

Left bars: Performance of food cup conditioned responses (CRs) to the conditioned stimulus (CS) alone in a test session after food was devalued by pairing it with lithium chloride (LiCl) or high speed rotation, or by satiation on that food. Right bars: Number of food pellets consumed in 10 min after 50 food pellets were placed in the food cup in the experimental chambers. Responding in the Maintain condition is a composite of responding to a test CS that was presented after unpaired presentations of food and LiCl or rotation, or in the absence of food satiation. Data are from Holland (1981b) and Holland and Straub (1979).

Another concern in devaluation experiments is the measure chosen to reflect outcome expectancy. First, some measures are more sensitive to devaluation than others. Holland and Straub (1979) found that later-chain behaviors (such as contacting and picking up the pellet) showed more sensitivity to food devaluation by pairings with LiCl than early-chain behaviors (approaching the food cup). Interestingly, some more indirect measures of learning, which are often thought to reflect conditioned incentive motivational properties, seem especially insensitive to devaluation. For example, Holland (2004) and Rescorla (1994) found no evidence that Pavlovian-instrumental transfer was affected by devaluation of the reinforcer. In Holland's (2004) study, after training with two CSs, two USs, and two instrumental responses, pairings of one US with LiCl reduced baseline levels of instrumental responding for that US, and reduced food cup responses to the CS that predicted that US, but had no effect on the extent to which that CS enhanced instrumental responding that normally earned that US. Likewise, Parkinson et al. (2005) found that although after repeated testing, US devaluation thoroughly eliminated responding evoked by a CS, that CS did not lose its ability to serve as a secondary or conditioned reinforcer.

Second, response measures and devaluation methods may interact. For example, although Holland and Straub (1979) found that devaluation of food by LiCl decreased food cup entry more than devaluation by high-speed rotation, rearing to visual cues and startle responses to auditory cues were more affected by high-spend rotation than by food-LiCl pairings. By contrast, satiation reduced both classes of responding by an intermediate degree. Clearly “US value” is not unidimensional in its relation to behavior systems. Different behavioral responses might be mediated by different sensory properties of US representations, which in turn may be differentially sensitive to various devaluation methods. I return to this issue briefly in section 3.

Third, appetitive and consummatory CRs may respond to reinforcer devaluation in quite different ways. Although as noted in Figure 2, food cup approach responses often show relatively little sensitivity to reinforcer devaluation, consummatory CRs such as licking and a variety of so-called taste reactivity (TR; Berridge, 2000; Grill & Norgren, 1978) measures thought to reflect hedonic or palatability aspects of rewards are often are more completely eliminated. Using rats fitted with chronic intraoral cannulas, Kerfoot et al. (2007) paired a tone CS with the intraoral delivery of sucrose solution in a single session in two groups of rats and presented those events unpaired in other rats. Rats in the Devalue group then were made ill by a LiCl injection after a single session of intraoral delivery of sucrose in the absence of tones, while rats in the Maintain group received sucrose and LiCl on separate days. Control rats trained with tone and sucrose unpaired also received sucrose and LiCl either paired or unpaired in this phase. Finally, all rats were tested with pairings of the tone with plain water infusions. Rats in the Maintain group showed high levels of TR measures of positive (hedonic) affect on test trials (Figure 5a). Although these responses were especially evident in the presence of water, which provided a substrate for their performance, they also occurred during the tone alone, prior to the infusion of water on each trial. By contrast, rats in the Devalue group showed no more licking or positive TRs than Control rats. Thus, devaluation was complete. In addition, these rats showed substantially more negative TRs, such as gaping, and abbreviated licks, which are normally shown in response to averted flavors themselves, than either the Control or Maintain rats (Figure 5b). The occurrence of these negative TRs is especially important because Devalue rats had never previously performed negative TRs to either the tone or the sucrose. At the time of tone-sucrose pairing, and during sucrose-alone administration prior to LiCl injection, sucrose elicited only positive TRs, and sucrose was not presented again after the single LiCl injection. Thus, the display of negative TR responses by Paired-Devalue rats in test could not reflect S-R learning, even via S-R (Osgood, 1953) or rg-sg (Miller, 1935) mediation mechanisms described earlier.

Which of the following is true about the cognitive processes that produce behavior?

Positive (panel A) and negative (panel B) taste-reactivity (TR) responses during test presentations of a tone conditioned stimulus (CS) and plain water in a devaluation test. Rats first received tone→sucrose pairings followed by either sucrose-lithium chloride (LiCl) pairings (Devalue) or unpaired presentations of those events (Maintain). Control rats received tone and sucrose unpaired in training followed by sucrose and LiCl either paired or unpaired. All fluids were delivered intraorally. Data are from Kerfoot et al., 2007.

It is tempting to speculate from Kerfoot et al.'s (2007) results that learned consummatory or TR responses are especially sensitive to devaluation. However, a simpler possibility is that the intraoral delivery of the reinforcer, and the use of that delivery method during both training and devaluation phases, are more critical factors. Consistent with that possibility, Colwill and Rescorla (1990) found nearly complete devaluation of operant lever pressing after training and devaluation with intraoral reinforcer delivery. More recently, Holland, Lasseter and Agarwal (2008) examined consummatory and TR responses in devaluation experiments with either intraoral or conventional (to a recessed cup) delivery of sucrose reinforcers. The food cups were equipped with cameras to permit measuring TR responses in the cup condition as well as the intraoral delivery condition. Positive TR responses showed the same pattern regardless of sucrose administration method. With both intraoral and cup delivery methods, rats in the Devalue condition showed nearly complete loss of positive TRs after sucrose devaluation, relative to rats in the Maintain condition, and only rats in the Devalue condition displayed negative TR responses (Figures 6a and 6c). At the same time, food cup entry showed only partial sensitivity to sucrose devaluation. In addition, in both delivery conditions, simple licking responses, characterized by Berridge (2000) as consummatory but nonhedonic, showed moderate but not complete sensitivity to devaluation. Thus, the results of Holland et al.'s (2008) study suggest that, regardless of reinforcer delivery method, evaluative TR responses are especially sensitive to devaluation, compared to nonevaluative consummatory responses (simple licking), and (with conventional delivery) to appetitive responses (food cup entry).

Which of the following is true about the cognitive processes that produce behavior?

A: Food cup and positive taste-reactivity (TR) responding during test presentations on tone and plain water after minimal numbers of tone-sucrose pairings, followed by either pairings (Devalue) of sucrose with lithium chloride (LiCl) or unpaired presentations of those events (Maintain). B: Same as panel A, except after extensive tone-sucrose training. C: Simple (nonevaluative) lick, positive TR and negative TR responding during test presentations of tone and plain water, after minimal numbers of tone-sucrose pairings, followed by either pairings (Devalue) of sucrose with LiCl or unpaired presentations of those events (Maintain). D: Same as panel C, except after extensive tone-sucrose training. All fluids were delivered to a recessed food cup (panels A and B; Holland et al., 2008, Experiment 1) or intraorally (panels C and D; Holland et al., 2008, Experiment 2).

A second purpose of Holland et al.'s (2008) study was to examine the effect of the amount of training on devaluation performance, with all of these measures. In the context of instrumental learning, it is often claimed (e.g., Adams, 1982; Kimble & Perlmuter, 1970; but see Colwill & Rescorla, 1985a; Colwill & Triola, 2002; Holland, 2004) that learned responding becomes increasingly immune to reinforcer devaluation with extended training. Indeed, one of Tolman's (1933) alibis (his word) for not obtaining a devaluation effect was that his rats had been overtrained, and had become “fixated” on running. We have never observed such an effect in Pavlovian conditioning; that is, Pavlovian CRs are at least as sensitive to food devaluation by pairings with LiCl after extended training as after minimal training (e.g., Holland, 1998; 2004, 2005). Nevertheless, we have frequently observed dramatic loss in the ability of a Pavlovian CS to participate in “mediated learning” as training proceeds.

Holland (1981a, 1990) found that mediated food aversion learning may be established when an associatively-activated representation of food is paired with illness (see Holland & Wheeler, in press, for a recent review). For example, in one experiment (Holland, 1981a), rats first received pairings of two auditory CSs with two food reinforcers that differed only in flavor (e.g., tone → wintergreen-flavored sucrose and noise → peppermint-flavored sucrose), to give those CSs the ability to selectively activate representations of one or the other of those reinforcers. Next, presentations of one of the auditory CSs alone were paired with the injection of the toxin lithium chloride (LiCl), in the absence of any flavors. Subsequent consumption tests showed the establishment of a mild aversion to the food whose CS partner had been paired with illness, as if the CS's activation of a representation of food just prior to the induction of illness permitted the formation of food-illness associations. Later, Holland (1990, 1998, 2005) found that although CS-LiCl pairings established significant mediated food aversions after 16-24 CS-food pairings, rats trained with 40 or more CS-food pairings showed no evidence for such mediated food aversions. At the same time, other rats trained with as many as 160 CS-US pairings continued to display evidence for S-S tone-food learning, as indicated by reinforcer devaluation procedures. These results led Holland (1990, 1998, 2005) to suggest that early in training, CSs activate a wide range of processing usually activated by the US itself, including early-stage perceptual processing, but as training continues, that access narrows to more limited sets of processing systems. Thus, within this view, mediated flavor aversion learning was not observed later in training because the CS no longer activated perceptual processing of the absent flavor, which could enter into associations with illness.

The logic behind Holland et al.'s (2008) studies was that if the conditioning of TR responses to cues paired with sucrose reflects the cues' activation of perceptual processing of the absent sucrose, and that such activation occurs only early in training, then only a minimally-trained CS would control these positive evaluative responses, and a more extensively-trained CS would not. In each of their experiments, Holland et al. (2008) trained each rat with two CSs, one paired with the sucrose US 16 times and the other paired with it 112 times. Figures 6a and 6c show the response to the minimally-trained CS (described earlier), whereas Figures 6b and 6c show the response to the extensively-trained CS. As anticipated, the display of conditioned TR responses was observed only after minimal training. At the same time, appetitive and nonevaluative consummatory responses (simple licks) were sensitive to sucrose devaluation after either minimal or extensive training.

These data, like those from mediated learning studies, are consistent with the view (Holland, 1990) that early in training the CS activates perceptual processing normally activated by the food itself. Within this perspective, any function normally controlled by the taste of the US should be accessible by the CS. Thus, mediated learning of an aversion to that reinforcer would be expected if the CS was paired with LiCl injection, because the rat would experience illness after tasting the absent US. Similarly, that CS should control evaluative TR responses appropriate to that taste, as observed by Holland et al. (2008) and Kerfoot et al. (2007). Because the CS would activate pre-evaluative perceptual processing of the food US, the TR observed would be consistent with the current evaluation of the taste of that food, and thus would be sensitive to post-training changes in reinforcer value, as observed in those studies. By contrast, as training continues, the CS loses access to this low-level perceptual processing of the US, and instead activates aspects of reinforcer processing that do not include perceptual experience itself or the ability to control evaluative responses appropriate to the current evaluation of that perceptual experience. These more conceptual aspects of the US nevertheless must include both sensory and some sort of motivational information, because reinforcer devaluation effects remain highly US-specific even after the CSs lose their ability to control TR responses or mediate new taste aversion learning (Holland, 1998).

Clearly, there is more to the story than S-S versus S-R associations. The representations activated by CSs after minimal and extended training can be distinguished by their ability to participate in new (mediated) learning and by their ability to control appropriate evaluative TR responses. Nevertheless, even after extended training, CSs can activate reinforcer representations that include rich sensory information, which can subserve reinforcer devaluation effects. This distinction resembles one suggested by Konorski (1967, e.g., pp. 170-181), who distinguished between the associative activation of an “hallucination” and of an “image” by a CS. In the former case, the CS was thought to evoke activity of sensory “projective” units, which could resemble the activity of those units when they were activated by the US itself. By contrast, in the latter case, the CS activated only “gnostic” units which coded information about the US, but which did not normally generate perceptual processing. Casually speaking, an hallucination includes sensory experience of an absent object whereas one can imagine that object and its sensory properties without experiencing them.

III. Can neural systems analysis help us understand the contents of learning?

In section I, I noted that Spence disparaged physiologizing, even Hull's, as more likely to be misleading than useful in characterizing mathematical and behavioral aspects of habit learning. However, the results of the devaluation tasks I've described imply that associative learning can engage a variety of levels of neural processing. Plasticity is not unique to any of the circles in Figure 1b. For example, studies of neural plasticity in eyelid conditioning (e.g., Medina et al., 2000; Steinmetz et al., 1989; Thompson et al., 1998) have identified cerebellar sites of plasticity that are characterizable as being within “output” paths and hence more “S-R”, as well as other sites earlier in the processing stream. A reasonable research program might imagine the circumstances under which cues come to control processing in different brain systems, which preferentially control different aspects of behavior. Understanding information flow in the brain may help us understand why, for example, some products of learning, such as the control of TR responses and the ability to support mediated learning, are related in their rapid bitonic acquisition functions, why others seem to maintain sensitivity to devaluation over extensive training, and still others seem to lose such sensitivity over training.

A variety of neurobiological techniques have been harnessed to provide information relevant to understanding the contents of learning. The methods of brain stimulation have gone far beyond the crude stimulation of motor cortex, as in Loucks's (1935) experiments. For example, in the neural systems analysis of eyelid conditioning, the replacement of real events such as CS and US (or both, e.g., Steinmetz et al., 1989) with brain stimulation has revealed events critical for eyelid conditioning. Similarly, patterned stimulation of amygdala has helped elucidate the mechanisms of acquisition and extinction of fear conditioning (Vidal-Gonzalez et al., 2006).

Precise, pharmacologically selective lesions and reversible inactivations of various cell groups can selectivity alter rats' sensitivity to devaluation, without affecting other aspects of performance. For example, damage to components of circuitry including the basolateral amygdala (BLA), orbitofrontal cortex (OFC), subregions of the medialprefrontal cortex, and in some situations the mediodorsal thalamus, interfere with rats' Pavlovian devaluation performance without affecting the acquisition of food-cup CRs or taste aversions (Gallagher et al., 1999; Hatfield et al., 1996; Maddux & Holland, 2007; Pickens, 2008; Pickens et al., 2003). Furthermore, these studies indicate that different portions of that circuitry have specialized functions within the devaluation task. For example, for rats to display devaluation effects with single-reinforcer procedures, BLA function is needed during initial CS-food pairings, but not thereafter. That is, if rats acquired CS-food associations while BLA function is intact, then subsequent lesions do not impair performance on this form of devaluation, as if BLA function is needed for rats to acquire S-S rather than S-R associations, but not to use previously-established S-S associations to control behavior (Pickens et al., 2003). By contrast, OFC function appears to be critical for the expression as well as the acquisition of such associations in devaluation tasks. Lesions of OFC disrupt devaluation performance (that is, rats fail to spontaneously reduce responding to the CS after US devaluation) whether they are performed prior to CS-food, food aversion, or final test phases (Pickens et al., 2003, 2005). Finally, involvement of these brain regions may vary with variations in task demands. For example, although when a single cue-reinforcer combination is used, BLA function is not required once cue-reinforcer associations are established, when two or more cues and reinforcers are used, BLA function must also be intact at the time of taste aversion training and/or devaluation testing (Johnson et al., 2007). Similarly, Pickens (2008) found that function of mediodorsal thalamus was important for performance on a Pavlovian devaluation task only if that task required a strategy shift from a previous task.

The results of recent electrophysiological recording studies also suggest that conditioning procedures may establish a variety of types of associations, both within and across interconnected brain regions. For example, using an odor-cued discrimination task, Schoenbaum, Chiba, and Gallagher (1998) found neurons in BLA and OFC that might be characterized as reflecting S-S and response-stimulus (R-S) associations. These neurons initially responded selectively to one of the two reinforcers used in the task, but over the course of training their activity came under the control of either a particular odor cue or a particular response. Later studies showed that the nature of neuronal coding of stimulus- or response-outcome information in these two brain regions depended on communication between them (e.g., Saddoris, Gallagher, & Schoenbaum, 2005; Schoenbaum, et al., 2003). Most recently, Furuyashiki, Holland, and Gallagher (2008) also contrasted the activity of OFC neurons that coded outcome information with those that coded response information. These latter neurons responded specifically to particular responses but not to particular reinforcers. Although these latter neurons cannot be construed as a substrate for S-R associations, because they increased their activity only after performance of the coded response, their existence shows that at least in some brain regions, individual neurons may code a variety of kinds of task information.

Finally, we have been using techniques of immediate-early gene expression to relate brain function to performance in simple devaluation (and other) tasks. Our intent is to relate variations in performance under different conditions to differences in brain systems engaged under those conditions. Post-mortem tissue analysis can reveal detailed information about individual neurons that were active (Lee et al., 2005) or undergoing plastic change (Guzowski & Worley, 2001; Petrovich et al., 2005) during fairly restricted time intervals, such as a particular behavioral test episode, before sacrifice. For example, after assessing TR responses to a minimally-trained tone CS, in the absence of sucrose, Kerfoot et al. (2007; Figure 5) sacrificed their rats to examine the expression of FOS, the protein product of the activity-dependent immediate-early gene c-fos. They found learning- and devaluation-dependent FOS expression in a number of brain regions known from lesion experiments (just described) to be critical for learning that is sensitive to outcome devaluation (basolateral amygdala and orbitofrontal cortex; Holland & Gallagher, 2004), in regions related to the display of TR responses (accumbens shell; Reynolds & Berridge, 2002), and in regions related to processing of taste information (gustatory cortex; Kiefer & Orr, 1992). These last observations are especially interesting from our earlier suggestion that conditioning may endow CSs with the ability to activate perceptual processing of absent USs. If as a result of tone-sucrose pairings, the tone aroused the perception of sweetness, making the plain water taste sweet, rats in both the Devalue and Maintain conditions might be expected to show enhanced FOS activity in gustatory cortex, which they did. Furthermore, because rats in the Maintain condition would respond to sweet with appetitive taste-reactivity responses, the tone alone would also elicit those responses and FOS activity in a portion of the accumbens shell correlated with such responses. By contrast, because rats in the Devalue condition would respond to sweet with aversive responses, the tone would provoke aversive responses and FOS in another subregion of accumbens shell correlated with aversive TR responses. All of these outcomes were observed. Given Holland et al.'s (2008) findings, it would be intriguing to determine if a more extensively-trained CS, which fails to elicit appetitive TR responses in training or evoke negative TR responses after devaluation, and which does not support mediated taste-aversion learning, would also fail to induce FOS in the accumbens shell or gustatory cortex.

Interestingly, although Kerfoot et al. (2007) found conditioning-dependent FOS expression in the central nucleus of the amygdala (CeA), this expression was unaffected by devaluation condition (devalued or maintained). This observation is notable for three reasons. First, it suggests that associatively-activated taste memories are not completely interchangeable with the tastes themselves. Taste and illness information converge in this region (Bernstein & Koh, 2007; Yamamoto, 2007), and CeA is known to be important in many aspects of the learning and expression of taste aversions (Lamprecht and Dudai, 2000; Yamamoto, 2007). However, if an associatively-activated taste memory instigated these aspects of taste processing, we would have observed differential CeA activity in Maintain and Devalue conditions. Second, lesions of CeA do not affect devaluation performance (Hatfield et al., 1996). Third, CeA function is critical for the acquisition of conditioned ORs, which in other circumstances Holland and Straub (1979) found to be relatively insensitive to devaluation by food-LiCl pairings. Taken together, these last two observations suggest converging evidence that relates differences in the sensitivity of conditioned ORs and food-related CRs to LiCl-based devaluation procedures (Holland & Straub, 1979) to differences in brain circuitry subserving those learned responses. Differences in operating characteristics of those systems may determine the differential sensitivity of different behavioral systems to parameters of devaluation.

These neural systems studies of devaluation and related phenomena make it clear that questions of the contents of learning are complex. Even in apparently simple behavioral systems such as the eyeblink reflex, opportunities for plastic change abound in the neural systems that subserve them. The consequences of experimental manipulations of environmental stimuli for learning may vary considerably across these systems and their components. It has been increasingly difficult to distinguish between S-S and S-R associations with any confidence on the basis of any single behavioral assay or aspect of neural processing. Characterization of multiple behavioral and neural consequences of variations in associative learning procedures should lead to richer, if perhaps less fathomable, descriptions of the nature of learning.

Conclusions

Within the context of S-S versus S-R associations in learning, the pendulum has swung decidedly to the cognitive side since Spence's day. There are now many demonstrations of encoding of detailed sensory aspects of reinforcers by CSs, from a range of behavioral paradigms, including sensory preconditioning, reinforcer devaluation, mediated learning and others. Perhaps because of these many examples, the acquisition of S-R associations is seldom directly examined. The S-S association has become the default, and absence of evidence for it is no longer taken as evidence for S-R associations, as it once was. For example, although devaluation effects are often small, the prevailing opinion seems to be that the residual responding after devaluation reflects inadequacy of the devaluation procedure rather than S-R learning. This opinion may not be well-grounded. There are many examples of devaluation failure, and it is notable that few of the many brain lesions that impair devaluation performance produce general reductions in the measured CRs. Instead, animals with those lesions perform at normal pre-devaluation levels, but fail to adjust their responding downward after devaluation. Thus, these lesions apparently alter what is learned from S-S to S-R associations without affecting the most obvious aspects of CR form. Furthermore, it is clear from neural systems studies that response-end plasticity occurs, and that through associative learning, neurons can come to code aspects of behavior as well as of stimuli.

What of Spence's other 5 characterizations of “Cognitive versus stimulus-response theories of learning”? In terms of metaphors, both telephone switchboards and map control rooms have become anachronisms. As in Spence's day, most modern quantitative models of learning and behavior emphasize associative structures that could be as equally construed as S-S or S-R. Rich representations of tasks and outcomes in learning and flexibility in their use in performance are widely assumed to be the rule, but are probably no better specified within formal models than the maps of yesteryear.

Spence was probably wise to ignore the neurophysiologizing of his day, because it had little basis in the study of the nervous system. Neurophysiological characterizations of learning are on much firmer ground today than they were in 1950, and are far more likely to be based on data than on speculation. Electrophysiological recoding methods can characterize neuronal coding of different kinds of information, and functional anatomical methods can be used to examine the bases of initially hypothetical brain networks underlying aspects of associative learning. Although many modern reductionist analyses of learning emphasize cellular and molecular changes underlying learning, these analyses are seldom taken as a substitute for behavioral analysis. Although I believe that the best way to further our understanding of behavior is to study behavior, neural analyses can often give us hints and clues about behavioral processes and organization. Perhaps just as important, the study of the neural basis of learning can connect the study of learning to a much broader scientific base and community.

In modern learning research, our choices of stimulus variables are much more eclectic than in Spence's day, with the same laboratories often investigating perceptual, motivational and other variables, and delving equally into determining the contents and conditions of learning. We are at least more receptive to the idea that the conditions and contents of learning are not independent, and are more likely to pursue programs that examine, for example, the conditions that favor the formation of S-S or S-R learning, than to ask simply, is learning S-S or S-R?

In addressing the state of theorizing in the study of learning, Spence (1950, p. 171) lamented that

“one almost gets the impression that the major goal is to prove the other conception wrong than to try to arrive at a more comprehensive interpretation of the phenomena. Psychologists interested and appreciative of the role of theory in the development of a scientific body of knowledge should resist such temptations… elimination of all theorizing, would certainly not help to speed up the acquisition of a scientific body of knowledge about learning.”

I think it is fair to say that the conflict between cognitive and stimulus-response theories of learning eventually went beyond this early competitive phase and stimulated great advances in our understanding of associative learning.

Acknowledgments

In this article, I have borrowed heavily from Spence's (1950) original language, but have provided marked quotations only for particularly lengthy or salient passages. I thank the organizers of this meeting, Edward Wasserman, John Freeman, and especially, Allan Wagner, whose excellent tutelage in the early 1970s endured long enough for me to recall these classic studies so many years later. Preparation of this manuscript was supported in part by NIH grants MH53667 and MH65879.

References

  • Adams CD. Variations in the sensitivity of instrumental responding to reinforcer devaluation. Quarterly Journal of Experimental Psychology. 1982;34B:77–98. [Google Scholar]
  • Adams C, Dickinson A. Instrumental responding following reinforcer devaluation. Quarterly Journal of Experimental Psychology. 1981;33B:109–121. [Google Scholar]
  • Balleine BW, Dickinson A. Instrumental performance following reinforcer devaluation depends upon incentive learning. Quarterly Journal of Experimental Psychology. 1991;40B:279–296. [Google Scholar]
  • Bernstein I, Koh MT. Molecular signaling during taste aversion learning. Chemical Senses. 2007;32:99–103. [PubMed] [Google Scholar]
  • Berridge KC. Measuring hedonic impact in animals and infants: microstructure of affective taste reactivity patterns. Neuroscience and Biobehavioral Reviews. 2000;24:173–198. [PubMed] [Google Scholar]
  • Brogden WJ. Sensory pre-conditioning of human subjects. Journal of Experimental Psychology. 1947;37:527–539. [PubMed] [Google Scholar]
  • Colwill RM, Rescorla RA. Instrumental responding remains sensitive to reinforcer devaluation after extensive training. Journal of Experimental Psychology: Animal Behavior Processes. 1985;11:520–526. (a) [Google Scholar]
  • Colwill RM, Rescorla RA. Post-conditioning devaluation of a reinforcer affects instrumental responding. Journal of Experimental Psychology: Animal Behavior Processes. 1985;11:120–132. (b) [Google Scholar]
  • Colwill RM, Rescorla RA. Associative structures in instrumental learning. In: Bower GH, editor. ThePsychology of Learning and Motivation. Vol. 20. San Diego, CA: Academic Press; 1986. pp. 55–104. [Google Scholar]
  • Colwill RM, Rescorla RA. Effect of reinforcer devaluation on discriminative control of instrumental behavior. Journal of Experimental Psychology: Animal Behavior Processes. 1990;16:40–47. [PubMed] [Google Scholar]
  • Colwill RM, Triola SM. Instrumental responding remains under the control of the consequent outcome after extended training. Behavioural Processes. 2002;57:51–64. [PubMed] [Google Scholar]
  • Cook RG, Riley DA, Brown MF. Spatial and configural factors in compound stimulus processing by pigeons. Animal Learning & Behavior. 1992;20:41–55. [Google Scholar]
  • Doty RW, Giurgea C. Conditioned reflexes established by coupling electrical excitation of two cortical areas. In: Delafresnaye JF, editor. Brain mechanisms and learning. Oxford: Blackwell; 1961. pp. 133–151. [Google Scholar]
  • Elliot MH. The effect of change of reward on the maze performance of rats. University of California Publications in Psychology. 1928;4:19–30. [Google Scholar]
  • Finch G. Salivary conditioning in atropinized dogs. American Journal of Psychology. 1938;124:136–141. [Google Scholar]
  • Fowler H, Miller NE. Facilitation and inhibition of runway performance by hind- and forepaw shock of various intensities. Journal of Comparative and Physiological Psychology. 1963;56:801–805. [PubMed] [Google Scholar]
  • Furuyashiki T, Holland PC, Gallagher M. Rat orbitofrontal cortex separately encodes response and outcome information during performance of goal-directed behavior. Journal of Neuroscience. 2008 in press. [PMC free article] [PubMed] [Google Scholar]
  • Galarce EM, Crombag HS, Holland PC. Reinforcer-specificity of appetitive and consummatory behavior of rats after Pavlovian conditioning with food reinforcers. Physiology and Behavior. 2007;91:95–105. [PMC free article] [PubMed] [Google Scholar]
  • Gallagher M, McMahan RW, Schoenbaum G. Orbitofrontal cortex and representation of incentive value in associative learning. Journal of Neuroscience. 1999;19:6610–6614. [PMC free article] [PubMed] [Google Scholar]
  • Grill HJ, Norgren R. The taste reactivity test. I. Mimetic responses to gustatory stimuli in neurologically normal rats. Brain Research. 1978;143:263–279. [PubMed] [Google Scholar]
  • Guthrie ER. The psychology of learning. New York: Harper & Row; 1935. [Google Scholar]
  • Guthrie ER. Psychological facts and psychological theories. Psychological Bulletin. 1946;43:1–20. [PubMed] [Google Scholar]
  • Guzowski JF, Worley PF. Cellular compartment analysis of temporal activity by Fluorescent In Situ Hybidization (catFISH) In: Taylor GP, editor. Current Protocols in Neuroscience. New York: John Wiley & Sons, Inc.; 2001. pp. 1–16. [Google Scholar]
  • Hatfield T, Han JS, Conley M, Gallagher M, Holland PC. Neurotoxic lesions of the basolateral, but not central, amygdala interfere with Pavlovian second order conditioning and reinforcer devaluation effects. Journal of Neuroscience. 1996;16:5256–5265. [PMC free article] [PubMed] [Google Scholar]
  • Holland PC. Acquisition of representation mediated conditioned food aversions. Learning and Motivation. 1981;12:1–18. (a) [Google Scholar]
  • Holland PC. The effects of satiation after first- and second-order appetitive conditioning. Pavlovian Journal of Biological Science. 1981;16:18–24. (b) [PubMed] [Google Scholar]
  • Holland PC. Event representation in Pavlovian conditioning: Image and action. Cognition. 1990;37:105–131. [PubMed] [Google Scholar]
  • Holland PC. Brain mechanisms for changes in processing of conditioned stimuli in Pavlovian conditioning: Implications for behavior theory. Animal Learning & Behavior. 1997;25:373–399. [Google Scholar]
  • Holland PC. Amount of training affects associatively-activated event representation. Neuropharmacology. 1998;37:461–469. [PubMed] [Google Scholar]
  • Holland PC. Relations between Pavlovian-instrumental transfer and reinforcer devaluation. Journal of Experimental Psychology: Animal Behavior Processes. 2004;30:104–117. [PubMed] [Google Scholar]
  • Holland PC. Amount of training effects in representation-mediated food aversion learning: No evidence for a role for associability changes. Learning & Behavior. 2005;33:464–478. [PMC free article] [PubMed] [Google Scholar]
  • Holland PC, Forbes DT. Control of conditional discrimination performance by CS-evoked event representations. Animal Learning & Behavior. 1982;10:249–256. [Google Scholar]
  • Holland PC, Gallagher M. Amygdala-frontal interactions and reward expectancy. Current Opinion in Neurobiology. 2004;14:148–155. [PubMed] [Google Scholar]
  • Holland PC, Lasseter H, Agarwal I. Amount of training and cue-evoked taste-reactivity responding in reinforcer devaluation. Journal of Experimental Psychology: Animal Behavior Processes. 2008;34:119–132. [PMC free article] [PubMed] [Google Scholar]
  • Holland PC, Rescorla RA. The effect of two ways of devaluing the unconditioned stimulus after first- and second-order appetitive conditioning. Journal of Experimental Psychology: Animal Behavior Processes. 1975;1:355–363. [PubMed] [Google Scholar]
  • Holland PC, Straub JJ. Differential effects of two ways of devaluing the unconditioned stimulus after Pavlovian appetitive conditioning. Journal of Experimental Psychology: Animal Behavior Processes. 1979;5:65–78. [PubMed] [Google Scholar]
  • Holland PC, Wheeler DS. Representation-mediated food aversions. In: Reilly S, Schachtman T, editors. Conditioned taste aversion: Behavioral and neural processes. Oxford: Oxford University Press; in press. [Google Scholar]
  • Hull CL. Principles of Behavior. New York: Appleton-Century Crofts; 1943. [Google Scholar]
  • Johnson AW, Gallagher M, Holland PC. Examining the role of the basolateral amygdala in the expression of reinforcer devaluation. Society for Neuroscience. 2007 poster 934.3/xx11. [Google Scholar]
  • Kerfoot EC, Agarwal I, Lee HJ, Holland PC. Control of appetitive and aversive taste-reactivity responses by an auditory conditioned stimulus in a devaluation task: a FOS and behavioral analysis. Learning and Memory. 2007;14:581–589. [PMC free article] [PubMed] [Google Scholar]
  • Kiefer SW, Orr MR. Taste avoidance, but not aversion, learning in rats lacking gustatory cortex. Behavioral Neuroscience. 1992;106:140–146. [PubMed] [Google Scholar]
  • Kimble GA, Perlmuter LC. The problem of volition. Psychological Review. 1970;77:361–384. [PubMed] [Google Scholar]
  • Koffka K. The principles of Gestalt psychology. New York: Harcourt, Brace; 1935. [Google Scholar]
  • Kohler W. Dynamics in psychology. New York: Liveright; 1940. [Google Scholar]
  • Konorski J. Integrative activity of the brain. Chicago: University of Chicago Press; 1967. [Google Scholar]
  • Lamprecht R, Dudai Y. The amygdala in conditioned taste aversion: it's there but where. In: Aggleton JP, editor. The amygdala: a functional analysis. Oxford: Oxford University Press; 2000. pp. 331–351. [Google Scholar]
  • Lee HJ, Groshek F, Petrovich GD, Cantalini JP, Gallagher M, Holland PC. Role of amygdalo-nigral circuitry in conditioning of a visual stimulus paired with food. Journal of Neuroscience. 2005;25:3881–3888. [PMC free article] [PubMed] [Google Scholar]
  • Lewin K. Principles of topological psychology. New York: McGraw-Hill; 1936. [Google Scholar]
  • Loucks RB. The experimental delimination of neural structures essential for learning: the attempt to condition striped muscle responses to faradization of the sigmoid gyri. Journal of Psychology. 1935;1:5–44. [Google Scholar]
  • Mackintosh NJ. A theory of attention: Variations in the associability of stimuli with reinforcement. Psychological Review. 1975;82:276–298. [Google Scholar]
  • Maddux JM, Holland PC. Lesions of the medial prefrontal cortex impair Pavlovian devaluation. Society for Neuroscience. 2007 poster 310.7. [Google Scholar]
  • Medina JF, Nores WL, Ohyama T, Mauk MD. Mechanisms of cerebellar learning suggested by eyelid conditioning. Current Opinion in Neurobiology. 2000;10:717–724. [PubMed] [Google Scholar]
  • Miller NE. A reply to “Sign Gestalt or Conditioned Reflex?” Psychological Review. 1935;42:280–292. [Google Scholar]
  • Osgood CE. Method and theory in experimental psychology. New York: Oxford University Press; 1953. [Google Scholar]
  • Packard MG, McGaugh JL. Inactivation of hippocampus or caudate nucleus with lidocaine differentially affects expression of place and response learning. Neurobiology of Learning and Memory. 1996;65:65–72. [PubMed] [Google Scholar]
  • Parkinson JA, Roberts AC, Everitt BJ, di Ciano P. Acquisition of instrumental conditioned reinforcement is resistant to the devaluation of the unconditioned stimulus. Quarterly Journal of Experimental Psychology. 2005;58B:19–30. [PubMed] [Google Scholar]
  • Pearce JM, Hall G. A model for Pavlovian learning: variations in the effectiveness of conditioned but not of unconditioned stimuli. Psychological Review. 1980;106:532–552. [PubMed] [Google Scholar]
  • Petrovich GD, Holland PC, Gallagher M. Amygdalar and prefrontal pathways to the lateral hypothalamus are activated by a learned cue that stimulates eating. Journal of Neuroscience. 2005;25:8295–8302. [PMC free article] [PubMed] [Google Scholar]
  • Pickens CL. A limited role for mediodorsal thalamus in devaluation tasks. Behavioral Neuroscience. 2008 in press. [PMC free article] [PubMed] [Google Scholar]
  • Pickens CL, Holland PC. Conditioning and cognition. Neuroscience and Biobehavioral Reviews. 2004;28:651–661. [PubMed] [Google Scholar]
  • Pickens CL, Saddoris MP, Gallagher M, Holland PC. Orbitofrontal lesions impair use of cue-outcome associations in a devaluation task. Behavioral Neuroscience. 2005;119:317–322. [PMC free article] [PubMed] [Google Scholar]
  • Pickens CL, Saddoris MP, Setlow B, Gallagher M, Holland PC, Schoenbaum G. Different roles for orbitofrontal cortex and basolateral amygdala in a reinforcer devaluation task. Journal of Neuroscience. 2003;23:11078–11084. [PMC free article] [PubMed] [Google Scholar]
  • Poldrack RA, Packard MG. Competition among multiple memory systems: Converging evidence from animal and human brain studies. Neuropsychologia. 2003;41:245–251. [PubMed] [Google Scholar]
  • Rescorla RA. Effect of US habituation following conditioning. Journal of Comparative and Physiological Psychology. 1973;82:137–143. [PubMed] [Google Scholar]
  • Rescorla RA. Effect of inflation of the unconditioned stimulus value following conditioning. Journal of Comparative and Physiological Psychology. 1974;86:101–106. [Google Scholar]
  • Rescorla RA. Pavlovian excitatory and inhibitory conditioning. In: Estes WK, editor. Handbook of learning and cognitive processes. Vol. 2. New York: Erlbaum; 1975. pp. 7–35. [Google Scholar]
  • Rescorla RA. Two perceptual variables in within-event learning. Animal Learning & Behavior. 1986;14:387–392. [Google Scholar]
  • Rescorla RA. A Pavlovian analysis of goal directed behavior. American Psychologist. 1987;42:119–129. [Google Scholar]
  • Rescorla RA. Pavlovian conditioning : It's not what you think it is. American Psychologist. 1988;43:151–160. [PubMed] [Google Scholar]
  • Rescorla RA. Transfer of instrumental control mediated by a devalued outcome. Animal Learning & Behavior. 1994;22:27–33. [Google Scholar]
  • Reynolds SM, Berridge KC. Positive and negative motivation in nucleus accumbens shell: bivalent rostrocaudal gradients for GABA-elicited eating, taste “liking”/“disliking” reactions, place preference/avoidance, and fear. Journal of Neuroscience. 2002;22:7308–7320. [PMC free article] [PubMed] [Google Scholar]
  • Rozeboom WW. “What is learned?”: An empirical enigma” Psychological Review. 1958;65:22–33. [PubMed] [Google Scholar]
  • Saddoris MP, Gallagher M, Schoenbaum G. Rapid associative encoding in basolateral amygdala depends on connections with orbitofrontal cortex. Neuron. 2005;46:321–331. [PubMed] [Google Scholar]
  • Schoenbaum G, Chiba AA, Gallagher M. Orbitofrontral cortex and basolateral amygdala encode expected outcomes during learning. Nature Neuroscience. 1998;1:155–159. [PubMed] [Google Scholar]
  • Schoenbaum G, Setlow B, Saddoris MP, Gallagher M. Encoding predicted outcome and acquired value in orbitofrontal cortex during cue sampling depends upon input from basolateral amygdala. Neuron. 2003;39:855–867. [PubMed] [Google Scholar]
  • Siegel S. Overtraining and transfer processes. Journal of Comparative and Physiological Psychology. 1967;64:471–477. [PubMed] [Google Scholar]
  • Spence KW. The nature of discrimination learning in animals. Psychological Review. 1936;43:427–49. [Google Scholar]
  • Spence KW. Cognitive versus stimulus-response theories of learning. Psychological Review. 1950;57:159–72. [PubMed] [Google Scholar]
  • Steinmetz JE, Lavond DG, Thompson RF. Classical-conditioning in rabbits using pontine nucleus stimulation as a conditioned-stimulus and inferior olive stimulation as an unconditioned stimulus. Synapse. 1989;3:225–233. [PubMed] [Google Scholar]
  • Terry WS, Wagner AR. Short term memory for “surprising” vs. “expected” unconditioned stimuli in Pavlovian conditioning. Journal of Experimental Psychology: Animal Behavior Processes. 1975;1:122–133. [PubMed] [Google Scholar]
  • Thomas E. The role of postural adjustments in conditioning of dogs with electrical stimulation of the motor cortex as the unconditioned stimulus. Journal of Comparative and Physiological Psychology. 1971;76:187–198. [PubMed] [Google Scholar]
  • Thompson RF. Sensory preconditioning. In: Thompson RF, Voss JS, editors. Topics in learning and performance. New York: Academic Press; 1972. pp. 105–129. [Google Scholar]
  • Thompson RF, Thompson JK, Kim JJ, Kripa DJ, Shinkman PG. the nature of reinforcement in cerebellar learning. Neurobiology of Learning and Memory. 1998;70:150–176. [PubMed] [Google Scholar]
  • Thorndike EL. Animal intelligence: An experimental study of the associative processes in animals. Psychological Review Monograph Supplements. 1898;2(4, Whole number 8) [Google Scholar]
  • Tinklepaugh OL. An experimental study of representative factors in monkeys. Journal of Comparative and Physiological Psychology. 1928;8:197–236. [Google Scholar]
  • Tolman EC. Purposive behavior in animals and men. New York: Appleton Century; 1932. [Google Scholar]
  • Tolman EC. Sign-gestalt or conditioned reflex. Psychological Review. 1933;40:246–255. [Google Scholar]
  • Tolman EC. Cognitive maps in rats and men. Psychological Review. 1948;55:189–208. [PubMed] [Google Scholar]
  • Tolman EC, Gleitman H. Studies in learning and motivation: I. equal reinforcements in both end-boxes, followed by shock in one end-box. Journal of Experimental Psychology. 1949;39:810–819. [PubMed] [Google Scholar]
  • Tolman EC, Ritchie BF, Kalish D. Studies in spatial learning. I: Orientation and the short-cut. Journal of Experimental Psychology. 1946;36:13–25. (a) [PubMed] [Google Scholar]
  • Tolman EC, Ritchie BF, Kalish D. Studies in spatial learning. II: Place learning versus response learning. Journal of Experimental Psychology. 1946;36:221–229. (b) [PubMed] [Google Scholar]
  • Vidal-Gonzalez I, Vidal-Gonzalez B, Rauch SL, Quirk GJ. Microstimulation reveals opposing influences of prelimbic and infralimbic cortex on the expression of conditioned fear. Learning and Memory. 2006;13:728–733. [PMC free article] [PubMed] [Google Scholar]
  • Wagner AR, Thomas E, Norton T. Conditioning with electrical stimulation of the motor cortex: Evidence of a possible source of motivation. Journal of Comparative and Physiological Psychology. 1967;64:191–199. [PubMed] [Google Scholar]
  • Ward-Robinson J, Hall G. Backward sensory preconditioning. Journal of Experimental Psychology: Animal Behavior Processes. 1996;22:395–404. [Google Scholar]
  • Yamamoto T. Brain regions responsible for the expression of conditioned taste aversion in rats. Chemical Senses. 2007;32:105–109. [PubMed] [Google Scholar]
  • Zener K. The significance of behavior accompanying conditioned salivary secretion for theories of the conditioned response. American Journal of Psychology. 1937;50:384–403. [Google Scholar]
  • Zener K, McCurdy HG. Analysis of motivational factions in conditioned behavior: I. The differential effect of changes in hunger on conditioned, unconditioned, and spontaneous salivary secretion. Journal of Psychology. 1939;8:321–350. [Google Scholar]

Which of the following provides the best definition of cognitive psychology?

Cognitive psychology is an area that focuses on the science of how people think. This branch of psychology explores a wide variety of mental processes, including how people think, use language, attend to information, and perceive their environments.

What are the four interrelated processes involved in observational learning of aggression?

Specifically, Bandura and Jeffrey (1973) described four processes that account for learning from observation: attentional, retention, motor reproduction, and motivational.

Who among the following is likely high in unmitigated communion quizlet?

Who among the following is likely high in unmitigated communion? A. Belinda, a surgeon who sacrifices her own needs and interests to care for all her patients.

Which of the following is a component of the depressive cognitive triad?

Which of the following is a component of the depressive cognitive triad? Interpreting ongoing experiences in a negative light, Pessimistic thoughts about the future, and Negative thoughts about oneself.