RESEARCH ARTICLE

Complementary contributions of basolateral and to value learning under uncertainty Alexandra Stolyarova1*, Alicia Izquierdo1,2,3,4*

1Department of Psychology, University of California, Los Angeles, Los Angeles, United States; 2Integrative Center for Learning and Memory, University of California, Los Angeles, Los Angeles, United States; 3Integrative Center for , University of California, Los Angeles, Los Angeles, United States; 4The Brain Research Institute, University of California, Los Angeles, Los Angeles, United States

Abstract We make choices based on the values of expected outcomes, informed by previous experience in similar settings. When the outcomes of our decisions consistently violate expectations, new learning is needed to maximize rewards. Yet not every surprising event indicates a meaningful change in the environment. Even when conditions are stable overall, outcomes of a single experience can still be unpredictable due to small fluctuations (i.e., expected uncertainty) in reward or costs. In the present work, we investigate causal contributions of the basolateral amygdala (BLA) and orbitofrontal cortex (OFC) in rats to learning under expected outcome uncertainty in a novel delay-based task that incorporates both predictable fluctuations and directional shifts in outcome values. We demonstrate that OFC is required to accurately represent the distribution of wait times to stabilize choice preferences despite trial-by-trial fluctuations in outcomes, whereas BLA is necessary for the facilitation of learning in response to surprising events. *For correspondence: DOI: 10.7554/eLife.27483.001 [email protected] (AS); [email protected] (AI)

Competing interests: The authors declare that no Introduction competing interests exist. Learning to predict rewards is a remarkable evolutionary adaptation that supports flexible behavior in complex and unstable environments. When circumstances change, previously-acquired knowledge Funding: See page 22 may no longer be informative and the behavior needs to be adapted to benefit from novel opportu- Received: 05 April 2017 nities. Frequently, alterations in environmental conditions are not signaled by external cues and can Accepted: 05 July 2017 only be inferred from deviations from anticipated outcomes, that is, surprise signals. Published: 06 July 2017 When making decisions, humans typically attempt to maximize benefits (i.e., amount of reward) Reviewing editor: Geoffrey received per invested resource (i.e., money, time, physical or cognitive effort). We, like many other Schoenbaum, National Institutes animals, compute economic value that takes into account rewards and costs associated with avail- of Health, United States able behavioral options and choose the alternative that is expected to result in outcomes of the highest value based on previous experiences under similar conditions (Padoa-Schioppa and Schoen- Copyright Stolyarova and baum, 2015; Sugrue et al., 2005). When the outcomes of choices consistently violate expectations, Izquierdo. This article is new learning is needed to maximize reward procurement. However, not every unexpected outcome distributed under the terms of is caused by meaningful changes in the environment. Even when conditions are stable overall, out- the Creative Commons Attribution License, which comes of a single experience can still be unpredictable due to small fluctuations (i.e., expected permits unrestricted use and uncertainty) in reward and costs. Such fluctuations complicate surprise-driven learning since animals redistribution provided that the need to distinguish between true changes in the environment from stochastic feedback under other- original author and source are wise stable conditions, known as the problem of change-point detection (Courville et al., 2006; credited. Dayan et al., 2000; Gallistel et al., 2001; Pearce and Hall, 1980; Yu and Dayan, 2005).

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 1 of 27 Research article Neuroscience

eLife digest Nobody likes waiting – we opt for online shopping to avoid standing in lines, grow impatient in traffic, and often prefer restaurants that serve food quickly. When making decisions, humans and other animals try to maximize the benefits by weighing up the costs and rewards associated with a situation. Many regions in the brain help us choose the best options based on quality and size of rewards, and required waiting times. Even before we make decisions, the activity in these brain regions predicts what we will choose. Sometimes, however, unexpected changes can lead to longer waiting times and our preferences suddenly become less desirable. The brain can detect such changes by comparing the outcomes we anticipate to those we experience. When the outcomes are surprising, specific areas in the brain such as the amygdala and the orbitofrontal cortex help us learn to make better choices. However, as surprising events can occur purely by chance, we need to be able to ignore irrelevant surprises and only learn from meaningful ones. Until now, it was not clear whether the amygdala and orbitofrontal cortex play specific roles in successfully learning under such conditions. Stolyarova and Izquierdo trained rats to select between two images and rewarded them with sugar pellets after different delays. If rats chose one of these images they received the rewards after a predictable delay that was about 10 seconds, while choosing the other one produced variable delays – sometimes the time intervals were either very short or very long. Then, the waiting times for one of the alternatives changed unexpectedly. Rats with healthy brains quickly learned to choose the option with the shorter waiting time. Stolyarova and Izquierdo repeated the experiments with rats that had damage in a part of the amygdala. These rats learned more slowly, particularly when the variable option changed for the better. Rats with damage to the orbitofrontal cortex failed to learn at all. Stolyarova and Izquierdo then examined the rats’ behavior during delays. Rats with damage to the orbitofrontal cortex could not distinguish between meaningful and irrelevant surprises and always looked for the food pellet (i.e. anticipated a reward) at the average delay interval. These findings highlight two brain regions that help us distinguish meaningful surprises from irrelevant ones. A next step will be to examine how the amygdala and orbitofrontal cortex interact during learning and see if changes to the activity of these brain regions may affect responses. Advanced methods to non-invasively manipulate brain activity in humans may help people who find it hard to cope with changes; or individuals suffering from substance use disorders, who often struggle to give up drugs that provide them immediate and predictable rewards. DOI: 10.7554/eLife.27483.002

Both the basolateral amygdala (BLA) and orbitofrontal cortex (OFC) participate in flexible reward- directed behavior. Representations of expected outcomes can be decoded from both brain regions during value-based decision making (Conen and Padoa-Schioppa, 2015; Haruno et al., 2014; Padoa-Schioppa, 2007, 2009; Salzman et al., 2007; van Duuren et al., 2009). Amygdala lesions render animals unable to adaptively track changes in reward availability or benefit from profitable periods in the environment (Murray and Izquierdo, 2007; Salinas et al., 1996; Salzman et al., 2007). Furthermore, a recent evaluation of the accumulated literature on BLA in appetitive behavior suggests that this region integrates both current reward value and long-term history information (Wassum and Izquierdo, 2015), and therefore may be particularly well-suited to guide behavior when conditions change. Importantly, single-unit responses in BLA track surprise signals (Roesch et al., 2010) that can drive learning. Similarly, a functionally-intact OFC is required for adaptive responses to changes in outcome val- ues (Elliott et al., 2000; Izquierdo and Murray, 2010; Murray and Izquierdo, 2007). Impairments produced by OFC lesions have been widely attributed to diminished cognitive flexibility or inhibitory control deficits (Bari and Robbins, 2013; Dalley et al., 2004; Elliott and Deakin, 2005; Winstan- ley, 2007). However, this view has been challenged recently by observations that selective medial OFC lesions cause potentiated switching between different option alternatives, rather than a failure to disengage from previously acquired behavior (Walton et al., 2010, 2011). Indeed, there is increasing evidence that certain sectors of OFC may not exert a canonical inhibitory control over

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 2 of 27 Research article Neuroscience action, but may instead contribute outcome representations predicted by specific cues in the envi- ronment and update expectations in response to surprising feedback (Izquierdo et al., 2017; Marquardt et al., 2017; Riceberg and Shapiro, 2012, 2017; Rudebeck and Murray, 2014; Stalnaker et al., 2015). Despite important contributions of both the BLA and OFC to several forms of adaptive value learning, some learning tasks progress normally without the recruitment of these brain regions. For example, the OFC is not required for acquisition of simple stimulus-outcome associations, both in Pavlovian and instrumental context, or for unblocking driven by differences in value when outcomes are certain and predictable. However, the OFC is needed for adaptive behavior that requires inte- gration of information from different sources, particularly when current outcomes need to be com- pared with a history in a different context (or state) as in devaluation paradigms (Izquierdo et al., 2004; McDannald et al., 2011, 2005; Stalnaker et al., 2015). Similarly, as has been shown in rats, BLA has an important role in early learning or decision making under ambiguous outcomes (Hart and Izquierdo, 2017; Ostrander et al., 2011), and seems to play a limited role in choice behavior when these outcomes are known or reinforced through extended training. These observa- tions hint at important roles for BLA and OFC in learning under conditions of uncertainty. Yet little is known about unique contributions of these brain regions to value learning when outcomes are fluc- tuating even under stable conditions (i.e., when there is expected uncertainty in outcome values). Furthermore, the functional dissociation between different OFC subregions (e.g. ventromedial vs. lateral) is presently debated (Dalton et al., 2016; Elliott et al., 2000; Morris et al., 2016). Recently-developed computational models based on reinforcement learning (RL) (Diederen and Schultz, 2015; Khamassi et al., 2011; Preuschoff and Bossaerts, 2007) and Bayesian inference principles (Behrens et al., 2007; Nassar et al., 2010) are well suited to test for unique contributions of different brain regions to value learning under uncertainty. These models rely on learning in response to surprise, or the deviation between expected and observed outcomes (i.e., reward pre- diction errors, RPEs); the learning rate, in turn, determines the degree to which prediction errors affect value estimates. Importantly, the RL principles do not only account for animal behavior, but are also reflected in underlying neuronal activity (Lee et al., 2012; Niv et al., 2015). In the present work, we first developed a novel delay-based behavioral paradigm to investigate the effects of expected outcome uncertainty on learning in rats. We demonstrated that rats can detect true changes in outcome values even when they occur against a background of stochastic feedback. Such behavioral complexity in rodents allowed us to assess causal contributions of the BLA and OFC to value learning under expected outcome uncertainty. Specifically, we examined the neuroadaptations that occur in these brain regions in response to experience with different levels of environmental uncertainty and employed fine-grained behavioral analyses partnered with computa- tional modeling of trial-by-trial performance of OFC- and BLA-lesioned animals on our task that incorporates both predictable fluctuations and directional shifts in outcome values.

Results Rats can detect true changes in values despite variability in outcomes Our delay-based task was designed to assess animals’ ability to detect true changes in outcome val- ues (i.e., upshifts and downshifts) even when they occur against the background of stochastic feed- back under baseline conditions (expected uncertainty). To probe the effects of expected outcome uncertainty on learning in rodents, we first presented a group of naı¨ve rats (n = 8) with two choice options identical in average wait time but different in the variance of the outcome distribution. Each response option was associated with the delivery of one sugar pellet after a delay interval. The delays were pooled from distributions that were identical in mean, but different in variability (low vs

high: LV vs HV; ~N(m, s): m = 10 s, sHV=4s sLV=1 s). Following the establishment of stable perfor- mance (defined as no statistical difference in any of the behavioral parameters across three consecu- tive testing sessions, including choice and initiation omissions, average response latencies and option preference), rats experienced value upshifts (delay mean was reduced to 5 s with variance kept constant) and downshifts (delay mean was increased to 20 s) on each option independently, fol- lowed by return to baseline conditions (Figure 1A,B; Video 1, Video 2). Each shift and baseline phase lasted five 60-trial testing sessions; therefore, the total duration of the main task was 43

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 3 of 27 Research article Neuroscience

Figure 1. Task design and performance of intact animals. Our task is designed to investigate the effects of expected outcome uncertainty on value learning. (A) Each trial began with stimulus presentation in the central compartment of the touchscreen. Rats (n = 8) were given 40 s to initiate a trial. If 40 s passed without a response, the trial was scored as an ‘initiation omission.’ Following a nosepoke to the central compartment, the central stimulus disappeared and two choice stimuli were presented concurrently in each of the side compartments of the touchscreen allowing an animal a free choice between two reward options. An animal was given 40 s to make a choice; failure to select an option within this time interval resulted in the trial being scored as ‘choice omission’ and beginning of an ITI. Each response option was associated with the delivery of one sugar pellet after a delay interval. (B) The delays associated with each option were pooled from distributions that are identical in mean value, but different in variability: LV (low variability, shown in blue) vs. HV (high variability, shown in red); ~N(m, s): m = 10 s, sHV=4s, sLV=1s. Following the establishment of stable performance, rats experienced value upshifts (m = 5 s; s kept constant) and downshifts (m = 20 s) on each option independently, followed by return to baseline conditions. Each shift and return to baseline phase lasted for five 60-trial sessions. (C) Regardless of the shift type, animals significantly changed their preference in response to all shifts (all p values<0.05). However, significant differences between HV and LV in choice adaptations were observed for both upshifts and downshifts: greater variance of outcome distribution at baseline facilitated behavioral adaptation in response to value upshifts (HV vs LV difference, p=0.004), but rendered animals suboptimal during downshifts (p=0.027); conversely, low expected uncertainty at baseline led to decreased reward procurement during upshifts in reward. The data are shown as group means for option preference during pre-baseline, shift and post-baseline conditions, ± SEM. The asterisks signify statistical differences between HV and LV conditions. (D) The number of initiation omissions was significantly increased during downshift (p=0.004) and decreased during upshifts (p=0.017) in value, regardless of the levels of expected uncertainty, demonstrating effects of overall environmental reward conditions on motivation to engage in the task. The data are shown as group means by condition +SEM. *p<0.05, **p<0.01. Summary statistics and individual animal data are provided in Figure 1—source data 1. DOI: 10.7554/eLife.27483.003 The following source data is available for figure 1: Source data 1. Summary statistics and individual data for naı¨ve animals performing the task. DOI: 10.7554/eLife.27483.004

testing days for each animal. Maximal changes in the choice of each option in response to shifts were analyzed with omnibus within-subject ANOVA with shift type (HV, LV; upshift, downshift) and shift phase (pre-shift baseline, shift, post-shift baseline) as within-subject factors. These analyses identified a significant shift type x phase interaction [F(6, 42)=16.412, p<0.0001]. Post-hoc analyses revealed no differences in preference at baseline conditions across assessments [F(3.08, 21.57)=0.98, p=0.422; Greenhouse-Geisser corrected], suggesting that rats were able to infer mean option values (wait times) and maintain stable choice preferences despite variability in outcomes. All animals significantly changed their preference in response to all shifts (Figure 1, all p val- ues<0.05). We then assessed the effects of the overall environmental reward conditions on rats’

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 4 of 27 Research article Neuroscience motivation to engage in the task. The number of initiation omissions (i.e., failure to respond to the central cue presented at the beginning of each trial within 40 s) was analyzed with omnibus ANOVA with reward conditions (stable, upshift, and downshift collapsed across HV and LV options) as within-subject factor. The main effect of condition was significant [F(1.09, 7.61)=16.772, p=0.03; Greenhouse-Geisser corrected]: the num- ber of omissions was significantly increased dur- ing downshifts (p=0.004) and decreased during Video 1. An animal performing the task during upshift upshifts (p=0.017) in value, revealing that task on HV option. During an upshift in value on each engagement was sensitive to overall environmen- option, the mean of the delays to reward was reduced to 5 s with variance kept the same as during baseline tal reward rate. conditions. Therefore, rodents are able to learn about fun- DOI: 10.7554/eLife.27483.005 damental directional changes in value means despite stochastic fluctuations in outcome values under baseline conditions (i.e., expected uncer- tainty). However, significant differences between HV and LV in choice adaptations were observed for both upshifts and downshifts: greater variance of outcome distribution at baseline facilitated behavioral adaptation in response to value upshifts (HV vs LV difference, p=0.004), but rendered animals suboptimal during downshifts (p=0.027); con- versely, low expected uncertainty at baseline led to decreased reward procurement during upshifts in reward. These effects may be explained by a hyperbolic nature of delay-discounting across spe- cies (Freeman et al., 2009; Green et al., 2013; Hwang et al., 2009; Mazur and Biondi, 2009; Mitchell et al., 2015; Rachlin et al., 1991).

Experience with uncertainty induces distinct patterns of neuroadaptations in the BLA and OFC We hypothesized that experience with different levels of outcome uncertainty would induce long- term neuroadaptations, affecting the response to the same magnitude of surprise signals. Specifi-

cally, we assessed expression of gephyrin (a reliable proxy for membrane-inserted GABAA receptors mediating fast inhibitory transmission; [Chhatwal et al., 2005; Tyagarajan et al., 2011]) and GluN1 (an obligatory subunit of glutamate NMDA receptors; [Soares et al., 2013]) in BLA and OFC. Three separate groups of animals were trained to respond to visual stimuli on a touchscreen to procure a reward after variable delays. The values of outcomes were identical to our task described above but no choice was given. One group was trained under LV conditions, the second under HV (matched in total number of rewards received), and the third control group received no rewards (n = 8 in each group, total n = 24). Given the limited amount of tissue, we focused on NMDA instead of AMPA receptors based on previous evidence demon- strating dissociable effects of ionotropic gluta- mate receptors in delay-based decision making (Yates et al., 2015). Protein expression analyses revealed unique adaptations to outcome variability in BLA, specif- ically in GABA-ergic sensitivity. Biochemical measures were analyzed with mixed ANOVA with brain region as a within-subject factor and reward experience (HV, LV or no reward) as a between-subject factor. There was a significant main effect of group [F(2,12)=6.002, p=0.016] Video 2. An animal performing the task during and brain region x group interaction [Figure 2A; downshift on HV option. During a downshift in value on F(2,12)=41.863, p<0.0001] for gephyrin. A signif- each option, the mean of the delays to reward was icant main effect of group [F(2,21)=4.084, increased to 20 s with variance kept constant. p=0.032] and group x brain region [F(2,21) DOI: 10.7554/eLife.27483.006 =5.291, p=0.014] interaction were also found for

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 5 of 27 Research article Neuroscience

Figure 2. Region-specific alterations in gephyrin and GluN1 expression induced by experience with outcome uncertainty. Three separate groups of animals were trained to respond to visual stimuli on a touchscreen to get a reward after variable delays. The values of outcomes were identical to the main task but no choice was given. One group was trained under LV conditions, the second under HV (matched in total number of rewards received), and the third control group received no rewards (n = 8 per group). We assessed expression of A gephyrin (a reliable proxy for membrane-inserted

GABAA receptors mediating fast inhibitory transmission) and B GluN1 (an obligatory subunit of glutamate NMDA receptors) in BLA and ventral OFC. Biochemical analyses revealed uncertainty-dependent upregulation in gephyrin in BLA, that was maximal following HV training (p<0.0001). Similarly, GluN1 showed robust upregulation in response to experienced reward in BLA (no reward vs LV p=0.045; no reward vs HV p=0.002), however post hoc analyses failed to detect a significant difference between HV and LV training (p=0.637). In ventral OFC, gephyrin was downregulated in response to experiences with reward in general (no reward vs LV p=0.045; no reward vs HV p=0.042) and did not depend on variability in outcome distribution; no changes were observed in GluN1. The data are shown as group means by condition +SEM. *p<0.05, **p<0.01 Summary statistics and individual animal data are provided in Figure 2—source data 1. DOI: 10.7554/eLife.27483.007 The following source data is available for figure 2: Source data 1. Summary statistics and individual data for GluN1 and gephyrin expression in BLA and OFC. DOI: 10.7554/eLife.27483.008

GluN1 expression. Subsequent analyses identified uncertainty-dependent upregulation of gephyrin in BLA [between-subject ANOVA: F(2,21)=45.448, p<0.0001), that was maximal following HV train- ing (all post hoc comparison p values<0.05). Similarly, GluN1 showed robust upregulation in response to experienced reward in BLA [Figure 2B; F(2,21)=7.092, p=0.004; no reward vs LV p=0.045; no reward vs HV p=0.002], however post hoc analyses failed to detect a significant differ- ence between HV and LV training (p=0.637). In OFC, gephyrin was instead downregulated in response to experiences with reward in general [F(2,12)=4.445, p=0.036; no reward vs LV p=0.045; no reward vs HV p=0.042] and did not depend on variability in outcome distribution (post hoc com- parison: HV vs LV, p=1); no changes were observed in GluN1 [F(2,21)=2.359, p=0.119]. Therefore, both the BLA and OFC undergo unique patterns of neuroadaptations in response to experience with variability, suggesting that these brain regions may play complementary, yet disso- ciable, roles in value learning under outcome uncertainty. Given the behavioral complexity that rodents exhibit on our task, we were able to directly test the causal contributions of the BLA and ventromedial OFC to value learning under conditions of expected uncertainty in outcome distribution.

Causal contributions of the BLA and OFC to value learning under uncertainty The results of lesion studies (lesion sites are shown in Figure 3) were in line with predictions sug- gested by protein data. Because we were primarily interested in the contributions of the BLA and OFC to surprise-driven learning, we first analyzed the maximal changes in option preference in response to up- and downshifts. This analysis allowed us to control for potential effects of brain lesions on choice behavior under baseline conditions in our task. An omnibus ANOVA with shift type as within- and experimental group (sham, BLA vs OFC lesion; n = 8 per group; total n = 24) as

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 6 of 27 Research article Neuroscience

Figure 3. Location and extent of intended lesion (colored regions) on standard coronal sections through ventral OFC and BLA. The extent of the lesions was assessed after the completion of behavioral testing by staining for a marker of neuronal nuclei, NeuN. (A) Top: representative photomicrograph of a NeuN stained coronal section showing ventral OFC lesion. Bottom: depictions of coronal sections adapted from (Paxinos and Watson, 1997). The numerals on the lower left of each matched section represent the anterior-posterior distance (mm) from Bregma. Light and dark blue represent maximum and minimum lesion area across animals, respectively. Though coordinates were aimed at the ventral orbital region, lesion extent includes anterior medial orbital cortex as well. (B) Top: representative photomicrograph of a NeuN stained coronal section showing BLA lesion. Bottom: depictions of coronal sections with numerals on the lower left of each matched section representing the anterior-posterior distance (mm) from Bregma. Light and dark red represent maximum and minimum lesion area across animals, respectively. DOI: 10.7554/eLife.27483.009

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 7 of 27 Research article Neuroscience between-subject factors detected a significant main effect of group [F(2,21)=11.193, p<0.0001] and group x shift type interaction [F(6,63)=9.472, p<0.0001). Subsequent analyses showed significant simple main effects of experimental group on all shift types: upshift on HV [F(2,21)=14.723, p<0.0001], upshift on LV [F(2,21)=5.663, p=0.011], downshift on HV [F(2,21)=19.081, p<0.0001], and downshift on LV [F(2,21)=7.189, p=0.004]. The OFC-lesioned rats were less optimal on our task: they changed their option preference to a significantly lesser degree compared to control animals during upshifts on HV (p=0.005) and LV (p=0.039), as well as the downshift on LV option (p=0.015; Figure 4A). Whereas OFC lesions produced a pronounced impairment in performance, it was less clear if alterations produced by BLA lesions lead to suboptimal behavior. BLA-lesioned animals changed their option preference to a lesser degree on HV upshifts (p<0.0001), but compensated by exaggerated adaptations to HV downshifts (p<0.0001; Figure 4A). In addition to examining the maximal changes in option preferences, we analyzed the behavioral data with an omnibus ANOVA with shift type and shift phase (pre-shift baseline, shift performance, and post-shift baseline) as within-subject and experimental group as between-subject factors. This test similarly detected a significant shift type x phase x group interaction [F(6.9,72.5)=7.41, p<0.0001; Greenhouse-Geisser corrected, Figure 4—figure supplement 1). Consistent with the pre- ceding analyses, post hoc tests revealed reduced adaptations to value uphifts on the HV option in both lesion groups (p<0.01). However, we also observed more frequent choices of the LV option when its value was increased in BLA-lesioned animals (p<0.01) as well as reduced HV option prefer- ence (p<0.01) and increased LV option preference (p<0.05) during downshifts in both lesion groups compared to control animals. This pattern of results may be explained by changes in choice behavior even under baseline conditions in BLA- and OFC-lesioned animals that interacted with rats’ ability to learn about shifts in value. Successful performance on our task required animals to distinguish between the variance of out- come distributions under stable conditions from surprising shifts in value, despite the fact that delay distributions at baseline and distributions during the shift partially overlapped. To evaluate whether the animals in lesioned groups adopted a different strategy and demonstrated altered sensitivity to surprising outcomes, we examined the win-stay/lose-shift responses. Win-stay and lose-shift scores were computed based on trial-by-trial data similar to previous reports (Faraut et al., 2016; Imhof et al., 2007; Worthy et al., 2013): a score of 1 was assigned when animals repeated the choice following better than average outcomes (win-stay) or switched to the other alternative follow- ing worse than average outcomes (lose-shift). Win-shift and lose-stay trials were counted as 0 s. To specifically address whether rats distinguished expected fluctuations from surprising changes, we divided the trials into two types: when the delays fell within distributions experienced for each option at baseline (expected outcomes) and those in which the degree of surprise exceeded that expected by chance. The algorithm used for this analysis kept track of all delays experienced under baseline conditions before the current trial for each animal individually. On each trial, we found the value of the minimal and maximal delay. If the current delay value fell within this interval, the out- come was classified as expected. If the current delay fell outside of this distribution, the outcome on this trial was classified as unexpected (surprising). Win-stay and lose-shift scores were calculated for each trial type separately and their probabilities (summary score divided by the number of trials) for both trial types were subjected to ANOVA with strategy as within-subject and experimental group as between-subject factors. Our analyses indi- cated significant strategy x experimental group interaction [F(6,63)=9.912, p<0.0001]. Critically, sham-lesioned animals demonstrated increased sensitivity to unexpected outcomes compared to predictable fluctuations for both wins and losses (Figure 4B, p values <0.0001). Similarly, the ability to distinguish between expected and unexpected outcomes was intact in BLA-lesioned animals (p values < 0.001), although their sensitivity to feedback decreased overall. In contrast, OFC-lesioned animals failed to distinguish predictable from surprising fluctuations. Interestingly, sham and BLA- lesioned animals demonstrated low win-stay and lose-shift scores when trial outcomes were expected; these animals were more likely to shift after better than average outcomes and persist with their choices after worse outcomes. In addition to feedback insensitivity, such behavior may result from increases in exploratory behavior in response to wins and behavioral inflexibility after losses. Additionally, when outcomes are relative stable and predictable, rats may be more sensitive to long-term reward history and rely less on the outcome of any one given trial.

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 8 of 27 Research article Neuroscience

Figure 4. Changes in choice preference in response to value shifts and learning strategies in experimental groups. (A) The OFC-lesioned rats (n = 8) were less optimal on our task: they changed their option preference to a significantly lesser degree compared to control animals (n = 8) during upshifts on HV (p=0.005) and LV (p=0.039), as well as the downshift on LV option (p=0.015). Conversely, animals with BLA lesions (n = 8) changed their option preference to a lesser degree on HV upshifts (p<0.0001), but compensated by exaggerated adaptations to HV downshifts (p<0.0001). Group means for option preference during pre-baseline, shift and post-baseline conditions are shown in Figure 4—figure supplement 1. We broke the trials into two types: when the delays fell within distributions experienced for each option at baseline (expected outcomes) and those in which the degree of surprise exceeded that expected by chance (unexpected outcomes). Win-stay/lose-shift scores were computed based on trial-by-trial data: a score of 1 was assigned when animals repeated the choice following better than average outcomes (win-stay) or switched to the other alternative following worse than average outcomes (lose-shift). Sham-lesioned animals demonstrated increased sensitivity to unexpected feedback (p values < 0.001). Similarly, the ability to distinguish between expected and unexpected outcomes was intact in BLA-lesioned animals (p values < 0.001), although their sensitivity to feedback decreased overall. In contrast, OFC-lesioned animals failed to distinguish expected from unexpected fluctuations. (C,D) To examine the learning trajectory we analyzed the evolution of option preference. BLA-lesioned animals were indistinguishable from controls during the shifts on LV option. Whereas, this experimental group demonstrated significantly attenuated learning during the upshift on HV (p values < 0.0001 for all sessions) and potentiated performance during sessions 3 through 5 on HV downshift (p values < 0.05) compared to sham group. Conversely, learning in OFC- lesioned animals was affected on the majority of the shift types: these animals demonstrated significantly slower learning during sessions 3 through 5 during upshift on HV (p values < 0.05), all sessions during upshift on LV (p values < 0.05) and sessions 3 through 5 during downshift on LV (p values < 0.05). Session 0 refers to baseline/pre-shift option preference. Despite these differences in responses to shifts in value under conditions of uncertainty, we did not observe any deficits in basic reward learning in either the BLA- or OFC-lesioned animals, shown in Figure 4—figure supplement 2. The data are shown as group means by condition +SEM. *p<0.05, **p<0.01. Summary statistics and individual animal data are provided in Figure 4—source data 1 and Figure 4—source data 2. DOI: 10.7554/eLife.27483.010 The following source data and figure supplements are available for figure 4: Source data 1. Summary statistics and individual data for changes in choice preference and learning strategies. DOI: 10.7554/eLife.27483.011 Source data 2. Summary statistics and individual data demonstrating experimental group differences in response to shifts. DOI: 10.7554/eLife.27483.012 Figure supplement 1. Changes in choice behavior in response to value shifts. DOI: 10.7554/eLife.27483.013 Figure supplement 2. The lack of group differences in basic reward learning. DOI: 10.7554/eLife.27483.014

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 9 of 27 Research article Neuroscience Lesions to the BLA and OFC alter learning trajectory To examine the learning trajectory we analyzed the evolution of option preference during shift con- ditions. Specifically, we subjected the session-by-session data during each swift to an omnibus ANOVA with testing session (1 through 5; session 0 in Figure 4C,D corresponds to pre-shift option preference) and shift type as within- and experimental group as between subject factors. This analy- sis revealed a three-way session x shift type x group interaction [F(8.73, 91.71)=8.418, p=0.002; Greenhouse-Geisser corrected, Figure 4C,D]. Subsequent analyses identified significant two-way session x group interactions for each shift type [upshift on HV: F(5.24, 55.04)=3.585, p=0.006; down- shift on HV: F(4.14, 43.452)=25.646, p<0.0001; upshift on LV: F(2.59,27.14) = 4.378, p=0.016; down- shift on LV: F(3.69, 38.767)=6.768, p<0.0001; all Greenhouse-Geisser corrected]. BLA-lesioned animals were indistinguishable from controls during the shifts on LV option. However, this experi- mental group demonstrated significantly attenuated learning during the upshift on HV (p val- ues < 0.0001 for all sessions) and potentiated performance during sessions 3 through five during the downshifts on HV (p values < 0.05) compared to the sham group. Conversely, learning in OFC- lesioned animals was affected on the majority of the shift types: these animals demonstrated signifi- cantly slower learning during sessions 3 through five during upshift on HV (p values < 0.05), all ses- sions during upshift on LV (p values < 0.05) and sessions 3 through five during downshift on LV (p values < 0.05). Despite these differences in responses to shifts in value, we did not observe any deficits in basic reward learning in either the BLA- or OFC-lesioned animals. Our surgeries took place prior to any exposure to the testing apparatus or behavioral training, yet both lesioned groups were indistin- guishable from controls at early stages of the task. All animals took a similar number of days to learn to nosepoke visual stimuli on the touchscreen to receive sugar rewards [F(2,21)=0.231, p=0.796] and to initiate a trial [F(2,21)=0.199, p=0.821]. Similarly, there were no group differences in their responses to the introduction of a 5 s delay interval during pre-training [F(2,21)=0.679, p=0.518] or the number of sessions to reach stable performance during the initial baseline phase of our uncer- tainty task [F92,21)=0.262, p=0.772; Figure 4—figure supplement 2].

Complementary contributions of the BLA and OFC to value learning under uncertainty revealed by computational modeling We fit different versions of RL models to trial-by-trial choices for each animal separately. Specifically, we considered the standard Rescorla-Wagner model (RW) and a dynamic learning rate model (Pearce-Hall, PH). The RW model updates option values in response to RPEs (i.e., the degree of sur- prise) with a constant learning rate, conversely the PH model allows for learning facilitation with sur- prising feedback (i.e., the learning rate is scaled according to absolute prediction errors). We also compared models in which expected outcome uncertainty is learned simultaneously with value and scales the impact of prediction errors on value (RW+expected uncertainty) and learning rate (Full model) updating. The total number of free parameters, BIC and parameter values for each model and experimental group are provided in Table 1. The behavior of the control group was best cap- tured by the dynamic learning rate model with RPE scaling proportional to expected outcome uncer- tainty and facilitation of learning in response to surprising feedback (Full model; Table 1, lower BIC values indicate better fit). Therefore, rats in our experiment increased learning rates in response to surprise to maximize reward acquisition rate, but only if unexpected outcomes were not likely to result from value fluctuations under otherwise stable conditions. Consistent with attenuated learning observed in animals with BLA lesions, trial-by-trial performance in these animals was best fit by RW +expected uncertainty model, demonstrating selective loss of learning potentiation in response to surprise and preserved RPE scaling with expected uncertainty in these animals, leading to slower learning compared to intact animals during the shifts on HV option. Conversely, performance of OFC-lesioned animals was best accounted for by PH model, suggesting that while these animals still increased learning rates in response to surprise, they were insensitive to expected outcome uncer- tainty. Furthermore, the overall learning rates were reduced in OFC-lesioned animals (p=0.01 com- pared to the sham group). Finally, we observed significantly lower values of b (inverse temperature parameter in softmax choice rule) in both BLA- and OFC-lesioned animals [F(2,21)=4.88, p=0.018; sham vs BLA: p<0.0001; sham vs OFC: p<0.0001], suggesting that their behavior is less stable, more

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 10 of 27 Research article Neuroscience

Table 1. Model comparison. Lower BIC values indicate better model fit (in bold); number of free parameters and parameter values ± SEM of the best fitting model are provided for each group. Trial-by-trial choices of the intact animals were best captured by the dynamic learning rate model incorporating RPE scaling proportional to expected uncertainty and facilitation of learning in response to surprising outcomes (Full model). BLA lesions selectively eliminated learning rate scaling in response to surprise (RW+expected uncer- tainty model provided the best fit). Whereas OFC lesioned animals still increased learning rates in response to surprising events (PH model), RPE scaling proportional to expected outcome uncertainty was lost in this group. Furthermore, the overall learning rates were reduced in OFC-lesioned animals (p=0.01). Finally, we observed significantly lower values of b (inverse temperature parameter in soft- max choice rule) in both BLA- and OFC-lesioned animals (p<0.0001), suggesting that their behavior is less stable, more exploratory and less dependent on the difference in learned outcome values. Asterisks indicate parameter values that were significantly different from the control group (in bold).

Model RW PH RW+expected uncertainty Full # parameters 3 4 5 6 BIC parameter value ± SEM k a, value b h a, risk w sham 26519.39 26900.66 26384.18 25681.7 0.29 ± 0.03 0.09 ± 0.01 14.1 ± 0.99 0.33 ± 0.04 0.56 ± 0.08 3.04 ± 0.11 BLA lesion 26201.89 26864.74 25153.82 27162.82 0.32 ± 0.02 0.07 ± 0.01 7.4 ± 0.6* n/a 0.58 ± 0.06 3.40 ± 0.4 OFC lesion 24292.54 23171.46 24630.92 23994.5 0.3 ± 0.05 0.05 ± 0.01* 5.5 ± 0.68* 0/32 ± 0.05 n/a n/a

DOI: 10.7554/eLife.27483.015

exploratory and less dependent on the difference in learned outcome values compared to control group.

Animals with ventral OFC lesions fail to represent expected uncertainty in wait time distributions To gain further insights into outcome representations in our experimental groups, we analyzed the microstructure of rats’ choice behavior. Specifically, we addressed whether BLA and ventral OFC lesions altered animals’ ability to form expectations about the timing of reward delivery. On each trial during all baseline conditions, where the overall values of LV and HV options were equivalent, reward port entries were recorded in 1 s bins during the waiting period (after a rat had indicated its choice and until reward delivery; histograms of true distributions of the delays and animals’ reward- seeking actions normalized to the total number of reward port entries are shown in Figure 5). These data were analyzed with an ANOVA with time bin as within- and lesion group as between-subject factors. There were no significant differences in the mean of expected reward delivery times across groups [F(5,42)=1.064, p=0.394]. Similarly, all groups were matched in the total number of reward port entries [F(2,21)=0.462, p=0.636; Figure 5—figure supplement 1]. However, a significant differ- ence in variances of reward port entry distributions was detected [c2(209)=4004.054, p<0.0001]. Whereas the distributions of reward-seeking times in BLA-lesioned rats were indistinguishable from control animals’ and the true delays, OFC-lesioned animals concentrated their reward port entries in the time interval corresponding to mean delays, suggesting that while these animals can infer the average outcomes, they fail to represent the variance (i.e., expected uncertainty). We also considered the changes in waiting times across our task. We calculated the variance of reward port entry times during each baseline (initial phase of the task and four baseline separating the shifts) for each animal. We then subjected the estimated variances to ANOVAs with baseline order (1st to 5th) as within- and lesion group as between-subject factors. Similar to our preceding analysis of combined baselines, we did not detect any group differences in waiting times for the LV option (all p values>0.2). However, there was a significant main effect of lesion group on waiting time variances for the HV option [F(2,21)=117.074, p<0.0001; Figure 5—figure supplement 1] with OFC-lesioned animals demonstrating consistently lower variability in their waiting behavior despite the experience with shifts. Importantly, since our analyses only included the waiting time prior to reward delivery, these results suggest that OFC-lesioned animals retain the ability to form simple outcome expectations based on long-term experience, yet their ability to represent the more com- plex outcome distributions is compromised.

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 11 of 27 Research article Neuroscience

Figure 5. Animals with ventral OFC lesions fail to represent expected uncertainty in reward delays. We assessed whether BLA and ventral OFC lesions alter animals’ ability to form expectations about the timing of reward delivery. On each trial during all baseline conditions where the overall value of LV and HV options were equivalent, reward port entries were recorded in 1 s bins during the waiting period. There were no significant differences in the means of expected reward delivery times across groups (p=0.394). Similarly, the groups were matched in the total number of reward port entries (p=0.636) as shown in Figure 5—figure supplement 1. Whereas the distributions of reward-seeking times in BLA-lesioned animals were indistinguishable from control animals’ and the true delays (A–F), OFC-lesioned animals concentrated their reward port entries in the time interval corresponding to mean delays (G,H), suggesting that while these animals can infer the average outcome, they fail to represent the variance (i.e., expected uncertainty). We also considered the changes in waiting times across our task; these data are shown in Figure 5—figure supplement 1. Each bar in histogram plots represents mean frequency normalized to total number of reward port entries ±SEM. DOI: 10.7554/eLife.27483.016 The following figure supplement is available for figure 5: Figure supplement 1. Total number of reward port entries and changes in waiting time variances across task phases. DOI: 10.7554/eLife.27483.017

Lesions to the BLA and ventral OFC induce an uncertainty-avoidant phenotype under baseline conditions To assess group differences in uncertainty-seeking or avoidance, we subjected HV option preference data under baseline conditions to an ANOVA with time (five repeated baseline tests separating the value shifts) as within- and lesion group as between-subject factors. In addition to their effects on value learning, lesions to both the BLA and ventral OFC induced an uncertainty-avoidant phenotype with animals in both experimental groups demonstrating reduced preference for the HV option under baseline conditions compared to the control group at the beginning of testing [time x group interaction: F(4.37,45.87) = 8.484, p<0.0001; post hoc sham vs BLA: p=0.002; sham vs OFC:

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 12 of 27 Research article Neuroscience p=0.002, Figure 6]. BLA-lesioned animals continued to avoid the uncertain option for the entire duration of our experiment (all p values < 0.05, except for baseline three assessment when this group was not different from control animals). However, OFC-lesioned animals increased their choices of the HV option during baseline conditions with repeated testing: they were indistinguish- able from controls during baselines 3 and 4 and even demonstrated a trend for higher preference than the control group during the last baseline [post hoc test, OFC vs sham: p=0.059].

Discussion Volatile reward statistics were one of the central characteristics of ancestral habitats, favoring the selection of behavioral phenotypes that are able to cope with uncertainty (Emery, 2006; Potts, 2004; Steppan et al., 2004). Most mammals are able to learn higher-order statistics of the environment (Cikara and Gershman, 2016; Gershman and Niv, 2010; Niv et al., 2015) and opti- mize learning rates based on the degree of uncertainty (Behrens et al., 2007; Nassar et al., 2010; Payzan-LeNestour and Bossaerts, 2011). Until recently, most of the studies have been carried out in the context of probabilistic feedback, where stochasticity in outcomes is driven by reward

Figure 6. BLA and ventral OFC lesions induce uncertainty-avoidance. We observed significantly reduced preference for the HV option under baseline conditions in both experimental groups compared to control animals at the beginning of testing (sham vs BLA: p=0.002; sham vs OFC: p=0.002). BLA-lesioned animals continued to avoid the risky option for most of the experiment (all p values < 0.05, except for baseline three assessment when this group was not different from control animals). OFC-lesioned animals progressively increased their choices of HV option during baseline conditions with repeated testing: they were indistinguishable from controls during baselines 3 and 4 and even demonstrated a trend for higher preference than control group during the last baseline [post hoc test, OFC vs sham: p=0.059]. The data are shown as group means by condition ±SEM, *p<0.05, **p<0.01. Summary statistics and individual animal data are provided in Figure 6—source data 1. DOI: 10.7554/eLife.27483.018 The following source data is available for figure 6: Source data 1. Summary statistics and individual data for HV option preference following lesions. DOI: 10.7554/eLife.27483.019

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 13 of 27 Research article Neuroscience omission in a subset of trials. Unlike in laboratory tasks, uncertainty in naturalistic settings is not lim- ited to probabilistic binary outcomes, but also includes variability in delays and effortful costs required to obtain the desired rewards. In the present work, we developed a delay-based task for rats to investigate the effects of expected outcome uncertainty on value learning. Our results pro- vide the first evidence that rats can detect and learn about the true changes in outcome values even when they occur against a background of stochastic delay costs. In our task, animals successfully changed their choice behavior in response to directional shifts in delay distributions (i.e., value up- and downshifts) to maximize the rate of reward acquisition, while maintaining stable choice preferen- ces despite variability in outcomes under baseline conditions. We note that the changes in option preference in response to shifts on HV and LV options were asymmetric: greater variance of outcome distribution facilitated behavioral adaptations in response to value upshifts; conversely, low expected outcome uncertainty led to potentiated responses to downshifts. This effect may be explained by the hyperbolic nature of delay-discounting across spe- cies (Freeman et al., 2009; Green et al., 2013; Hwang et al., 2009; Mazur and Biondi, 2009; Mitchell et al., 2015; Rachlin et al., 1991). Specifically, the delays in our task were normally distrib- uted, but the perceived value distributions may be skewed. Since the HV option produces a greater proportion of immediate or short-delayed rewards, and therefore more valuable outcomes, it may generally be easier for animals to detect upshifts on this option. These more immediate rewards may be more salient and/or more preferred. Conversely, during the downshifts as delays get longer, differences in waiting time become less meaningful and the LV option that produces more delays of similar value could promote faster learning about worsening of reward conditions. Despite these effects of delays on outcome valuation, our results demonstrated that rats can learn about shifts in values even when outcomes are uncertain. We then directly assessed the uncertainty- induced neural adaptations within the BLA and OFC and investigated causal contributions of these brain regions to value learning and decision making under expected outcome uncertainty.

The BLA and ventral OFC undergo distinct patterns of neuroadaptations in response to outcome uncertainty One of the most difficult challenges faced by an animal learning in an unstable habitat is correctly distinguishing between true changes in the environment that require new learning from stochastic feedback under mostly stable conditions. Indeed, the problem of change-point detection has long been studied in relation to modulation of learning rates in RL and Bayesian learning theory (Behrens et al., 2007; Courville et al., 2006; Dayan et al., 2000; Gallistel et al., 2001; Pearce and Hall, 1980; Pearson and Platt, 2013; Yu and Dayan, 2005). Long-term neuroadaptations in response to experience with outcome uncertainty may benefit learning by altering signal-to-noise processing (Hoshino, 2014; Liguz-Lecznar et al., 2015; Ro¨ssert et al., 2011), such that only those surprising events that exceed the levels of expected variability in the environment produce neuronal responses and affect behavior. We directly assessed the changes in expression of gephyrin (a reliable proxy for membrane-

inserted GABAA receptors mediating fast inhibitory transmission; [Chhatwal et al., 2005; Tyagarajan et al., 2011]) and GluN1 (an obligatory subunit of glutamate NMDA receptors; [Soares et al., 2013]) in BLA and ventral OFC in three separate groups of animals following pro- longed experience with low and high levels of expected uncertainty in outcome distribution. Both gephyrin and GluN1 showed robust uncertainty-dependent upregulation in the BLA that was maxi- mal after experience with highly uncertain conditions. Conversely, within the ventral OFC, gephyrin was downregulated following reward experience in general and did not depend on the degree of uncertainty in outcomes. However, our experiments did not include a certain control group (i.e., ani- mals receiving rewards following a predictable delay on all trials). Therefore, we cannot exclude the possibility that changes in protein expression in the OFC in response to reward experience required some, albeit small, levels of outcome uncertainty. Adaptations to expected uncertainty at the protein level are likely to diminish responses to subse- quent trial-by-trial surprise signals in BLA. Concurrent increases in the sensitivity to excitation and inhibition benefit signal-to-noise processing, providing further evidence in support of this view (Hoshino, 2014; Liguz-Lecznar et al., 2015; Ro¨ssert et al., 2011). To detect environmental changes, animals need to compare current prediction errors to the levels of expected outcome uncertainty. Previous work has shown that GABA-ergic interneurons in BLA gate the information

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 14 of 27 Research article Neuroscience flow and determine the signal intensity that is passed to postsynaptic structures (Wolff et al., 2014). The intrinsic excitability of pyramidal neurons (Motanis et al., 2014; Paton et al., 2006) and activity of interneurons in the BLA are shaped by reward experiences, possibly via a dopamine-dependent

mechanism (Chu et al., 2012; Merlo et al., 2015). Upregulation of functional GABAA receptors as suggested by our data may decrease sensitivity to surprising events when outcome variability is high even under mostly stable conditions, while increases in GluN1 could support learning facilitation when the environment changes. Several psychiatric conditions such as anxiety, schizophrenia, obses- sive compulsive and autism spectrum disorders, share pathological uncertainty processing as a core deficit, manifesting as a preference for stable, certain outcomes (Winstanley and Clark, 2016a, 2016b). Interestingly, recent studies have similarly implicated mutations in the gephyrin gene as risk for autism and schizophrenia (Chen et al., 2014; Lionel et al., 2013). Future research may address the role of this synaptic organizer in surprise-driven learning and decision making under uncertainty in animal models of these disorders. Contrary to the pattern of neuroadaptations observed in BLA, gephyrin in the OFC was downre- gulated in response to reward mean, but not expected uncertainty. These changes in protein expres- sion may leave OFC responsivity to noisy value signals intact or even amplified, suggesting that one of its normal functions is to encode the richness of the outcome distribution or expected uncertainty signal. Indeed, previous reports demonstrated that at least some subpopulations of OFC neurons carry expected uncertainty representations during option evaluation and outcome receipt (Li et al., 2016; van Duuren et al., 2009). Based on these findings we hypothesized that BLA and ventral OFC may play complementary, yet dissociable, roles in decision making and learning under uncertainty.

Ventral OFC causally contributes to learning under expected outcome uncertainty Lesions to the ventral OFC produced a pronounced behavioral impairment on our task. These ani- mals failed to change their choice preference in response to the majority of shifts. Paradoxically, the results of computational modeling revealed that responsivity to surprising outcomes was facilitated in these rats. Specifically, performance of OFC-lesioned animals was best accounted for by the PH model, suggesting that while these animals still increased learning rates in response to surprise (i.e., absolute prediction errors), they were insensitive to expected outcome uncertainty. Due to the lack of prediction error scaling based on the variability in experienced outcomes, OFC-lesioned animals treated every surprising event as indicative of a fundamental change in the value distribution and updated their expectations, rendering trial-by-trial value representations noisier, preventing consis- tent changes in preference. Because the delay distributions encountered during baseline and shift conditions in our task partially overlapped, inability to ignore meaningless fluctuations in outcomes would lead to unstable choice behavior and attenuated learning. Complementary analyses of win-stay/ lose-shift strategy provide further support for potentiated sensitivity to surprising feedback in these animals: increased responsivity to both wins and losses emerged following ventral OFC lesions. Note that increased reliance on this strategy is highly subop- timal under stochastic environmental reward conditions (Faraut et al., 2016; Imhof et al., 2007; Worthy et al., 2013). Furthermore, we observed significantly reduced b (inverse temperature parameter in softmax decision rule) values in OFC-lesioned group, indicating a noisier choice pro- cess and decreased reliance on learned outcome values in these animals. These results are in agree- ment with previous findings demonstrating increased switching and inconsistent economic preferences following ventral OFC lesions in monkeys (Walton et al., 2010, 2011). Similarly, lesions to ventromedial prefrontal cortex, encompassing the ventral OFC, in humans render subjects unable to make consistent preference judgments (Fellows and Farah, 2003, 2007). Importantly, human subjects with OFC damage cannot distinguish between degrees of uncertainty (Hsu et al., 2005). Similarly, previous work has implicated this brain region in prediction of reward timing (Bakhurin et al., 2017). We directly addressed whether BLA and ventral OFC lesions alter animals’ ability to form expectations about the expected uncertainty in timing of reward delivery on our task. Whereas the distributions of reward-seeking times in BLA-lesioned animals were indistinguishable from control animals’ and the true delays, OFC-lesioned animals concentrated their reward port entries in the time interval corresponding to mean delays, suggesting that while these animals can infer the average outcomes, they fail to represent the variance (i.e., expected uncertainty). These

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 15 of 27 Research article Neuroscience findings are consistent with emerging evidence that more ventromedial regions, unlike lateral, OFC may be critical for decision making involving outcome uncertainty, but not response inhibition or impulsive choice behavior as suggested previously (Stopper et al., 2014). Although frequently framed as a deficit in inhibitory control (Bari and Robbins, 2013; Dalley et al., 2004; Elliott and Deakin, 2005), medial OFC lesions or inactivations induce analogous effects in probabilistic reversal learning tasks where surprising changes in the reward distribution occur against the background of stochastic outcomes during the baseline conditions. For example, a recent study in rodents systematically compared the contributions of five different regions of the frontal cortex to reversal learning (Dalton et al., 2016). The results revealed unique contributions of the OFC to successful performance under probabilistic, but not deterministic conditions. Intriguingly, inactivations of the medial OFC impaired both the acquisition and reversal phases, suggesting that this subregion might be critical for many types of reward learning under conditions of expected out- come uncertainty. Since our lesions also intruded on medial OFC, our present observations are in agreement with these findings and suggest that one of the normal functions of more ventromedial sectors of OFC might be to stabilize value representations by adjusting responses to surprising out- comes based on expected outcome uncertainty. Similar to previous work demonstrating that the OFC is not required for acquisition of simple stimulus-outcome associations or for unblocking driven by differences in value when outcomes are certain and predictable (Izquierdo et al., 2004; McDannald et al., 2011, 2005; Stalnaker et al., 2015), we observed intact performance in OFC-lesioned animals during training to respond for rewards. It has been previously proposed that the OFC may provide value expectations that can be used to calculate RPEs to drive learning under more complex task conditions (Schoenbaum et al., 2011a, Schoenbaum et al., 2011b). Although this initial proposal was based on findings after tar- geting more lateral OFC subregions, our observations are generally consistent with this view and add a nuanced perspective. Specifically, if the OFC is needed to provide expectations about the value to which the observed outcomes are then compared, lesions of this brain region may result in attenuated learning driven by violation of expectations. The results of computational modeling in our work revealed a reduction in learning rates in OFC-lesioned animals consistent with this account. Yet our data provide further evidence that the representations of expected outcomes in ventral OFC are not limited to a single-point estimate of value, but also include information about expected uncertainty of variability in outcomes. This would allow an animal not only to detect if the outcomes violate expectations, but also to assess whether such surprising events are meaningful and informa- tive to the current state of the world. If such events are important, an animal will shift its behavior, but if they may have occurred by chance, choices should remain unchanged. Finally, more recently it has also been suggested that the OFC represents an animal’s current location within an abstract cognitive map of the task it is facing (Chan et al., 2016; Schuck et al., 2016; Wilson et al., 2014), particularly when task states are not signaled by external sensory infor- mation, but rather need to be inferred from experience. In our task, animals may similarly represent different conditions, stable environment vs. shifted value, as separate states. Attenuated learning may result from state misrepresentations, where an animal incorrectly infers that it is currently in a stable environment and persists with the previous choice policy, despite the shift in value. As has been reported recently, neuronal activity in the lateral OFC organizes the task space according to the sequence of behaviorally significant events, or trial epochs. Conversely, neural ensembles in more medial OFC do not track the sequence of the events, but instead segregate between states depending on the trial value (Lopatina et al., 2017). In our study, ventromedial OFC may be espe- cially well-positioned to encode upshifts and downshifts in value on long timescales, and loss of this function could cause an inability to recover appropriate state representations at the time of option choice. Taken together with previous findings, our results implicate the OFC in representing fine-grained value distributions, including the expected uncertainty in outcomes (that may be task state-depen- dent). Consequently, lacking access to the complex outcome distribution, animals with OFC lesions over-rely on the average cached value.

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 16 of 27 Research article Neuroscience Functionally intact BLA is required for facilitation of learning in response to surprise Whereas OFC lesions produced a pronounced impairment in performance on our uncertainty task, whether alterations induced by BLA lesions lead to suboptimal behavior is less clear. These animals changed their option preference to a lesser degree on HV upshifts, but compensated by exagger- ated adaptations to HV downshifts. More detailed analyses of session-by-session data revealed spe- cific alteration in responses to surprising value shifts under HV, but not LV, conditions in this group. Consistent with attenuated learning observed in animals with BLA lesions, trial-by-trial performance in this group was best fit by a RW+expected uncertainty model, demonstrating a selective loss of learning rate scaling in response to surprise and preserved RPE scaling with expected outcome uncertainty, leading to slower value learning compared to intact animals during the HV upshift. Note that suboptimal performance during even two or three sessions in our task (each session lasting 60 trials) means that BLA-lesioned animals are less efficient at reward procurement for 120–180 experi- ences. In naturalistic settings, such an early-learning impairment can have detrimental consequences. In agreement with the results of computational modeling, BLA-lesioned animals were less likely to adopt the win-stay/lose-shift strategy compared to the control group, demonstrating decreased sen- sitivity to surprising outcomes. Whereas the lack of learning facilitation can account for reduced changes in preference in response to HV upshifts in BLA-lesioned animals, it may seem at odds with potentiated responses to downshifts on this option. Our computational modeling results suggest that control animals potenti- ate their learning in response to highly surprising outcomes, which leads to greater behavioral adap- tations in the first few sessions during the shifts. In BLA-lesioned animals, this function is lost, and learning proceeds at the same rate. This results in significantly reduced choice adaptations through- out the HV upshift sessions. Yet BLA-lesioned animals adapt much more to the downshift on HV option. This difference appears to be in the performance asymptote, as learning still progresses line- arly in BLA-lesioned group. A couple of factors may drive this effect. Firstly, as discussed earlier hyperbolic discounting leads to a larger impact of short delays on behavior. Immediate or short- delayed rewards encountered during upshift on HV option will potentiate learning in control animals early on during the shift, but fail to do so in BLA lesioned animals. During the downshift on HV option, as delays get longer, differences in waiting times become less meaningful as there is a smaller effect of larger delays on perceived outcome values. Thus, learning will be potentiated, but only briefly in control animals, but will still proceed linearly in BLA-lesioned rats. Additionally, poten- tiated responses to downshifts on HV option in this group may result from uncertainty avoidance interacting with surprise-driven learning. Indeed, we observed a consistent increase in uncertainty aversion in BLA-lesioned animals. Our computational models did not include an explicit uncertainty avoidance parameter as we were primarily interested in exploring alterations in learning. Previous findings have implicated the BLA in updating reward expectancies when the predictions and outcomes are incongruent and facilitating learning in response to surprising events (Ramirez and Savage, 2007; Savage et al., 2007; Wassum and Izquierdo, 2015). Indeed, predic- tive value learning in the amygdala involves a neuronal signature that accords with an RL algorithm (Dolan, 2007). Specifically, single-unit responses in the BLA correspond to the unsigned prediction error signals (Roesch et al., 2010) that are necessary for learning rate scaling in both RL and Bayes- ian updating models. The BLA utilizes positive and negative prediction errors to boost cue process- ing, potentially directing attention to relevant stimuli and potentiating learning (Chang et al., 2012; Esber and Holland, 2014) as demonstrated in downshift procedures with reductions in reward amount. These effects are frequently interpreted as surprise-induced enhancement of cue associabil- ity. Notably, a similar computational role for the amygdala has been proposed based on Pavlovian fear conditioning in humans, where cue-shock associations were also probabilistic, highlighting the general role for the amygdala in fine-tuning learning according to the degree of surprise (Li et al., 2011). Taken together, the accumulated literature suggests that this contribution of the BLA is apparent for both appetitive and aversive outcomes, for cues in different sensory modalities, and as we demonstrate here, the role is not limited to changes in outcome contingencies, but also supports learning about surprising changes in delay costs.

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 17 of 27 Research article Neuroscience BLA and OFC lesions induce uncertainty-avoidance In addition to their effects on value learning, lesions to both the BLA and ventral OFC induced an uncertainty-avoidant phenotype with animals in both experimental groups demonstrating reduced preference for the HV option under baseline conditions compared to control group at the beginning of testing. Similarly, previous findings demonstrated that lesions or inactivations of the BLA shift the behavior away from uncertain options and promote choices of safer outcomes (Ghods-Sharifi et al., 2009; Zeeb and Winstanley, 2011). However, inactivations of the medial OFC have been shown to produce consistent shifts toward the uncertain option (Winstanley and Floresco, 2016b). Despite demonstrating pronounced risk-aversion at the beginning of the task, OFC-lesioned animals in our experiments progressively increased their preference for HV option with experience, suggesting that the effects on stable choice preference depend critically on the timing of OFC manipulations. In summary, we show that both BLA and ventral OFC are causally involved in decision making and value learning under conditions of outcome uncertainty. Functionally-intact BLA is required for facilitation of learning in response to surprise, whereas ventral OFC is necessary for an accurate representation of outcome distributions to stabilize value expectations and maintain choice preferences.

Materials and methods Subjects were 56 naı¨ve male outbred Long Evans rats (Charles River Laboratories, Crl:LE, Strain code: 006). All animals arrived at our facility at PND 70 (weight range 300–350 at arrival). Vivaria were maintained under a reversed 12/12 hr light/dark cycle at 22˚C. Rats were left undisturbed for 3 days after arrival to our facility to acclimate to the vivarium. Each rat was then handled for a mini- mum of 10 min once per day for 5 days. Animals were food-restricted to ensure motivation to work for food for a week prior to and during the behavioral testing, while water was available ad libitum, except during behavioral testing. All animals were pair-housed at arrival and separated on the last day of handling to minimize aggression during food restriction. We ensured that animals did not fall below 85% of their free-feeding body weight. On the two last days of food restriction prior to behavioral training, rats were fed 20 sugar pellets in their home cage to accustom them to the food rewards. All behavioral procedures took place 5 days a week between 8am and 6pm during the rats’ active period. Because we utilized a novel decision making task, we did not use an a priori power analysis to determine sample size for initial cohort of naı¨ve animals. The chosen group size (n = 8) is consistent with previous reports in our lab. For subsequent behavioral experiments with sham, OFC, or BLA lesions we determined the animal numbers using an a priori sample size estimation for F test family in G*Power 3.1 (http://www.gpower.hhu.de/en.html). The analyses were based on the vari- ance parameters obtained in the pilot experiments (reported in Figure 1 and associated Figure 1— source data 1) and the number of independent variables as well as the interactions of interest in planned analyses. The analysis yielded a projected minimum of 7–8 animals per group when no sur- gical procedures are required. However, considering the possibility of surgical attrition, we set n = 8 per group. Research protocols were approved by the Chancellor’s Animal Research Committee at the University of California, Los Angeles.

Behavioral training Behavioral training was conducted in operant conditioning chambers (Model 80604, Lafayette Instru- ment Co., Lafayette, IN) that were housed within sound- and light- attenuating cubicles. Each cham- ber was equipped with a house light, tone generator, video camera, and LCD touchscreen opposing the pellet dispenser. The pellet dispenser delivered 45 mg dustless precision sucrose pellets. Soft- ware (ABET II TOUCH; Lafayette Instrument Co., Model 89505) controlled the hardware. All testing schedules were programmed by our group and can be requested from the corresponding author. During habituation, rats were required to eat five sugar pellets out of the dispenser inside of the chambers within 15 min before exposure to any stimuli on the touchscreen. They were then trained to respond to visual stimuli presented in the central compartment of the screen within 40 s time interval in order to receive the sugar reward. During the next stage of training, animals learned to initiate the trial by nose-poking the bright white square stimulus presented in the central compart- ment of the touchscreen within 40 s; this response was followed by disappearance of the central

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 18 of 27 Research article Neuroscience stimulus and presentation of a target image in one of the side compartments of the touchscreen (immediately to the left or right of the initiation stimulus). Rats were given 40 s to respond to the tar- get image, which was followed by an immediate reward. The last stage of training was administered to familiarize animals with delayed outcomes. The protocol was identical to the previous stage, except the nosepoke to the target image and reward delivery were separated by a 5 s stable delay. Across all stages of pre-training, failure to respond to a visual stimulus within the allotted time resulted in the trial being scored as an omission and beginning of a 10 s ITI. All images used in pre- training were pulled from the library of over 100 visual stimuli and were never the same as the images used in behavioral testing described below. This was done to ensure that none of the visual cues acquired incentive value that could affect subsequent performance. Criterion for advancement into the next stage was set to 60 rewards collected within 45 min.

Behavioral testing Task design and behavior of intact animals are illustrated in Figure 1, Video 1 and Video 2. Our task is designed to assess the effects of expected outcome uncertainty on learning. We have elected to focus on reward rate (outcome value was determined by delay to reward receipt) rather than reward magnitude to avoid the issue of satiety throughout the testing session. Each trial began with stimulus (bright white square) presentation in the central compartment of the touchscreen. Rats were given 40 s to initiate a trial. If 40 s passed without a response, the trial was scored as an ‘initia- tion omission’. Following a nosepoke to the central compartment, the central cue disappeared and two choice stimuli were presented concurrently in each of the side compartments of the touchscreen allowing an animal a free choice between two reward options. In our task stimulus-response side assignments were held constant for each animal to facilitate learning. Side-stimulus assignments were counterbalanced across animals, and held constant between sessions. Each response option was associated with the delivery of one sugar pellet after a delay interval. The delays associated with each option were pooled from distributions that are identical in mean value, but different in variabil-

ity (LV vs HV; ~N(m, s): m = 10 s, sHV=4s sLV=1s). An animal was given 40 s to make a choice; failure to select an option within this time interval resulted in the trial being scored as ‘choice omission’ and beginning of an ITI. Therefore, rats were presented with two options identical in mean (10 s) but different in the vari- ance of the delay distribution (i.e., expected outcome uncertainty). Following the establishment of stable performance (defined as no statistical difference in any of the behavioral parameters across three consecutive testing sessions), rats experienced reward upshifts (delay mean was reduced to 5 s with variance kept constant) and downshifts (20 s) on each option independently, followed by return to baseline conditions. Thus, in upshifts rats were required to wait less on average for a single sugar pellet, whereas in downshifts, rats were required to wait longer, on average. The order of shift expe- riences was counterbalanced across animals. Animals were given one testing session per day that was terminated when an animal had collected 60 rewards or when 45 min had elapsed. Each shift and return to baseline phase lasted for five sessions. Therefore, rats experienced a total number of 43 sessions with varying delays. We first trained a group of naı¨ve rats (n = 8) on this task to probe the ability to distinguish true changes in the environment from stochastic fluctuations in outcomes under baseline conditions in rodents. The animals in lesion experiments (n = 24: n sham = 8, n BLA lesion = 8; n OFC lesion = 8) were tested under identical conditions. Each animal participated in a single experiment. For each experiment, rats were randomly assigned into groups.

Protein expression analyses Three separate groups of animals were trained to respond to visual stimuli on a touchscreen to pro- cure a reward after variable delays. The values of outcomes were identical to our task described above but no choice was given. One group was trained under LV conditions, the second under HV (matched in total number of rewards received), and the third control group received no rewards (n = 8 in each group; total n = 24). The training criterion was set to a 60 sugar pellets for three con- secutive days to mimic the baseline testing duration in animal trained on our main task. Rats were euthanized 1d after the last day of reward experience with an overdose of sodium pentobarbital (250 mg/kg, i.p.) and decapitated. The brains were immediately extracted and two mm-thick coronal sections of ventral OFC and BLA were further rapidly dissected, using a brain matrix, over wet ice at

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 19 of 27 Research article Neuroscience 4˚C. To prepare the tissues for the assays 0.2 mL of PBS (0.01 mol/L, pH 7.2) containing a protease and phosphatase inhibitor cocktail (aprotinin, bestatin, E-64; leupeptin, NaF, sodium orthovanadate, sodium pyrophosphate, b-glycerophosphate; EDTA-free; Thermo Scientific, Rockford, IL; Product # 78441) was added to each sample. Each tissue was minced, homogenized, sonicated with an ultra- sonic cell disrupter, and centrifuged at 5000 g at 4˚C for 10 min. Supernatants were removed and stored at +4˚C until ELISA assays were performed (within 24 hr period). Bradford protein assays were also performed to determine total protein concentrations in each sample. The assays were per- formed according to the manufacturer’s instructions. The sensitivity of the assays is 0.1 ng/ml for gephyrin (Cat# MBS9324933) and GluN1 (Cat# MBS724735, MyBioSource, Inc, San Diego, CA) and the detection range is 0.625 ng/ml - 20 ng/ml. The concentration of each protein was quantified as ng/mg of total protein accounting for dilution factor and presented as percent of no reward group.

Surgery Excitotoxic lesions of BLA (n = 8) and ventral OFC (n = 8) were performed using aseptic stereotaxic techniques under isoflurane gas (1–5% in O2) anesthesia prior to behavioral testing and training. Before surgeries, all animals were administered 5 mg/kg s.c. carprofen (NADA #141–199, Pfizer, Inc., Drug Labeler Code: 000069) and 1cc saline. After being placed into a stereotaxic apparatus (David Kopf; model 306041), the scalp was incised and retracted. The skull was then leveled to ensure that bregma and lambda are in the same horizontal plane. Small burr holes were drilled in the skull to allow cannulae with an injection needle to be lowered into the BLA (AP: 2.5; ML: ± 5.0; À DV: 7.8 (0.1 ml) and 8.1 (0.2 ml) from skull surface) or OFC (0.2 ml, AP =+3.7; ML = ±2.0; DV = À À 4.6). The injection needle was attached to polyethylene tubing connected to a Hamilton syringe À mounted on a syringe pump. N-Methyl-D-aspartic acid (NMDA, Sigma-Aldrich; 20 mg/ml in 0.1 m PBS, pH 7.4; Product # M3262) was bilaterally infused at a rate of 0.1 ml/min to destroy intrinsic neu- rons. After each injection, the needle was left in place for 3–5 min to allow for diffusion of the drug. Sham-lesioned group (n = 8) underwent identical surgical procedures, except no NMDA was infused. All animals were given one-week recovery period prior to food restriction and subsequent behavioral testing. During this week, the rats were administered 5 mg/kg s.c. carprofen (NADA #141–199, Pfizer, Inc., Drug Labeler Code: 000069) and their health condition was monitored daily.

Histology The extent of the lesions was assessed by staining for NeuN, a marker for neuronal nuclei. After the termination of training, animals were sacrificed by pentobarbital overdose (Euthasol, 0.8 mL, 390 mg/mL pentobarbital, 50 mg/mL phenytoin; Virbic, Fort Worth, TX) and transcardial perfusion. Brains were post-fixed in 10% buffered formalin acetate for 24 hr followed by 30% sucrose for 5 days. Forty mm coronal sections containing the OFC and BLA were first incubated for 24 hr at 4˚C in solution containing primary NeuN antibody (Anti-NeuN (rabbit), 1:1000, EMD. Millipore, Cat. # ABN78), 10% normal goat serum (Abcam, Cambridge, MA, Cat. # ab7481), and 0.5% Triton-X (Sigma, St. Louis, MO, Cat. # T8787) in 1X PBS, followed by three 10 min washes in PBS. The tissue was then incubated for 4 hr in solution containing 1X PBS, Triton-X and a secondary antibody (Goat anti-Rabbit IgG (H+L), Alexa Fluor 488 conjugate, 1:400, Fisher Scientific, Catalog #A-11034), fol- lowed by three 10 min washes in PBS. Slides were subsequently mounted and cover-slipped, visual- ized using a BZ-X710 microscope (Keyence, Itasca, IL), and analyzed with BZ-X Viewer software. Lesions were determined by comparison with a standard rat brain atlas (Paxinos and Watson, 1997).

Computational analyses We fit different versions of reinforcement learning models to trial-by-trial choices for each animal separately. Specifically, we considered the standard Rescorla-Wagner model (RW) and a dynamic learning rate model (Pearce-Hall, PH). Trials from all sessions were treated as contiguous. Option val-

ues were updated in response to RPE, dt, weighted by the learning rate, a (constrained to the inter- val [0 1]). The expected value for each option was updated according to delta rule:

Qt 1 Qt a*dt: þ þ

The dt is the difference between current outcome Vt and expected value Qt. Given that the value

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 20 of 27 Research article Neuroscience

of each outcome was determined by delay to reward of a constant magnitude, Vt was specified as 1/ (1-kD), where D is the duration of delay and k [0, +¥] is a free parameter setting the steepness of the discounting curve. In dynamic learning rate models (PH and PH+expected uncertainty described

below), a was updated in response to the degree of surprise (absolute dt) according to:

at 1 dt *h 1 h *at: þ j j þ ð À Þ We set initial a for HV and LV options to the same value, but allowed independent updating with experience. We also considered models in which expected outcome uncertainty is learned simulta- neously with value and scales the impact of prediction errors on value (RW+expected uncertainty) and learning rate (Full model) updating. Uncertainty prediction errors are the difference between squared expected and realized RPEs. Expected uncertainty expectations are subsequently updated according to delta rule. Therefore, in the Full model: ffiffiffiffi Qt 1 Qt at *dt=!*exp pst0 ; þ þ ð Þ where w [1, +¥] is a free parameter determining individual sensitivity to expected uncertainty. ffiffiffiffi at 1 h* dt =!*exp pst0 1 h *at: þ j j ð Þ þ ð À2 Þ at 10 st0 arisk*drisk;t;drisk;t dt dt0 þ þ ¼ À The option choice probability on each trial was determined according to a softmax rule with an inverse-temperature parameter b; exp(b*Q ). / t The model parameters were estimated to maximize probability of obtaining the observed vector of choices given the model and its parameters (by minimizing negative log likelihood computed based on the difference between predicted choice probability and the actual choice on each trial using fmincon in MatLab). We used Bayesian information criterion (BIC) instead of AIC as a more conservative measure to determine the best model. The total number of free parameters, BIC and parameter values for each model and experimental group are provided in Table 1.

Behavioral and statistical analyses Software package SPSS (SAS Institute, Inc., Version 24) and MatLab (MathWorks, Natick, Massachu- setts; Version R2016b) were used for statistical analyses. Statistical significance was noted when p-values were less than 0.05. Shapiro Wilk tests of normality, Levene’s tests of equality of error var- iances, Box’s tests of equality of covariance matrices, and Mauchly’s tests of sphericity were used to characterize the data structure. Protein expression data were analyzed with univariate ANOVA with reward experience group (HV, LV, or no reward) as the between-subject factor. Maximal changes in choice of each option in response to shifts were analyzed with omnibus ANOVA with shift type (HV, LV; upshift, downshift) and shift phase (pre-baseline, shift, post-baseline) as within-subject factors (total number of animals, n, in this analysis = 8). Similar analyses were performed on data obtained from lesion experiments with an additional between-subject factor of experimental group (sham, BLA vs OFC lesion; total n = 24, n = 8 per group). Furthermore, we subjected the session-by-session data during each swift to an omnibus ANOVA with testing session (1 through 5) and shift type as within- and experimental group as between subject factors.

Win-stay/Lose-shift To evaluate whether the animals in lesioned groups adopted a different strategy and demonstrated altered sensitivity to surprising outcomes, we examined the win-stay/lose-shift response strategy. Win-stay/lose-shift score was computed based on trial-by-trial data similar to previous reports (Faraut et al., 2016; Imhof et al., 2007; Worthy et al., 2013). The algorithm used for this analysis kept track of all delays experienced before the current trial under baseline conditions for each animal individually. On each trial, we calculated the mean of the experienced baseline delay distribution and found the value of the minimal and maximal delay. If the current delay value fell within this inter- val (i.e., min prior delay current delay max prior delay), the outcome was classified as expected.   If the current delay fell outside of this distribution (current delay min prior delay or current delay   max prior delay), the outcome on this trial was classified as unexpected (surprising). Trials on which

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 21 of 27 Research article Neuroscience the current delay exceeded the mean of experienced delay distribution were counted as wins and delays lower than the mean were classified as losses. We counted rats’ decisions as stays when they chose the same option on the subsequent trial and as shifts when the animals switched to the other alternative. Therefore, each trial could be classified as win-stay, win-shift, lose-stay, or lose-shift. Win-stay and lose-shift trials were given a score of 1 and win-shifts and lose-stays were counted as 0 s. We considered all baseline and value-shift trials; however, trials with delays equal to the mean of previously experienced distribution or trials followed by choice omissions were excluded from this analysis. Win-stay and lose-shift scores we calculated for each trial type separately and their proba- bilities (summary score divided by the number of trials) for both trial types were subjected to ANOVA with strategy as within-subject and experimental group as between-subject factors.

Reward port entries To gain further insights into outcome representations in our experimental groups, we addressed whether BLA and ventral OFC lesions altered animals’ ability to form expectations about the timing of reward delivery. On each trial during all baseline conditions, where the overall values of LV and HV options were equivalent, reward port entries were recorded during the waiting period. This anal- ysis included all trials under initial baseline conditions and baselines separating the shifts. Since reward delivery in our task was signaled to animals by illumination of the magazine and sounds made by the dispenser and pellet drop, rats generally collected rewards immediately (median reac- tion time from reward delivery to consumption = 0.84 s). Because our aim was to assess outcome expectations, rather than reactions to reward delivery, we only analyzed the time interval starting at disappearance of visual stimuli following the choice and terminating at the end of the delay period (magazine entries after the pellet has been dispensed were excluded from this analysis). The waiting period was split into 1 s bins and all magazine entries were recorded in each interval. We then divided the number of entries in each bin by the total number of entries to obtain probabilities. These data we analyzed with multivariate ANOVA with option (LV, HV) and time bin as within- and experimental group as between-subject factors. Mauchly’s tests of sphericity were used to compare variances across groups. When significant interactions were found, post hoc simple main effects were reported. Dunnett t (2-sided) comparisons were applied when assessing the differences between experimental and a sin- gle control groups, whereas Bonferroni correction was applied to multiple comparisons. Where the assumptions of sphericity were violated, Greenhouse-Geisser p-value corrections were applied (Epsi- lon <0.75). Group mean values and associated SEM are reported in figures (individual data are pro- vided in Source_Data files).

Acknowledgements This work was supported by UCLA’s Division of Life Sciences Recruitment and Retention fund and UCLA Opportunity Fund (Izquierdo). We are grateful to Jose Gonzalez for his assistance with data collection, as well as Evan Hart and members of Lau lab for their valuable input.

Additional information

Funding Funder Author UCLA Division of Life Sciences Alicia Izquierdo Recruitment and Retention Fund UCLA Opportunity Fund Alicia Izquierdo

The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 22 of 27 Research article Neuroscience Author contributions AS, Conceptualization, Data curation, Formal analysis, Investigation, Visualization, Methodology, Writing—original draft, Writing—review and editing; AI, Conceptualization, Data curation, Formal analysis, Supervision, Funding acquisition, Visualization, Methodology, Writing—original draft, Writ- ing—review and editing

Author ORCIDs Alexandra Stolyarova, http://orcid.org/0000-0003-4397-4895 Alicia Izquierdo, http://orcid.org/0000-0001-9897-2091

Ethics Animal experimentation: This study was performed in strict accordance with the recommendations in the Guide for the Care and Use of Laboratory Animals of the National Institutes of Health. Research protocols (#2013-094-13A) were approved by the Chancellor’s Animal Research Commit- tee at the University of California, Los Angeles. All surgeries were performed under isoflurane anes- thesia, and every effort was made to minimize suffering.

References Bakhurin KI, Goudar V, Shobe JL, Claar LD, Buonomano DV, Masmanidis SC. 2017. Differential encoding of Time by Prefrontal and Striatal Network Dynamics. The Journal of Neuroscience 37:854–870. doi: 10.1523/ JNEUROSCI.1789-16.2016, PMID: 28123021 Bari A, Robbins TW. 2013. Inhibition and impulsivity: behavioral and neural basis of response control. Progress in Neurobiology 108:44–79. doi: 10.1016/j.pneurobio.2013.06.005, PMID: 23856628 Behrens TE, Woolrich MW, Walton ME, Rushworth MF. 2007. Learning the value of information in an uncertain world. Nature Neuroscience 10:1214–1221. doi: 10.1038/nn1954, PMID: 17676057 Chan SC, Niv Y, Norman KA. 2016. A Probability distribution over latent causes, in the Orbitofrontal Cortex. The Journal of Neuroscience 36:7817–7828. doi: 10.1523/JNEUROSCI.0659-16.2016, PMID: 27466328 Chang SE, McDannald MA, Wheeler DS, Holland PC. 2012. The effects of basolateral amygdala lesions on unblocking. Behavioral Neuroscience 126:279–289. doi: 10.1037/a0027576, PMID: 22448857 Chen J, Yu S, Fu Y, Li X. 2014. Synaptic proteins and receptors defects in autism spectrum disorders. Frontiers in Cellular Neuroscience 8:276. doi: 10.3389/fncel.2014.00276, PMID: 25309321 Chhatwal JP, Myers KM, Ressler KJ, Davis M. 2005. Regulation of gephyrin and GABAA receptor binding within the amygdala after fear acquisition and extinction. Journal of Neuroscience 25:502–506. doi: 10.1523/ JNEUROSCI.3301-04.2005, PMID: 15647495 Chu HY, Ito W, Li J, Morozov A. 2012. Target-specific suppression of GABA release from parvalbumin interneurons in the basolateral amygdala by dopamine. Journal of Neuroscience 32:14815–14820. doi: 10. 1523/JNEUROSCI.2997-12.2012, PMID: 23077066 Cikara M, Gershman SJ. 2016. Medial prefrontal cortex updates its Status. Neuron 92:937–939. doi: 10.1016/j. neuron.2016.11.040, PMID: 27930908 Conen KE, Padoa-Schioppa C. 2015. Neuronal variability in orbitofrontal cortex during economic decisions. Journal of Neurophysiology 114:1367–1381. doi: 10.1152/jn.00231.2015, PMID: 26084903 Courville AC, Daw ND, Touretzky DS. 2006. Bayesian theories of conditioning in a changing world. Trends in Cognitive Sciences 10:294–300. doi: 10.1016/j.tics.2006.05.004, PMID: 16793323 Dalley JW, Cardinal RN, Robbins TW. 2004. Prefrontal executive and cognitive functions in rodents: neural and neurochemical substrates. Neuroscience & Biobehavioral Reviews 28:771–784. doi: 10.1016/j.neubiorev.2004. 09.006, PMID: 15555683 Dalton GL, Wang NY, Phillips AG, Floresco SB. 2016. Multifaceted contributions by different regions of the Orbitofrontal and medial prefrontal cortex to Probabilistic reversal Learning. Journal of Neuroscience 36:1996– 2006. doi: 10.1523/JNEUROSCI.3366-15.2016, PMID: 26865622 Dayan P, Kakade S, Montague PR. 2000. Learning and selective attention. Nature Neuroscience 3 Suppl:1218– 1223. doi: 10.1038/81504, PMID: 11127841 Diederen KM, Schultz W. 2015. Scaling prediction errors to reward variability benefits error-driven learning in humans. Journal of Neurophysiology 114:1628–1640. doi: 10.1152/jn.00483.2015, PMID: 26180123 Dolan RJ. 2007. The human amygdala and orbital prefrontal cortex in behavioural regulation. Philosophical Transactions of the Royal Society B: Biological Sciences 362:787–799. doi: 10.1098/rstb.2007.2088, PMID: 17403643 Elliott R, Deakin B. 2005. Role of the orbitofrontal cortex in reinforcement processing and inhibitory control: evidence from functional magnetic resonance imaging studies in healthy human subjects. International Review of Neurobiology 65:89–116. doi: 10.1016/S0074-7742(04)65004-5, PMID: 16140054 Elliott R, Dolan RJ, Frith CD. 2000. Dissociable functions in the medial and lateral orbitofrontal cortex: evidence from human neuroimaging studies. Cerebral Cortex 10:308–317. doi: 10.1093/cercor/10.3.308, PMID: 10731225

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 23 of 27 Research article Neuroscience

Emery NJ. 2006. Cognitive ornithology: the evolution of avian intelligence. Philosophical Transactions of the Royal Society B: Biological Sciences 361:23–43. doi: 10.1098/rstb.2005.1736, PMID: 16553307 Esber GR, Holland PC. 2014. The basolateral amygdala is necessary for negative prediction errors to enhance cue salience, but not to produce conditioned inhibition. European Journal of Neuroscience 40:3328–3337. doi: 10.1111/ejn.12695, PMID: 25135841 Faraut MC, Procyk E, Wilson CR. 2016. Learning to learn about uncertain feedback. Learning & Memory 23:90– 98. doi: 10.1101/lm.039768.115, PMID: 26787780 Fellows LK, Farah MJ. 2003. Ventromedial frontal cortex mediates affective shifting in humans: evidence from a reversal learning paradigm. Brain 126:1830–1837. doi: 10.1093/brain/awg180, PMID: 12821528 Fellows LK, Farah MJ. 2007. The role of ventromedial prefrontal cortex in decision making: judgment under uncertainty or judgment per se? Cerebral Cortex 17:2669–2674. doi: 10.1093/cercor/bhl176, PMID: 17259643 Freeman KB, Green L, Myerson J, Woolverton WL. 2009. Delay discounting of saccharin in rhesus monkeys. Behavioural Processes 82:214–218. doi: 10.1016/j.beproc.2009.06.002, PMID: 19540317 Gallistel CR, Mark TA, King AP, Latham PE. 2001. The rat approximates an ideal detector of changes in rates of reward: implications for the law of effect. Journal of Experimental Psychology: Animal Behavior Processes 27: 354–372. doi: 10.1037/0097-7403.27.4.354, PMID: 11676086 Gershman SJ, Niv Y. 2010. Learning latent structure: carving nature at its joints. Current Opinion in Neurobiology 20:251–256. doi: 10.1016/j.conb.2010.02.008, PMID: 20227271 Ghods-Sharifi S, St Onge JR, Floresco SB. 2009. Fundamental contribution by the basolateral amygdala to different forms of decision making. Journal of Neuroscience 29:5251–5259. doi: 10.1523/JNEUROSCI.0315-09. 2009, PMID: 19386921 Green L, Myerson J, Oliveira L, Chang SE. 2013. Delay discounting of monetary rewards over a wide range of amounts. Journal of the Experimental Analysis of Behavior 100:269–281. doi: 10.1002/jeab.45, PMID: 24037 826 Hart EE, Izquierdo A. 2017. Basolateral amygdala supports the maintenance of value and effortful choice of a preferred option. European Journal of Neuroscience 45:388–397. doi: 10.1111/ejn.13497, PMID: 27977047 Haruno M, Kimura M, Frith CD. 2014. Activity in the and amygdala underlies individual differences in prosocial and individualistic economic choices. Journal of Cognitive Neuroscience 26:1861–1870. doi: 10.1162/jocn_a_00589, PMID: 24564471 Hoshino O. 2014. Balanced crossmodal excitation and inhibition essential for maximizing multisensory gain. Neural Computation 26:1362–1385. doi: 10.1162/NECO_a_00606, PMID: 24708367 Hsu M, Bhatt M, Adolphs R, Tranel D, Camerer CF. 2005. Neural systems responding to degrees of uncertainty in human decision-making. Science 310:1680–1683. doi: 10.1126/science.1115327, PMID: 16339445 Hwang J, Kim S, Lee D. 2009. Temporal discounting and inter-temporal choice in rhesus monkeys. Frontiers in Behavioral Neuroscience 3:9. doi: 10.3389/neuro.08.009.2009, PMID: 19562091 Imhof LA, Fudenberg D, Nowak MA. 2007. Tit-for-tat or win-stay, lose-shift? Journal of Theoretical Biology 247: 574–580. doi: 10.1016/j.jtbi.2007.03.027, PMID: 17481667 Izquierdo A, Brigman JL, Radke AK, Rudebeck PH, Holmes A. 2017. The neural basis of reversal learning: An updated perspective. Neuroscience 345:12–26. doi: 10.1016/j.neuroscience.2016.03.021, PMID: 26979052 Izquierdo A, Murray EA. 2010. Functional interaction of medial mediodorsal thalamic nucleus but not nucleus accumbens with amygdala and orbital prefrontal cortex is essential for adaptive response selection after reinforcer devaluation. Journal of Neuroscience 30:661–669. doi: 10.1523/JNEUROSCI.3795-09.2010, PMID: 20071531 Izquierdo A, Suda RK, Murray EA. 2004. Bilateral orbital prefrontal cortex lesions in rhesus monkeys disrupt choices guided by both reward value and reward contingency. Journal of Neuroscience 24:7540–7548. doi: 10. 1523/JNEUROSCI.1921-04.2004, PMID: 15329401 Khamassi M, Lalle´e S, Enel P, Procyk E, Dominey PF. 2011. Robot cognitive control with a neurophysiologically inspired reinforcement learning model. Frontiers in Neurorobotics 5:1. doi: 10.3389/fnbot.2011.00001, PMID: 21808619 Lee D, Seo H, Jung MW. 2012. Neural basis of reinforcement learning and decision making. Annual Review of Neuroscience 35:287–308. doi: 10.1146/annurev-neuro-062111-150512, PMID: 22462543 Li J, Schiller D, Schoenbaum G, Phelps EA, Daw ND. 2011. Differential roles of human striatum and amygdala in associative learning. Nature Neuroscience 14:1250–1252. doi: 10.1038/nn.2904, PMID: 21909088 Li Y, Vanni-Mercier G, Isnard J, Mauguie`re F, Dreher JC. 2016. The neural dynamics of reward value and risk coding in the human orbitofrontal cortex. Brain 139:1295–1309. doi: 10.1093/brain/awv409, PMID: 26811252 Liguz-Lecznar M, Lehner M, Kaliszewska A, Zakrzewska R, Sobolewska A, Kossut M. 2015. Altered glutamate/ GABA equilibrium in aged mice cortex influences cortical plasticity. Brain Structure and Function 220:1681– 1693. doi: 10.1007/s00429-014-0752-6, PMID: 24659256 Lionel AC, Vaags AK, Sato D, Gazzellone MJ, Mitchell EB, Chen HY, Costain G, Walker S, Egger G, Thiruvahindrapuram B, Merico D, Prasad A, Anagnostou E, Fombonne E, Zwaigenbaum L, Roberts W, Szatmari P, Fernandez BA, Georgieva L, Brzustowicz LM, et al. 2013. Rare exonic deletions implicate the synaptic organizer gephyrin (GPHN) in risk for autism, schizophrenia and seizures. Human Molecular Genetics 22:2055– 2066. doi: 10.1093/hmg/ddt056, PMID: 23393157 Lopatina N, Sadacca BF, McDannald MA, Styer CV, Peterson JF, Cheer JF, Schoenbaum G. 2017. Ensembles in medial and lateral orbitofrontal cortex construct cognitive maps emphasizing different features of the behavioral landscape. Behavioral Neuroscience 131:201–212. doi: 10.1037/bne0000195, PMID: 28541078

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 24 of 27 Research article Neuroscience

Marquardt K, Sigdel R, Brigman JL. 2017. Touch-screen visual reversal learning is mediated by value encoding and signal propagation in the orbitofrontal cortex. Neurobiology of Learning and Memory 139:179–188. doi: 10.1016/j.nlm.2017.01.006, PMID: 28111339 Mazur JE, Biondi DR. 2009. Delay-amount tradeoffs in choices by pigeons and rats: hyperbolic versus exponential discounting. Journal of the Experimental Analysis of Behavior 91:197–211. doi: 10.1901/jeab.2009. 91-197, PMID: 19794834 McDannald MA, Lucantonio F, Burke KA, Niv Y, Schoenbaum G. 2011. Ventral striatum and orbitofrontal cortex are both required for model-based, but not model-free, reinforcement learning. Journal of Neuroscience 31: 2700–2705. doi: 10.1523/JNEUROSCI.5499-10.2011, PMID: 21325538 McDannald MA, Saddoris MP, Gallagher M, Holland PC. 2005. Lesions of orbitofrontal cortex impair rats’ differential outcome expectancy learning but not conditioned stimulus-potentiated feeding. Journal of Neuroscience 25:4626–4632. doi: 10.1523/JNEUROSCI.5301-04.2005, PMID: 15872110 Merlo E, Ratano P, Ilioi EC, Robbins MA, Everitt BJ, Milton AL. 2015. Amygdala dopamine receptors are required for the destabilization of a reconsolidating appetitive memory(1,2). eNeuro 2:ENEURO.0024-14.2015. doi: 10.1523/ENEURO.0024-14.2015, PMID: 26464966 Mitchell SH, Wilson VB, Karalunas SL. 2015. Comparing hyperbolic, delay-amount sensitivity and present-bias models of delay discounting. Behavioural Processes 114:52–62. doi: 10.1016/j.beproc.2015.03.006, PMID: 257 96454 Morris LS, Kundu P, Dowell N, Mechelmans DJ, Favre P, Irvine MA, Robbins TW, Daw N, Bullmore ET, Harrison NA, Voon V. 2016. Fronto-striatal organization: defining functional and microstructural substrates of behavioural flexibility. Cortex 74:118–133. doi: 10.1016/j.cortex.2015.11.004, PMID: 26673945 Motanis H, Maroun M, Barkai E. 2014. Learning-induced bidirectional plasticity of intrinsic neuronal excitability reflects the valence of the outcome. Cerebral Cortex 24:1075–1087. doi: 10.1093/cercor/bhs394, PMID: 23236201 Murray EA, Izquierdo A. 2007. Orbitofrontal cortex and amygdala contributions to affect and action in primates. Annals of the New York Academy of Sciences 1121:273–296. doi: 10.1196/annals.1401.021, PMID: 17846154 Nassar MR, Wilson RC, Heasly B, Gold JI. 2010. An approximately bayesian delta-rule model explains the dynamics of belief updating in a changing environment. Journal of Neuroscience 30:12366–12378. doi: 10. 1523/JNEUROSCI.0822-10.2010, PMID: 20844132 Niv Y, Daniel R, Geana A, Gershman SJ, Leong YC, Radulescu A, Wilson RC. 2015. Reinforcement learning in multidimensional environments relies on attention mechanisms. Journal of Neuroscience 35:8145–8157. doi: 10.1523/JNEUROSCI.2978-14.2015, PMID: 26019331 Ostrander S, Cazares VA, Kim C, Cheung S, Gonzalez I, Izquierdo A. 2011. Orbitofrontal cortex and basolateral amygdala lesions result in suboptimal and dissociable reward choices on cue-guided effort in rats. Behavioral Neuroscience 125:350–359. doi: 10.1037/a0023574, PMID: 21639604 Padoa-Schioppa C, Schoenbaum G. 2015. Dialogue on economic choice, learning theory, and neuronal representations. Current Opinion in Behavioral Sciences 5:16–23. doi: 10.1016/j.cobeha.2015.06.004, PMID: 26613099 Padoa-Schioppa C. 2007. Orbitofrontal cortex and the computation of economic value. Annals of the New York Academy of Sciences 1121:232–253. doi: 10.1196/annals.1401.011, PMID: 17698992 Padoa-Schioppa C. 2009. Range-adapting representation of economic value in the orbitofrontal cortex. Journal of Neuroscience 29:14004–14014. doi: 10.1523/JNEUROSCI.3751-09.2009, PMID: 19890010 Paton JJ, Belova MA, Morrison SE, Salzman CD. 2006. The primate amygdala represents the positive and negative value of visual stimuli during learning. Nature 439:865–870. doi: 10.1038/nature04490, PMID: 164 82160 Paxinos G, Watson C. 1997. The Rat Brainin Stereotaxic Coordinates. Cambridge: Academic Press. Payzan-LeNestour E, Bossaerts P. 2011. Risk, unexpected uncertainty, and estimation uncertainty: bayesian learning in unstable settings. PLoS Computational Biology 7:e1001048. doi: 10.1371/journal.pcbi.1001048, PMID: 21283774 Pearce JM, Hall G. 1980. A model for pavlovian learning: variations in the effectiveness of conditioned but not of unconditioned stimuli. Psychological Review 87:532–552. doi: 10.1037/0033-295X.87.6.532, PMID: 7443916 Pearson JM, Platt ML. 2013. Change detection, multiple controllers, and dynamic environments: insights from the brain. Journal of the Experimental Analysis of Behavior 99:74–84. doi: 10.1002/jeab.5, PMID: 23344989 Potts R. 2004. Paleoenvironmental basis of cognitive evolution in great apes. American Journal of Primatology 62:209–228. doi: 10.1002/ajp.20016, PMID: 15027093 Preuschoff K, Bossaerts P. 2007. Adding prediction risk to the theory of reward learning. Annals of the New York Academy of Sciences 1104:135–146. doi: 10.1196/annals.1390.005, PMID: 17344526 Rachlin H, Raineri A, Cross D. 1991. Subjective probability and delay. Journal of the Experimental Analysis of Behavior 55:233–244. doi: 10.1901/jeab.1991.55-233, PMID: 2037827 Ramirez DR, Savage LM. 2007. Differential involvement of the basolateral amygdala, orbitofrontal cortex, and nucleus accumbens core in the acquisition and use of reward expectancies. Behavioral Neuroscience 121:896– 906. doi: 10.1037/0735-7044.121.5.896, PMID: 17907822 Riceberg JS, Shapiro ML. 2012. Reward stability determines the contribution of orbitofrontal cortex to adaptive behavior. Journal of Neuroscience 32:16402–16409. doi: 10.1523/JNEUROSCI.0776-12.2012, PMID: 23152622 Riceberg JS, Shapiro ML. 2017. Orbitofrontal Cortex signals expected outcomes with predictive codes when stable contingencies promote the integration of reward history. The Journal of Neuroscience 37:2010–2021. doi: 10.1523/JNEUROSCI.2951-16.2016, PMID: 28115481

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 25 of 27 Research article Neuroscience

Roesch MR, Calu DJ, Esber GR, Schoenbaum G. 2010. Neural correlates of variations in event processing during learning in basolateral amygdala. Journal of Neuroscience 30:2464–2471. doi: 10.1523/JNEUROSCI.5781-09. 2010, PMID: 20164330 Ro¨ ssert C, Moore LE, Straka H, Glasauer S. 2011. Cellular and network contributions to vestibular signal processing: impact of ion conductances, synaptic inhibition, and noise. Journal of Neuroscience 31:8359–8372. doi: 10.1523/JNEUROSCI.6161-10.2011, PMID: 21653841 Rudebeck PH, Murray EA. 2014. The orbitofrontal oracle: cortical mechanisms for the prediction and evaluation of specific behavioral outcomes. Neuron 84:1143–1156. doi: 10.1016/j.neuron.2014.10.049, PMID: 25521376 Salinas JA, Parent MB, McGaugh JL. 1996. Ibotenic acid lesions of the amygdala basolateral complex or central nucleus differentially effect the response to reductions in reward. Brain Research 742:283–293. doi: 10.1016/ S0006-8993(96)01030-X, PMID: 9117406 Salzman CD, Paton JJ, Belova MA, Morrison SE. 2007. Flexible neural representations of value in the primate brain. Annals of the New York Academy of Sciences 1121:336–354. doi: 10.1196/annals.1401.034, PMID: 17 872400 Savage LM, Koch AD, Ramirez DR. 2007. Basolateral amygdala inactivation by muscimol, but not ERK/MAPK inhibition, impairs the use of reward expectancies during working memory. European Journal of Neuroscience 26:3645–3651. doi: 10.1111/j.1460-9568.2007.05959.x, PMID: 18052977 Schoenbaum G, Roesch MR, Stalnaker TA, Takahashi YK. 2011a. Orbitofrontal Cortex and Outcome Expectancies: Optimizing Behavior and Sensory Perception. In: Gottfried J. A (Ed). Neurobiology of Sensation and Reward. Boca Raton: CRC Press/Taylor & Francis. Schoenbaum G, Takahashi Y, Liu TL, McDannald MA. 2011b. Does the orbitofrontal cortex signal value? Annals of the New York Academy of Sciences 1239:87–99. doi: 10.1111/j.1749-6632.2011.06210.x, PMID: 22145878 Schuck NW, Cai MB, Wilson RC, Niv Y. 2016. Human Orbitofrontal Cortex represents a cognitive map of State Space. Neuron 91:1402–1412. doi: 10.1016/j.neuron.2016.08.019, PMID: 27657452 Soares C, Lee KF, Nassrallah W, Be´ı¨que JC. 2013. Differential subcellular targeting of glutamate receptor subtypes during homeostatic synaptic plasticity. Journal of Neuroscience 33:13547–13559. doi: 10.1523/ JNEUROSCI.1873-13.2013, PMID: 23946413 Stalnaker TA, Cooch NK, Schoenbaum G. 2015. What the orbitofrontal cortex does not do. Nature Neuroscience 18:620–627. doi: 10.1038/nn.3982, PMID: 25919962 Steppan SJ, Storz BL, Hoffmann RS. 2004. Nuclear DNA phylogeny of the squirrels (Mammalia: rodentia) and the evolution of arboreality from c-myc and RAG1. Molecular Phylogenetics and Evolution 30:703–719. doi: 10. 1016/S1055-7903(03)00204-5, PMID: 15012949 Stopper CM, Green EB, Floresco SB. 2014. Selective involvement by the medial orbitofrontal cortex in biasing risky, but not impulsive, choice. Cerebral Cortex 24:154–162. doi: 10.1093/cercor/bhs297, PMID: 23042736 Sugrue LP, Corrado GS, Newsome WT. 2005. Choosing the greater of two goods: neural currencies for valuation and decision making. Nature Reviews Neuroscience 6:363–375. doi: 10.1038/nrn1666, PMID: 15832198 Tyagarajan SK, Ghosh H, Ye´venes GE, Nikonenko I, Ebeling C, Schwerdel C, Sidler C, Zeilhofer HU, Gerrits B, Muller D, Fritschy JM. 2011. Regulation of GABAergic synapse formation and plasticity by GSK3beta- dependent phosphorylation of gephyrin. PNAS 108:379–384. doi: 10.1073/pnas.1011824108, PMID: 21173228 van Duuren E, van der Plasse G, Lankelma J, Joosten RN, Feenstra MG, Pennartz CM. 2009. Single-cell and population coding of expected reward probability in the orbitofrontal cortex of the rat. Journal of Neuroscience 29:8965–8976. doi: 10.1523/JNEUROSCI.0005-09.2009, PMID: 19605634 Walton ME, Behrens TE, Buckley MJ, Rudebeck PH, Rushworth MF. 2010. Separable learning systems in the macaque brain and the role of orbitofrontal cortex in contingent learning. Neuron 65:927–939. doi: 10.1016/j. neuron.2010.02.027, PMID: 20346766 Walton ME, Behrens TE, Noonan MP, Rushworth MF. 2011. Giving credit where credit is due: orbitofrontal cortex and valuation in an uncertain world. Annals of the New York Academy of Sciences 1239:14–24. doi: 10. 1111/j.1749-6632.2011.06257.x, PMID: 22145871 Wassum KM, Izquierdo A. 2015. The basolateral amygdala in reward learning and . Neuroscience & Biobehavioral Reviews 57:271–283. doi: 10.1016/j.neubiorev.2015.08.017, PMID: 26341938 Wilson RC, Takahashi YK, Schoenbaum G, Niv Y. 2014. Orbitofrontal cortex as a cognitive map of task space. Neuron 81:267–279. doi: 10.1016/j.neuron.2013.11.005, PMID: 24462094 Winstanley CA, Clark L. 2016a. Translational models of Gambling-Related Decision-Making. Current Topics in Behavioral Neurosciences 28:93–120. doi: 10.1007/7854_2015_5014, PMID: 27418069 Winstanley CA, Floresco SB. 2016b. Deciphering decision making: variation in Animal models of effort- and Uncertainty-Based choice reveals distinct neural circuitries underlying Core Cognitive Processes. Journal of Neuroscience 36:12069–12079. doi: 10.1523/JNEUROSCI.1713-16.2016, PMID: 27903717 Winstanley CA. 2007. The orbitofrontal cortex, impulsivity, and addiction: probing orbitofrontal dysfunction at the neural, neurochemical, and molecular level. Annals of the New York Academy of Sciences 1121:639–655. doi: 10.1196/annals.1401.024, PMID: 17846162 Wolff SB, Gru¨ ndemann J, Tovote P, Krabbe S, Jacobson GA, Mu¨ ller C, Herry C, Ehrlich I, Friedrich RW, Letzkus JJ, Lu¨ thi A. 2014. Amygdala interneuron subtypes control fear learning through disinhibition. Nature 509:453– 458. doi: 10.1038/nature13258, PMID: 24814341 Worthy DA, Hawthorne MJ, Otto AR. 2013. Heterogeneity of strategy use in the Iowa gambling task: a comparison of win-stay/lose-shift and reinforcement learning models. Psychonomic Bulletin & Review 20:364– 371. doi: 10.3758/s13423-012-0324-9, PMID: 23065763

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 26 of 27 Research article Neuroscience

Yates JR, Batten SR, Bardo MT, Beckmann JS. 2015. Role of ionotropic glutamate receptors in delay and probability discounting in the rat. Psychopharmacology 232:1187–1196. doi: 10.1007/s00213-014-3747-3, PMID: 25270726 Yu AJ, Dayan P. 2005. Uncertainty, neuromodulation, and attention. Neuron 46:681–692. doi: 10.1016/j.neuron. 2005.04.026, PMID: 15944135 Zeeb FD, Winstanley CA. 2011. Lesions of the basolateral amygdala and orbitofrontal cortex differentially affect acquisition and performance of a rodent gambling task. Journal of Neuroscience 31:2197–2204. doi: 10.1523/ JNEUROSCI.5597-10.2011, PMID: 21307256

Stolyarova and Izquierdo. eLife 2017;6:e27483. DOI: 10.7554/eLife.27483 27 of 27