Skip to main content

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Blog
    • Collections
    • Podcast
  • TOPICS
    • Cognition and Behavior
    • Development
    • Disorders of the Nervous System
    • History, Teaching and Public Awareness
    • Integrative Systems
    • Neuronal Excitability
    • Novel Tools and Methods
    • Sensory and Motor Systems
  • ALERTS
  • FOR AUTHORS
  • ABOUT
    • Overview
    • Editorial Board
    • For the Media
    • Privacy Policy
    • Contact Us
    • Feedback
  • SUBMIT

User menu

Search

  • Advanced search
eNeuro

eNeuro

Advanced Search

 

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Blog
    • Collections
    • Podcast
  • TOPICS
    • Cognition and Behavior
    • Development
    • Disorders of the Nervous System
    • History, Teaching and Public Awareness
    • Integrative Systems
    • Neuronal Excitability
    • Novel Tools and Methods
    • Sensory and Motor Systems
  • ALERTS
  • FOR AUTHORS
  • ABOUT
    • Overview
    • Editorial Board
    • For the Media
    • Privacy Policy
    • Contact Us
    • Feedback
  • SUBMIT
PreviousNext
Research ArticleNew Research, Cognition and Behavior

Valence-Dependent Coupling of Prefrontal-Amygdala Effective Connectivity during Facial Affect Processing

David Willinger, Iliana I. Karipidis, Selina Beltrani, Sarah V. Di Pietro, Ronald Sladky, Susanne Walitza, Philipp Stämpfli and Silvia Brem
eNeuro 9 July 2019, 6 (4) ENEURO.0079-19.2019; DOI: https://doi.org/10.1523/ENEURO.0079-19.2019
David Willinger
1Department of Child and Adolescent Psychiatry and Psychotherapy, University Hospital of Psychiatry Zurich, University of Zurich, Zurich 8032, Switzerland
3Neuroscience Center Zurich, University of Zurich and ETH Zurich, Zurich 8057, Switzerland
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for David Willinger
Iliana I. Karipidis
1Department of Child and Adolescent Psychiatry and Psychotherapy, University Hospital of Psychiatry Zurich, University of Zurich, Zurich 8032, Switzerland
3Neuroscience Center Zurich, University of Zurich and ETH Zurich, Zurich 8057, Switzerland
4Center for Interdisciplinary Brain Sciences Research, Stanford University School of Medicine, Stanford, CA 94305
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Selina Beltrani
1Department of Child and Adolescent Psychiatry and Psychotherapy, University Hospital of Psychiatry Zurich, University of Zurich, Zurich 8032, Switzerland
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Sarah V. Di Pietro
1Department of Child and Adolescent Psychiatry and Psychotherapy, University Hospital of Psychiatry Zurich, University of Zurich, Zurich 8032, Switzerland
3Neuroscience Center Zurich, University of Zurich and ETH Zurich, Zurich 8057, Switzerland
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Ronald Sladky
2Department of Psychiatry, Psychotherapy and Psychosomatics, University Hospital of Psychiatry Zurich, University of Zurich, Zurich 8008, Switzerland
5Social, Cognitive and Affective Neuroscience Unit, Department of Basic Psychological Research and Research Methods, Faculty of Psychology, University of Vienna, Vienna 1010, Austria
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Ronald Sladky
Susanne Walitza
1Department of Child and Adolescent Psychiatry and Psychotherapy, University Hospital of Psychiatry Zurich, University of Zurich, Zurich 8032, Switzerland
3Neuroscience Center Zurich, University of Zurich and ETH Zurich, Zurich 8057, Switzerland
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Susanne Walitza
Philipp Stämpfli
1Department of Child and Adolescent Psychiatry and Psychotherapy, University Hospital of Psychiatry Zurich, University of Zurich, Zurich 8032, Switzerland
2Department of Psychiatry, Psychotherapy and Psychosomatics, University Hospital of Psychiatry Zurich, University of Zurich, Zurich 8008, Switzerland
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Philipp Stämpfli
Silvia Brem
1Department of Child and Adolescent Psychiatry and Psychotherapy, University Hospital of Psychiatry Zurich, University of Zurich, Zurich 8032, Switzerland
3Neuroscience Center Zurich, University of Zurich and ETH Zurich, Zurich 8057, Switzerland
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Silvia Brem
  • Article
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF
Loading

Abstract

Despite the importance of the prefrontal-amygdala (AMY) network for emotion processing, valence-dependent coupling within this network remains elusive. In this study, we assessed the effect of emotional valence on brain activity and effective connectivity. We tested which functional pathways within the prefrontal-AMY network are specifically engaged during the processing of emotional valence. Thirty-three healthy adults were examined with functional magnetic resonance imaging while performing a dynamic faces and dynamic shapes matching task. The valence of the facial expressions varied systematically between positive, negative, and neutral across the task. Functional contrasts determined core areas of the emotion processing circuitry, comprising the medial prefrontal cortex (MPFC), the right lateral prefrontal cortex (LPFC), the AMY, and the right fusiform face area (FFA). Dynamic causal modelling demonstrated that the bidirectional coupling within the prefrontal-AMY circuitry is modulated by emotional valence. Additionally, Bayesian model averaging showed significant bottom-up connectivity from the AMY to the MPFC during negative and neutral, but not positive, valence. Thus, our study provides strong evidence for alterations of bottom-up coupling within the prefrontal-AMY network as a function of emotional valence. Thereby our results not only advance the understanding of the human prefrontal-AMY circuitry in varying valence context, but, moreover, provide a model to examine mechanisms of valence-sensitive emotional dysregulation in neuropsychiatric disorders.

  • amygdala
  • DCM
  • emotional valence
  • fMRI
  • prefrontal cortex

Significance Statement

Recent neuroimaging studies have emphasized the importance of valence-sensitivity within the prefrontal-amygdala (AMY) network during emotion processing. Yet, it remains elusive which specific pathways are involved in processing affective information, and how this information is integrated in the brain’s network. In particular, the AMY’s role in signaling valence information to the cortex is subject to ongoing discussions. Moreover, as aberrant brain function has been found in the AMY and the prefrontal cortex in various debilitating psychiatric disorders, understanding the mechanisms of processing emotional stimuli with different valence (positive, negative, neutral) is particularly relevant for the field. Our findings indicate changes in coupling strength as a function of emotional valence within the prefrontal-AMY network.

Introduction

The prefrontal-amygdala (AMY) network plays a pivotal role in adapting human behavior to constantly changing environmental demands. Previous neuroimaging research has emphasized the importance of interactions between the prefrontal cortex and the AMY during affective processing (Phillips et al., 2008; Ochsner et al., 2009) and has tried to disentangle bottom-up from top-down mechanisms of emotion processes (Ochsner et al., 2009; Whalen et al., 2013; Comte et al., 2016; Pessoa, 2017). Emotional salience related to the perceptual properties of a stimulus, as mediated by emotional faces, is thought to be propagated from the AMY to the prefrontal cortex via bottom-up connections (McRae et al., 2012). It has long been recognized that the AMY plays a crucial role in immediate, automatic processing of emotional information and the modulation of attention (Anderson et al., 2003; Phelps, 2006; Ochsner et al., 2009). Conversely, top-down signaling during emotion processing has been attributed to different forms of emotion regulation, where the lateral prefrontal cortex (LPFC) supports top-down evaluation of contextual significance and altering of the affective response by exerting cognitive control over limbic regions (Ochsner and Gross, 2005; Quirk and Beer, 2006; Dima et al., 2011), even without explicit instruction (Drabant et al., 2009).

This coupling between the LPFC and the AMY is central to theoretical models of emotion processing. Nevertheless, emotion processing involves complex interactions between AMY driven bottom-up salience processing, and top-down contextualization and evaluation of stimuli, supported by the LPFC, whose strength and directions can differ substantially depending on context, e.g., emotional valence or task demands (Kim et al., 2004; Pessoa, 2017). Lately, this has led to new conceptions, where emotion processing is strongly interwoven with other mental entities that constitute cognition (e.g., memory or attention), and relies on dynamic, context-sensitive interactions of top-down and bottom-up processes (Pessoa, 2017).

Given that structural connections between the LPFC and the AMY are sparse (Ray and Zald, 2012), regulatory signals from the LPFC are likely mediated to the AMY via the medial prefrontal cortex (MPFC). The MPFC is situated perfectly to pass on top-down appraisal and regulation signals to limbic structures as it shares rich bidirectional connections with the LPFC and the AMY (Price, 2005; Ray and Zald, 2012). As such, the MPFC has not only been implicated in regulation of emotional responses, in particular to aversive stimuli, but also in integrating affective and contextual information, i.e., bottom-up and top-down signals, to support generation of affective meaning (Delgado et al., 2008; Ochsner et al., 2009; Roy et al., 2012; Etkin et al., 2015; Silvers et al., 2015; Comte et al., 2016; Lindquist et al., 2016). However, the valence-dependent coupling between regions comprising the emotion processing circuitry is only poorly understood. Particularly, the role of the AMY in encoding valence is still debated (Ball et al., 2009; Jin et al., 2015), and thus far, valence-dependent alterations of directed coupling between the AMY and the prefrontal cortex during emotion processing has not been investigated, despite it being strongly implicated in psychopathology (Dichter et al., 2009; Disner et al., 2011; Sladky et al., 2015b).

In this study, we used a novel dynamic face-matching and shape-matching task to investigate the effect of valence of facial expressions on effective connectivity within the prefrontal-AMY circuitry in 33 healthy adults. Dynamic faces have a higher ecological validity than traditionally used static faces and have been shown to elicit strong responses in brain networks of interest in several fMRI paradigms (Arsalidou et al., 2011; Kessler et al., 2011; Furl et al., 2015). Negatively, neutrally, and positively valenced facial expressions were used to examine the effect of valence on the prefrontal-AMY network, shapes served as a control condition.

In agreement with previous findings using static face processing tasks (Gläscher et al., 2004; Vuilleumier et al., 2004) or affective pictures (Urry et al., 2006), we expected an increased activation for negative valence in bilateral AMYs and the MPFC compared to the neutral and positive valence conditions of our dynamic paradigm. Moreover, dynamic causal modeling (DCM) was used to clarify the contextual influence of valence on the functional architecture of the emotion-processing network. Thus, we investigated whether valence of facial affect modulates effective connectivity within the hierarchical network architecture in a bottom-up, a top-down, or, as recently suggested (Pessoa, 2017), a bidirectional manner. Given the role of the MPFC in integrating context and salience to shape emotional responses (Roy et al., 2012; Etkin et al., 2015), we hypothesized that affective information would modulate bidirectional connections between MPFC and AMY, as well as between MPFC and LPFC.

Materials and Methods

Subjects

A group of 33 healthy volunteers (mean age and SD, 27.4 ± 5.2 years, 24 females and nine males, 30 right handed and three left handed) was recruited for this study. Inclusion criteria were age of 18–45 years and signed informed consent. Exclusion criteria included any MRI contraindication, pregnancy, a history of brain injury, psychiatric disorders, other major medical illnesses, and drug abuse. No subject reported any past or current psychiatric disorder. During scanning, none of the subjects exceeded our motion threshold of a mean framewise displacement (Power et al., 2012) of 0.5 mm (mean 0.14 ± 0.09 mm).

This study was approved by the ethics committee of the Kanton Zurich and was conducted in accordance with the Declaration of Helsinki.

Experimental design

All participants completed a 6-min fMRI dynamic face-matching and shape-matching task (Fig. 1A), which is based on the static task used by Hariri et al. (2002).

Figure 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 1.

A, Experimental design of the study. All participants were presented with triplets of emotional faces (blocks of positive, negative, and neutral valence) and shapes (polygons). The task comprised matching the dynamic target image on top with one of the two static probe images at the bottom by the emotion (faces, with distinct emotional expressions for the static probe images) or number of vertices (shapes). B, Behavioral results of the dynamic face-matching and shape-matching paradigm. RTs were comparable across different valence conditions. Response accuracy in trials with neutral faces was lower than in trials with positive or negative faces.

Face-matching and shape-matching blocks had a length of 20 s and were presented alternatingly. Each of the 12 blocks (six face and six shape blocks) consisted of five trials with a length of 4 s. In each trial, participants were presented with a dynamic target item and two static probe images below it, one of which matched with the target item with regard to shape or facial expression. Subjects were instructed to match either the left or the right static item at the bottom with the dynamic item on top and to press either the left or the right button with their dominant hand on a response-pad to indicate their choice as soon as they recognized which probe matched.

We used face images from the Radboud face database (Langner et al., 2010), including the faces of 38 white actors (19 females). In total, we presented six face-matching blocks (30 trials, 10 trials per valence), two positively valenced, including happy, surprised and neutral faces, two negatively valenced, sad and disgusted faces, and two neutrally valenced, neutral and surprised faces (for a view on surprise as neutrally valenced, see Sander and Scherer, 2014; Soriano Salinas et al., 2015). In the negatively valenced condition, we used sad and disgusted faces instead of widely used fearful faces to capture negative emotion processing not related to arousal (Remmington et al., 2000; Trautmann et al., 2009; Langner et al., 2010). To construct the stimuli for the positive condition, we used the inherently positive valence “happy,” and included faces with neutrally valenced expressions, surprised and neutral, for the face-matching task. This formed overall positively valenced stimulus triplets in all trials, as ambiguous faces (i.e., surprised or neutral) have been shown to be perceived more positive when being presented within a positive context (Neta et al., 2011). Importantly, in each trial of the positive condition, subjects were presented with at least one happy face, either as target or probe. Before the study, we established that a positive condition comprising neutral and positive faces only had lower task difficulty than the other two conditions (neutral and negative blocks). The selection of both, neutral and surprised faces, in the positive condition rendered the task difficulty across conditions comparable. Based on the face scores determined by Langner et al. (2010), the average valence (mean ± SD) of the faces used was 3.55 ± 0.08 for the positive condition, 2.94 ± 0.04 for the neutral condition, and 2.01 ± 0.04 for the negative condition, where 1 represents the most negative and 5 the most positive possible valence rating. Shapes were superimposed to a whirled face in six shape-matching blocks.

We adapted the original task used by Hariri et al. (2002) using a dynamic video sequence of the target emotion or shape to make our task ecologically more valid. Target faces on top were morphed from 0% (neutral faces) to 100% (emotional faces) within the trial time of 4 s. Neutral dynamic target stimuli were morphed to surprised emotion from 0% to 30% intensity and back to 0% intensity to introduce neutrally valenced facial motion. Similarly, during the shape-matching task, the target shape was morphed from a round circle into a polygon with three to eight vertices. During face-matching and shape-matching, probe images on the bottom always remained static. After the button press, the video sequence continued until the end of the morphing sequence of the target face or shape (100% morphed, 4 s). We ensured correct understanding of the task by familiarizing the subjects with the task outside the scanner.

Data acquisition and preprocessing

All MRI recordings were performed on a Philips Achieva 3 Tesla scanner (Philips Medical Systems) using a 32-channel head coil. Functional images were acquired with a multiband echoplanar images (EPIs) sequence (175 volumes, repetition time TR = 2 s, echo time TE = 35 ms, 15° tilted downwards of AC-PC, 54 slices, voxel size = 2.0 × 2.0 × 2.5 mm3, matrix size = 96 × 94 px, flip angle = 80°, no gap, SENSE-factor = 2, MB-factor = 2). Before the actual data acquisition, we acquired five dummy scans to establish steady-state conditions. After performing the task, we acquired a T1-weighted anatomic image for each subject that was used for coregistration and normalization of functional data with a 3D magnetization-prepared rapid gradient-echo sequence (MP-RAGE; time between two inversion pulses = 2484 ms, inversion time TI = 900 ms, inter-echo delay = 6.7 ms, aligned at AC-PC, flip angle = 9°, voxel size = 1.05 × 1.05 × 1.2 mm3, field of view = 270 × 354 mm2, 170 sagittal slices).

Preprocessing of the images included slice-timing correction, realignment, coregistration, and segmentation. Normalization to the Montreal Neurologic Institute (MNI)-152 template space was performed using the deformations derived from the segmentation step. In addition, preprocessing included resampling to 2-mm isometric voxels, and smoothing with a 6-mm full-width-half-maximum (FWHM) Gaussian kernel. All steps were performed using SPM12 (7219) software.

Behavioral analysis

To analyze the behavioral data, we conducted a repeated measures ANOVA to test for effects between valence conditions and paired t tests to test for any difference in performance between face and shape matching (Table 1). Trials without response or a response time (RT) <100 ms were excluded from the behavioral analysis (3.9% of all trials).

Whole-brain analysis

The first-level analysis was conducted by building a general linear model using the individual onset and length of each trial (4 s) for face-matching and shape-matching convolved with the canonical hemodynamic response function as implemented in SPM12. To model the valence of faces, we added three regressors for each of the respective conditions. The final GLM for the whole-brain analysis included five regressors of interest: a regressor for all faces, three parametric modulation regressors for each valence, that is for positive, negative, and neutral faces, and one regressor for shapes. The regressor “all faces” included 30 events, while parametric modulation regressors modeling positive, negative, and neutral conditions comprised 10 events each. The regressor “shapes” included 30 events. In addition, we added the six realignment parameters derived from preprocessing as nuisance regressors.

The main effect of our task (face-matching > shape-matching) was investigated with a one-sample t test using the respective contrast files of each subject. To examine the effect of valence, we performed an F test in a second-level repeated measures ANOVA design across the positive, negative, and neutral valence conditions. For both analyses the cluster-based family-wise error corrected significance threshold was set to pFWEc = 0.05, the uncorrected voxel-wise cluster-defining threshold was set to pCTD = 0.001. Labels of resulting brain regions were determined using the SPM Anatomy Toolbox (Eickhoff et al., 2007).

Dynamic causal modeling (DCM)

DCM is a hypothesis-driven Bayesian model comparison procedure for inferring effective connectivity between brain regions (Friston et al., 2003). DCM allows for the creation of different models to investigate the directed interactions of specific brain regions under experimentally controlled perturbations. These interactions are modeled at the neuronal level and related to the observable measurement via a hemodynamic forward model (Buxton et al., 1998). Importantly, it allows for estimation of endogenous coupling and context-specific, modulatory coupling (Friston et al., 2003; Penny et al., 2004). The neural model is given by the neural state equationEmbedded Image in which the vector z represents the time series of the neural signal in a given region of interest and u represents the experimental inputs (1 … j). Intrinsic (endogenous) coupling parameters between regions are stored in matrix A, modulatory parameters for a stimulus uj are stored in matrix B, and direct driving inputs for regions are described in matrix C.

Regional time series extraction

In our study, we focused on the analysis of an emotion processing network model comprising four regions, whose adequacy has been demonstrated in previous studies (Almeida et al., 2009a, 2011; Sladky et al., 2015a). In particular, we included (1) the ventrolateral part of the LPFC that is associated with emotion regulation (Hariri et al., 2003; Morris et al., 2012; Wagner and Heatherton, 2013); (2) the MPFC that is involved in integrating affective and contextual information, valence processing (Roy et al., 2012), and autonomous emotion regulation (Phillips et al., 2008); (3) the AMY for its role in salience detection and facial emotion processing (Phelps, 2006); and (4) the fusiform face area (FFA) as part of the visual system, that is sensitive to faces (Kanwisher et al., 1997).

The selection and functional localization of our volumes of interest (VOI) in the AMY-prefrontal network was guided by the results of the second-level group analyses (Table 2), similar to previous work (Hauser et al., 2014; Sladky et al., 2015a). For the AMY, the LPFC, and the FFA, we specified a spherical search volume at the peak of the face-matching > shape-matching contrast: AMY x = 21, y = –10, z = –14 mm (MNI); the right LPFC x = 47, y = 30, z = 8 mm (MNI); the right FFA x = 41, y = –44, z = –22 mm (MNI). In addition, we defined a search volume for the MPFC at the peak of the main effect of valence in the second-level ANOVA comparing positive, negative, and neutral valence conditions [x = 3, y = 50, z = –2 mm (MNI)]. The individual VOI center coordinates were restricted to not differ >12 mm (corresponding to twice the FWHM of the smoothing kernel) from the group maximum to ensure comparability between subjects.

Subjects’ individual spherical VOIs were centered at the individual peaks (r = 6 mm, p < 0.05, uncorrected) in the respective contrast and the first eigenvariate was extracted as summary statistic for all active voxels within the VOI. One subject was excluded from the DCM analysis, because we did not find any active voxels in the LPFC for the chosen threshold. We restricted our analysis to the right hemisphere, as previous studies suggested that it preferentially engages in processing of nonverbal emotional cues, such as emotional faces (Puce et al., 1996; Kanwisher et al., 1997; Anderson et al., 2003; Ochsner et al., 2004; Fairhall and Ishai, 2007; Sladky et al., 2015a).

Model space

We assumed bidirectional connection between MPFC and AMY, and MPFC and LPFC. Although there is evidence that direct connections between LPFC and AMY are only very sparse (Ray and Zald, 2012), it is possible that they exert influence via indirect pathways over each other. Hence, we included models with all possible intrinsic connectivity patterns between the LPFC and the AMY in the model space. In addition, we specified bidirectional intrinsic connections between the FFA and the LPFC and the FFA and the AMY, respectively. Modulation by valence was varied systematically across connections between MPFC and LPFC, and MPFC and AMY in all possible modulation patterns, spanning a model space of 256 models.

For the DCM analysis we specified a second GLM that included five regressors of interest (all stimuli, all faces, positive faces, negative faces, and shapes) and the six realignment parameters as nuisance regressors. The “all stimuli” regressor included 60 trials, “all faces” included all 30 face events, regressors modeling positive, negative comprised 10 events each, and the “shape” regressor included all 30 shape-matching trials. The “all faces” regressor served as driving input of the FFA.

We performed random-effects family-wise Bayesian model selection (BMS; Penny et al., 2010) as implemented in SPM12 to compute the expected posterior probabilities and the exceedance probabilities of model families within our sample. For model comparison, the BMS procedure uses the free energy that is a lower-bound approximation to the log-model evidence that accounts for both model accuracy and model complexity (Penny et al., 2004; Penny, 2012).

To test different functional architectures of contextual modulation, we created four different families of models (Fig. 2). These model families differed in terms of connections on which emotional valence modulated effective connectivity. We created families with no contextual (valence) modulations (one model), bottom-up modulations (15 models), top-down modulations (15 models), and bidirectional modulations (225 models).

Figure 2.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 2.

DCM model specification. We specified 256 models and grouped them into four families, depending on the location of the modulations of positive and negative valence. The modulations are depicted as red dots on the connections. In each family, all possible combinations of modulations were grouped together yielding one model with no modulation, 15 models with bottom-up modulations, 15 models with top-down modulations, and 225 models with bidirectional modulations. All faces were the driving input to the FFA.

We used Bayesian model averaging (BMA) across models to make further inferences on the significance of connections and modulation by valence (Penny et al., 2010). BMA allows for averaging the parameters while they are weighted by the posterior probability of the model and thereby accounting for the uncertainty of individual models (Stephan et al., 2010).

Subsequent one-sample t tests of averaged parameter estimates were conducted in MATLAB (MATLAB and Statistics Toolbox Release 2017a, The MathWorks, Inc.). We accounted for multiple t tests of the connectivity parameters by using the procedure of Benjamini and Hochberg (1995) to control the false discovery rate (FDR; adjusted pFDRc < 0.05).

Results

Behavioral analysis

The behavioral analysis of the RT and the accuracy across different conditions is summarized in Table 1 and depicted in Figure 1B.

View this table:
  • View inline
  • View popup
Table 1.

Behavioral results of the analysis of the behavioral data

Responses during shape-matching were significantly faster than during face-matching (t(32) = 5.97, p < 10–5). Accuracy (% correct) did not differ significantly between face-matching and shape-matching (t(32) = –2.01, p = 0.053). There was a main effect of valence on accuracy for the three valence conditions (F(2.32) = 24.02, p < 10–7). Pairwise comparisons indicated that accuracy during trials with neutral faces was lower than trials with positive (t(32) = 5.56, p < 10–5) and negative faces (t(32) = 4.74, p < 0.0001), suggesting a higher difficulty in matching faces of the neutral condition. Positive and negative face conditions did not differ in accuracy (t(32) = 1.28, p = 0.21). A similar accuracy pattern was reported in previous work (Aybek et al., 2015). One-sample t tests across positive (t(32) = 32.69, p < 10–25), negative (t(32) = 36.92, p < 10–275) and neutral valence (t(32) = 9.69, p < 10–10) showed that accuracies were well beyond chance level (50%). Importantly, in a repeated measures ANOVA we did not find any significant differences in RTs across valence conditions (F(2.32) = 1.79, p = 0.175).

Whole-brain results

The dynamic face-matching and shape-matching task showed a significant effect of task (face-matching > shape-matching) in brain regions commonly recruited during face processing (Fusar-Poli et al., 2009), including the AMY, the fusiform gyrus, the LPFC, the middle and superior temporal gyrus (Fig. 3A; Table 2). A repeated measures ANOVA (Fig. 3B) across valence conditions using the respective contrast images revealed a main effect of valence in the MPFC (F(2.32) = 17.54, pFWEc < 10–7), the right medial temporal lobe (F(2.32) = 18.14, pFWEc = 0.043), the superior temporal gyrus (F(2.32) = 16.13, pFWEc = 0.002), the left medial temporal lobe (F(2.32) = 15.34, pFWEc < 10–5), the left cerebellum (F(2.32) = 14.79, pFWEc = 0.041), the left AMY (F(2.32) = 14.72, pFWEc = 0.009), and the right parahippocampal gyrus (F(2.32) = 14.70, pFWEc = 0.001). Post-hoc t tests showed that the effect in the MPFC was driven by negative valence (Table 2).

Figure 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 3.

Whole-brain group analyses. A, Main effect of task face-matching > shape-matching. B, Main effect of valence in the MPFC. Both images thresholded at pFWEc < 0.05, with a voxel-wise cluster-defining threshold of pCDT < 0.001, n = 33. Color is mapped to t values (A) and F values (B).

View this table:
  • View inline
  • View popup
Table 2.

Results of the group analysis (n = 33)

DCM

Family-wise model comparison

In a first step, we compared different model families (Fig. 4). The model family with bidirectional modulations of connections outperformed all other families with an expected posterior probability of 42% and an exceedance probability of 82%. As the model space incorporated a wide range of plausible models, we subsequently performed Bayesian model averaging to infer on the model parameters of the winning family.

Figure 4.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 4.

Family-wise Bayesian model comparison. Expected posterior probabilities and exceedance probabilities for the four specified model families. Asterisks (*) indicate the winning model family.

Bayesian model averaging

The results from Bayesian model averaging (BMA; Table 3; Fig. 5) emphasize the relevance of connections between the AMY and the MPFC during processing of emotional faces. One-sample t tests for consistency across subjects showed that the average endogenous connectivity is significant between those regions. In addition, we found significant modulation of connectivity by valence on these connections. While the effective connectivity from the MPFC to the AMY was modulated by positive and negative valence, the modulation of effective connectivity of the bottom-up connection from the AMY to the MPFC was significant for the positive but not the negative condition.

View this table:
  • View inline
  • View popup
Table 3.

Mean and SD of endogenous and modulatory parameter estimates for all connections across all subjects and across the models of the winning bidirectional family, and the respective p value resulting from a one-sample t test (corrected for multiple comparisons)

Figure 5.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 5.

Effective connectivity during face-matching and its modulation by positive and negative valence. Parameters have been averaged with Bayesian model averaging, across all subjects and models of the winning model family. We found significant dampening of effective connectivity from the AMY to the MPFC during processing of positively valenced faces. Arrow thickness indicate effective connectivity values: thick > 0.15, medium > 0.10, thin > 0.05, dashed: not significant.

Average intrinsic connectivity between the MPFC and LPFC differed significantly from zero across subjects. In addition, the connection between the LPFC and the MPFC showed a significant modulation effect of positive and negative valence, suggesting a specific role of this connection during processing of emotional stimuli.

After averaging the intrinsic and modulatory connectivity parameters, we correlated each subject’s individual connectivity parameters with the behavioral data from the task (mean accuracy and the mean RTs for the different valence conditions). However, none of the correlations remained significant after correction for multiple comparisons.

Discussion

Our study examined the valence-dependent functional architecture of the prefrontal-AMY network during emotion processing using statistical parametric mapping and DCM. We used a dynamic face-matching and shape-matching paradigm in healthy subjects to assess activity and connectivity in regions supporting emotion processing and, subsequently, whether emotional valence modulates effective connectivity of bottom-up (salience signals), top-down (evaluation and regulation signals), or bidirectional connections. The results of our study suggest three main conclusions.

First, we corroborated earlier studies by showing that the MPFC as a core region of emotional response regulation is especially sensitive to negative affect (Ochsner et al., 2012). Our data suggest that during processing of negative valence the MPFC and the right AMY are more strongly activated than during processing of neutral valence. Second, we directly demonstrated that activity in key regions of the prefrontal-AMY network during emotion processing is best explained by bidirectional contextual modulation of effective connectivity by valence. Accordingly, processing emotional valence directly induces changes of coupling strengths within the prefrontal-AMY circuitry. In particular, model averaging showed that the bidirectional coupling between MPFC and AMY and unidirectional coupling between LPFC and MPFC were modulated by affective cues. This suggests that the MPFC not only serves the integration of bottom-up and top-down signals, but also continuously exerts influence on the AMY during face-matching. Third, we found evidence for a differential effect of valence on coupling between regions. On the one hand, effective connectivity from the MPFC to the AMY was modulated by both positive and negative valence, while on the other hand effective connectivity from the AMY to the MPFC was only significantly modulated by positive valence. Additionally, the connectivity from the LPFC to the MPFC was augmented during positive and negative valence processing. Previous studies have highlighted the role of the MPFC during emotion processing and they have underlined the role of the MPFC in processing of valence (Kawasaki et al., 2001; Winecoff et al., 2013). Studies using explicit emotion regulation paradigms have repeatedly shown that the activation of the MPFC is increased during the reappraisal of negative emotion (Urry et al., 2006; Delgado et al., 2008). Thus, it has been suggested that the MPFC supports the control of emotional responses. Moreover, recent work proposed that the involvement of the MPFC during emotion processing is related to the encoding of an integrated affective value of a stimulus (Smith et al., 2010; Winecoff et al., 2013). Importantly, this integrated affective value encoded in the MPFC might be crucially dependent on the confidence in the aggregated information (Lebreton et al., 2015), which might modulate the BOLD signal in the MPFC following a U-shape pattern (Barron et al., 2015). Our data show significantly increased activity in the MPFC during the processing of emotional stimuli with negative valence. Our findings may therefore reflect the encoding of biological significance of negatively valenced faces and provide further evidence for the encoding of stimulus valence in the MPFC. Accordingly, the lower responses in the MPFC in the neutral condition could reflect either lower relevance or lower confidence in the nature of stimuli. This would emphasize the role of the MPFC in the integration of affective information within a valence-sensitive network, computing a value for biological significance for a given stimulus.

This is supported by our modeling results showing that the MPFC integrates affective information from multiple routes. Model averaging demonstrated that the bidirectional coupling between the AMY and MPFC and the coupling from the LPFC to the MPFC are valence sensitive. This is not only in line with recent theories of distributed processing of emotional stimuli along multiple parallel pathways (Pessoa and Adolphs, 2010), but also provides direct evidence for the idea that the prefrontal-AMY circuitry can change its functional state to support appropriate mental functions for a given context (Pessoa, 2017), which potentially requires action.

Bayesian model averaging showed that positive valence significantly dampened the intrinsic connectivity between the AMY and the MPFC, whereas negatively valenced faces did not. As task difficulty between negative and positive conditions was comparable, the observed difference in connectivity strongly suggests valence-sensitivity of the coupling between these regions. That said, the similar effective connectivity of the bottom-up connectivity during neutral and negative blocks was particularly interesting to us, since the pathway from the AMY to the MPFC is thought to be specifically sensitive to negative valence (e.g., fear conditioning; for review, see Kim et al., 2011). Notably, this is the first study that investigated valence-dependent effective connectivity within the prefrontal-AMY network using dynamic faces. The similar connectivity pattern of the neutral and negative condition might stem from the AMY’s role as a significance detector. The AMY has been extensively studied and there is a broad consensus on the relevance of this brain structure in face processing (Adolphs, 2002), and more generally, the immediate detection of biological significance (Sander et al., 2003), or resolving uncertainty (Whalen, 1998). Neutrally (i.e., ambiguously) and negatively valenced faces might induce increased predictive uncertainty compared to positively valenced faces (Whalen et al., 2013). A plausible brain response to react to predictive uncertainty would be to relocate cognitive resources to resolve it (Bubic et al., 2010). The coupling between AMY and MPFC during processing neutral and negative facial expressions might therefore reflect a signal that translates into a need for action to increase precision and, hence, regain confidence in a volatile environment. In line with this, we found significant effective connectivity from the AMY to MPFC during processing of neutral and negative valence, reflected in a significant intrinsic connectivity and its non-significant modulation during negative blocks, which might reflect a bottom-up confidence signal from the AMY.

Furthermore, we found positive intrinsic connectivity from the MPFC to the AMY during the dynamic face processing that was modulated negatively during positive and negative blocks. This is in line with previous work that observed negative effective connectivity between these regions in healthy subjects using a similar task with static emotional faces (Sladky et al., 2015a) and might reflect a downregulation mechanism of automatic dampening the emotional response of the AMY by the MPFC to negative emotional cues (Ochsner et al., 2012).

Corticocortical effective connectivity between LPFC and MPFC was significant during face processing. A general valence-independent face-sensitive coupling between LPFC and MPFC in our dynamic task could reflect a cognitive attenuation of significance of emotional stimuli, which would eventually yield an adaption of emotional responses mediated by the MPFC. In this regard, the LPFC has been implicated in emotion regulation strategies such as repression (Anderson et al., 2004) or (spontaneous) reappraisal (Drabant et al., 2009). Based on our modeling results, we therefore propose that the valence-sensitive recruitment of the MPFC originates from the integration of affective information stemming from valence-dependent coupling within the prefrontal-AMY network. Our findings for the afferent connections of the LPFC are in agreement with the results reported in a previous study (Sladky et al., 2015a), that found an up-modulating effect of the AMY on the LPFC and a down-modulating effect of medial prefrontal regions over lateral ones. On the one hand, enhanced activation of the LPFC via the AMY could reflect the allocation of attentional resources toward emotionally salient stimuli of high biological significance, on the other hand, the downmodulating signal from the MPFC might support the continuous release of these resources (Bishop et al., 2004; Sladky et al., 2015a). Interestingly, we did not observe any evidence of valence sensitivity of the connection from the MPFC to the LPFC, suggesting a general downmodulating role of this connection during face-matching.

Forward connections from the FFA showed significant intrinsic connectivity to the AMY and the LPFC, while the backward connections were not. This feed-forward functional architecture during face processing has been described previously (Fairhall and Ishai, 2007), and our results confirm these findings.

One limitation regarding the interpretation of our findings are the emotional categories of the faces used in our paradigm. Unlike many previous studies (Fusar-Poli et al., 2009; Zinchenko et al., 2018), we did not use angry or fearful faces for the negative condition, but sad and disgusted faces. This choice was made deliberately to reduce the effects of arousal (Remmington et al., 2000; Trautmann et al., 2009). Thus, our findings should be only interpreted in regard to the emotional expressions used in our paradigm. Despite this limitation, our results are in concordance with previous findings in the literature and provide further evidence that the state of the prefrontal-AMY network is sensitive to valence. The goal of future investigations should be to assess whether our results are generalizable to other negative emotional expressions and to negatively valenced stimuli, other than faces, in general.

To conclude, using DCM analysis we showed valence-dependent coupling changes within the emotion processing circuitry during a dynamic face-matching paradigm. Our findings are in agreement with recent theories of affect processing that stress the highly dynamic nature of network interactions. It has been suggested that these interactions do not only depend on task-demands, but, as our empirical data suggest, on the emotional valence of a stimulus (Kim et al., 2004; Pessoa, 2017). Understanding mechanisms of dynamic integration of affective value in the emotion processing network might be pivotal for explaining psychopathologies. A dysregulation of the prefrontal-AMY network has been found in various psychiatric disorders. A disruption of neural circuitry underlying successful emotion regulation is a hallmark of various psychiatric conditions such as mood and anxiety disorders in adults (Johnstone et al., 2007; Almeida et al., 2009b; Etkin et al., 2010; Liao et al., 2010; Demenescu et al., 2013; Sladky et al., 2015a; Minkova et al., 2017) and adolescents (Monk et al., 2008; Perlman et al., 2012; Keshavan et al., 2014) and dysfunctional valence-dependent coupling might underlie the attention and processing bias in mood disorders (Disner et al., 2011; Groenewold et al., 2013; Clark et al., 2018). Our study provides strong evidence for alterations of coupling as a function of valence within the prefrontal-AMY network. Based on our results, such a dynamic face-matching task thus may aid future studies to probe and disentangle mechanisms of attentional bias and valence-sensitive emotional dysregulation in neuropsychiatric disorders.

Acknowledgments

Acknowledgements: We thank Nathalie Holz for her suggestions on the task design.

Footnotes

  • S.W. has received lecture honoraria from Opopharma in the last three years. Outside professional activities and interests of S.Brem and S.W. are declared under the link of the University of Zurich www.uzh.ch/prof/ssl-dir/interessenbindungen/client/web/. All other authors declare no competing financial interests.

  • This work was supported by the Department of Children and Adolescent Psychiatry and Psychotherapy, University Hospital of Psychiatry Zurich, University of Zurich.

This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International license, which permits unrestricted use, distribution and reproduction in any medium provided that the original work is properly attributed.

References

  1. ↵
    Adolphs R (2002) Recognizing emotion from facial expressions: psychological and neurological mechanisms. Behav Cogn Neurosci Rev 1:21–62. doi:10.1177/1534582302001001003 pmid:17715585
    OpenUrlCrossRefPubMed
  2. ↵
    Almeida J, Mechelli A, Hassel S, Versace A, Kupfer DJ, Phillips ML (2009a) Abnormally increased effective connectivity between parahippocampal gyrus and ventromedial prefrontal regions during emotion labeling in bipolar disorder. Psychiatry Res 174:195–201. doi:10.1016/j.pscychresns.2009.04.015
    OpenUrlCrossRefPubMed
  3. ↵
    Almeida J, Versace A, Mechelli A, Hassel S, Quevedo K, Kupfer DJ, Phillips ML (2009b) Abnormal amygdala-prefrontal effective connectivity to happy faces differentiates bipolar from major depression. Biol Psychiatry 66:451–459. doi:10.1016/j.biopsych.2009.03.024
    OpenUrlCrossRefPubMed
  4. ↵
    Almeida J, Kronhaus D, Sibille E, Langenecker S, Versace A, LaBarbara E, Phillips ML (2011) Abnormal left-sided orbitomedial prefrontal cortical–amygdala connectivity during happy and fear face processing: a potential neural mechanism of female MDD. Front Psychiatry 2:69. doi:10.3389/fpsyt.2011.00069 pmid:22163223
    OpenUrlCrossRefPubMed
  5. ↵
    Anderson AK, Christoff K, Panitz D, De Rosa E, Gabrieli JD (2003) Neural correlates of the automatic processing of threat facial signals. J Neurosci 23:5627–5633. pmid:12843265
    OpenUrlAbstract/FREE Full Text
  6. ↵
    Anderson MC, Ochsner KN, Kuhl B, Cooper J, Robertson E, Gabrieli SW, Glover GH, Gabrieli JD (2004) Neural systems underlying the suppression of unwanted memories. Science 303:232–235. doi:10.1126/science.1089504 pmid:14716015
    OpenUrlAbstract/FREE Full Text
  7. ↵
    Arsalidou M, Morris D, Taylor MJ (2011) Converging evidence for the advantage of dynamic facial expressions. Brain Topogr 24:149–163. doi:10.1007/s10548-011-0171-4 pmid:21350872
    OpenUrlCrossRefPubMed
  8. ↵
    Aybek S, Nicholson TR, O’Daly O, Zelaya F, Kanaan RA, David AS (2015) Emotion-motion interactions in conversion disorder: an FMRI study. PLoS One 10:e0123273. doi:10.1371/journal.pone.0123273 pmid:25859660
    OpenUrlCrossRefPubMed
  9. ↵
    Ball T, Derix J, Wentlandt J, Wieckhorst B, Speck O, Schulze-Bonhage A, Mutschler I (2009) Anatomical specificity of functional amygdala imaging of responses to stimuli with positive and negative emotional valence. J Neurosci Methods 180:57–70. doi:10.1016/j.jneumeth.2009.02.022 pmid:19427530
    OpenUrlCrossRefPubMed
  10. ↵
    Barron HC, Garvert MM, Behrens TE (2015) Reassessing VMPFC: full of confidence? Nat Neurosci 18:1064. doi:10.1038/nn.4076 pmid:26216461
    OpenUrlCrossRefPubMed
  11. ↵
    Benjamini Y, Hochberg Y (1995) Controlling the false discovery rate: a practical and powerful approach to multiple testing. J R Stat Soc Series B Stat Methodol 57:289–300. doi:10.1111/j.2517-6161.1995.tb02031.x
    OpenUrlCrossRef
  12. ↵
    Bishop S, Duncan J, Brett M, Lawrence AD (2004) Prefrontal cortical function and anxiety: controlling attention to threat-related stimuli. Nat Neurosci 7:184. doi:10.1038/nn1173 pmid:14703573
    OpenUrlCrossRefPubMed
  13. ↵
    Bubic A, Von Cramon DY, Schubotz RI (2010) Prediction, cognition and the brain. Front Hum Neurosci 4:25. doi:10.3389/fnhum.2010.00025 pmid:20631856
    OpenUrlCrossRefPubMed
  14. ↵
    Buxton RB, Wong EC, Frank LR (1998) Dynamics of blood flow and oxygenation changes during brain activation: the balloon model. Magn Reson Med 39:855–864. pmid:9621908
    OpenUrlCrossRefPubMed
  15. ↵
    Clark JE, Watson S, Friston KJ (2018) What is mood? A computational perspective. Psychol Med 48:2277–2284. doi:10.1017/S0033291718000430 pmid:29478431
    OpenUrlCrossRefPubMed
  16. ↵
    Comte M, Schön D, Coull JT, Reynaud E, Khalfa S, Belzeaux R, Ibrahim EC, Guedj E, Blin O, Weinberger DR, Fakra E (2016) Dissociating bottom-up and top-down mechanisms in the cortico-limbic system during emotion processing. Cereb Cortex 26:144–155. doi:10.1093/cercor/bhu185 pmid:25165065
    OpenUrlCrossRefPubMed
  17. ↵
    Delgado MR, Nearing KI, LeDoux JE, Phelps EA (2008) Neural circuitry underlying the regulation of conditioned fear and its relation to extinction. Neuron 59:829–838. doi:10.1016/j.neuron.2008.06.029 pmid:18786365
    OpenUrlCrossRefPubMed
  18. ↵
    Demenescu L, Kortekaas R, Cremers H, Renken R, van Tol M, van der Wee N, Veltman D, den Boer J, Roelofs K, Aleman A (2013) Amygdala activation and its functional connectivity during perception of emotional faces in social phobia and panic disorder. J Psychiatr Res 47:1024–1031. doi:10.1016/j.jpsychires.2013.03.020 pmid:23643103
    OpenUrlCrossRefPubMed
  19. ↵
    Dichter GS, Felder JN, Smoski MJ (2009) Affective context interferes with cognitive control in unipolar depression: an fMRI investigation. J Affect Disord 114:131–142. doi:10.1016/j.jad.2008.06.027 pmid:18706701
    OpenUrlCrossRefPubMed
  20. ↵
    Dima D, Stephan KE, Roiser JP, Friston KJ, Frangou S (2011) Effective connectivity during processing of facial affect: evidence for multiple parallel pathways. J Neurosci 31:14378–14385. doi:10.1523/JNEUROSCI.2400-11.2011 pmid:21976523
    OpenUrlAbstract/FREE Full Text
  21. ↵
    Disner SG, Beevers CG, Haigh EA, Beck AT (2011) Neural mechanisms of the cognitive model of depression. Nat Rev Neurosci 12:467–477. doi:10.1038/nrn3027 pmid:21731066
    OpenUrlCrossRefPubMed
  22. ↵
    Drabant EM, McRae K, Manuck SB, Hariri AR, Gross JJ (2009) Individual differences in typical reappraisal use predict amygdala and prefrontal responses. Biol Psychiatry 65:367–373. doi:10.1016/j.biopsych.2008.09.007 pmid:18930182
    OpenUrlCrossRefPubMed
  23. ↵
    Eickhoff SB, Paus T, Caspers S, Grosbras M-H, Evans AC, Zilles K, Amunts K (2007) Assignment of functional activations to probabilistic cytoarchitectonic areas revisited. Neuroimage 36:511–521. doi:10.1016/j.neuroimage.2007.03.060 pmid:17499520
    OpenUrlCrossRefPubMed
  24. ↵
    Etkin A, Büchel C, Gross JJ (2015) The neural bases of emotion regulation. Nat Rev Neurosci 16:693–700. doi:10.1038/nrn4044 pmid:26481098
    OpenUrlCrossRefPubMed
  25. ↵
    Etkin A, Prater KE, Hoeft F, Menon V, Schatzberg AF (2010) Failure of anterior cingulate activation and connectivity with the amygdala during implicit regulation of emotional processing in generalized anxiety disorder. Am J Psychiatry 167:545–554. doi:10.1176/appi.ajp.2009.09070931 pmid:20123913
    OpenUrlCrossRefPubMed
  26. ↵
    Fairhall SL, Ishai A (2007) Effective connectivity within the distributed cortical network for face perception. Cereb Cortex 17:2400–2406. doi:10.1093/cercor/bhl148 pmid:17190969
    OpenUrlCrossRefPubMed
  27. ↵
    Friston KJ, Harrison L, Penny W (2003) Dynamic causal modelling. Neuroimage 19:1273–1302. doi:10.1016/S1053-8119(03)00202-7
    OpenUrlCrossRefPubMed
  28. ↵
    Furl N, Henson RN, Friston KJ, Calder AJ (2015) Network interactions explain sensitivity to dynamic faces in the superior temporal sulcus. Cereb Cortex 25:2876–2882. doi:10.1093/cercor/bhu083 pmid:24770707
    OpenUrlCrossRefPubMed
  29. ↵
    Fusar-Poli P, Placentino A, Carletti F, Landi P, Allen P, Surguladze S, Benedetti F, Abbamonte M, Gasparotti R, Barale F, Perez J, McGuire P, Politi P (2009) Functional atlas of emotional faces processing: a voxel-based meta-analysis of 105 functional magnetic resonance imaging studies. J Psychiatry Neurosci 34:418–432. pmid:19949718
    OpenUrlCrossRefPubMed
  30. ↵
    Gläscher J, Tüscher O, Weiller C, Büchel C (2004) Elevated responses to constant facial emotions in different faces in the human amygdala: an fMRI study of facial identity and expression. BMC Neurosci 5:45. doi:10.1186/1471-2202-5-45 pmid:15548329
    OpenUrlCrossRefPubMed
  31. ↵
    Groenewold NA, Opmeer EM, de Jonge P, Aleman A, Costafreda SG (2013) Emotional valence modulates brain functional abnormalities in depression: evidence from a meta-analysis of fMRI studies. Neurosci Biobehav Rev 37:152–163. doi:10.1016/j.neubiorev.2012.11.015 pmid:23206667
    OpenUrlCrossRefPubMed
  32. ↵
    Hariri AR, Tessitore A, Mattay VS, Fera F, Weinberger DR (2002) The amygdala response to emotional stimuli: a comparison of faces and scenes. Neuroimage 17:317–323. pmid:12482086
    OpenUrlCrossRefPubMed
  33. ↵
    Hariri AR, Mattay VS, Tessitore A, Fera F, Weinberger DR (2003) Neocortical modulation of the amygdala response to fearful stimuli. Biol Psychiatry 53:494–501. pmid:12644354
    OpenUrlCrossRefPubMed
  34. ↵
    Hauser TU, Iannaccone R, Stämpfli P, Drechsler R, Brandeis D, Walitza S, Brem S (2014) The feedback-related negativity (FRN) revisited: new insights into the localization, meaning and network organization. Neuroimage 84:159–168. doi:10.1016/j.neuroimage.2013.08.028 pmid:23973408
    OpenUrlCrossRefPubMed
  35. ↵
    Jin J, Zelano C, Gottfried JA, Mohanty A (2015) Human amygdala represents the complete spectrum of subjective valence. J Neurosci 35:15145–15156. doi:10.1523/JNEUROSCI.2450-15.2015 pmid:26558785
    OpenUrlAbstract/FREE Full Text
  36. ↵
    Johnstone T, van Reekum CM, Urry HL, Kalin NH, Davidson RJ (2007) Failure to regulate: counterproductive recruitment of top-down prefrontal-subcortical circuitry in major depression. J Neurosci 27:8877–8884. doi:10.1523/JNEUROSCI.2063-07.2007 pmid:17699669
    OpenUrlAbstract/FREE Full Text
  37. ↵
    Kanwisher N, McDermott J, Chun MM (1997) The fusiform face area: a module in human extrastriate cortex specialized for face perception. J Neurosci 17:4302–4311. pmid:9151747
    OpenUrlAbstract/FREE Full Text
  38. ↵
    Kawasaki H, Adolphs R, Kaufman O, Damasio H, Damasio AR, Granner M, Bakken H, Hori T, Howard IIM (2001) Single-neuron responses to emotional visual stimuli recorded in human ventral prefrontal cortex. Nat Neurosci 4:15. doi:10.1038/82850
    OpenUrlCrossRefPubMed
  39. ↵
    Keshavan MS, Giedd J, Lau JY, Lewis DA, Paus T (2014) Changes in the adolescent brain and the pathophysiology of psychotic disorders. Lancet Psychiatry 1:549–558. doi:10.1016/S2215-0366(14)00081-9 pmid:26361314
    OpenUrlCrossRefPubMed
  40. ↵
    Kessler H, Doyen-Waldecker C, Hofer C, Hoffmann H, Traue HC, Abler B (2011) Neural correlates of the perception of dynamic versus static facial expressions of emotion. Psychosoc Med 8:Doc03.
    OpenUrlPubMed
  41. ↵
    Kim H, Somerville LH, Johnstone T, Polis S, Alexander AL, Shin LM, Whalen PJ (2004) Contextual modulation of amygdala responsivity to surprised faces. J Cogn Neurosci 16:1730–1745. doi:10.1162/0898929042947865 pmid:15701225
    OpenUrlCrossRefPubMed
  42. ↵
    Kim MJ, Loucks RA, Palmer AL, Brown AC, Solomon KM, Marchante AN, Whalen PJ (2011) The structural and functional connectivity of the amygdala: from normal emotion to pathological anxiety. Behav Brain Res 223:403–410. doi:10.1016/j.bbr.2011.04.025 pmid:21536077
    OpenUrlCrossRefPubMed
  43. ↵
    Langner O, Dotsch R, Bijlstra G, Wigboldus DHJ, Hawk ST, van Knippenberg A (2010) Presentation and validation of the Radboud faces database. Cogn Emot 24:1377–1388. doi:10.1080/02699930903485076
    OpenUrlCrossRef
  44. ↵
    Lebreton M, Abitbol R, Daunizeau J, Pessiglione M (2015) Automatic integration of confidence in the brain valuation signal. Nat Neurosci 18:1159. doi:10.1038/nn.4064 pmid:26192748
    OpenUrlCrossRefPubMed
  45. ↵
    Liao W, Qiu C, Gentili C, Walter M, Pan Z, Ding J, Zhang W, Gong Q, Chen H (2010) Altered effective connectivity network of the amygdala in social anxiety disorder: a resting-state FMRI study. PLoS One 5:e15238. doi:10.1371/journal.pone.0015238 pmid:21203551
    OpenUrlCrossRefPubMed
  46. ↵
    Lindquist KA, Satpute AB, Wager TD, Weber J, Barrett LF (2016) The brain basis of positive and negative affect: evidence from a meta-analysis of the human neuroimaging literature. Cereb Cortex 26:1910–1922. doi:10.1093/cercor/bhv001 pmid:25631056
    OpenUrlCrossRefPubMed
  47. ↵
    McRae K, Misra S, Prasad AK, Pereira SC, Gross JJ (2012) Bottom-up and top-down emotion generation: implications for emotion regulation. Soc Cogn Affect Neurosci 7:253–262. doi:10.1093/scan/nsq103 pmid:21296865
    OpenUrlCrossRefPubMed
  48. ↵
    Minkova L, Sladky R, Kranz GS, Woletz M, Geissberger N, Kraus C, Lanzenberger R, Windischberger C (2017) Task-dependent modulation of amygdala connectivity in social anxiety disorder. Psychiatry Res Neuroimaging 262:39–46. doi:10.1016/j.pscychresns.2016.12.016 pmid:28226306
    OpenUrlCrossRefPubMed
  49. ↵
    Monk CS, Telzer EH, Mogg K, Bradley BP, Mai X, Louro HM, Chen G, McClure-Tone EB, Ernst M, Pine DS (2008) Amygdala and ventrolateral prefrontal cortex activation to masked angry faces in children and adolescents with generalized anxiety disorder. Arch Gen Psychiatry 65:568–576. doi:10.1001/archpsyc.65.5.568 pmid:18458208
    OpenUrlCrossRefPubMed
  50. ↵
    Morris R, Sparks A, Mitchell P, Weickert C, Green M (2012) Lack of cortico-limbic coupling in bipolar disorder and schizophrenia during emotion regulation. Transl Psychiatry 2:e90. doi:10.1038/tp.2012.16 pmid:22832855
    OpenUrlCrossRefPubMed
  51. ↵
    Neta M, Davis FC, Whalen PJ (2011) Valence resolution of ambiguous facial expressions using an emotional oddball task. Emotion 11:1425. doi:10.1037/a0022993 pmid:21707167
    OpenUrlCrossRefPubMed
  52. ↵
    Ochsner KN, Gross JJ (2005) The cognitive control of emotion. Trends Cogn Sci 9:242–249. doi:10.1016/j.tics.2005.03.010 pmid:15866151
    OpenUrlCrossRefPubMed
  53. ↵
    Ochsner KN, Ray RD, Cooper JC, Robertson ER, Chopra S, Gabrieli JD, Gross JJ (2004) For better or for worse: neural systems supporting the cognitive down-and up-regulation of negative emotion. Neuroimage 23:483–499. doi:10.1016/j.neuroimage.2004.06.030
    OpenUrlCrossRefPubMed
  54. ↵
    Ochsner KN, Ray RR, Hughes B, McRae K, Cooper JC, Weber J, Gabrieli JD, Gross JJ (2009) Bottom-up and top-down processes in emotion generation: common and distinct neural mechanisms. Psychol Sci 20:1322–1331. doi:10.1111/j.1467-9280.2009.02459.x pmid:19883494
    OpenUrlCrossRefPubMed
  55. ↵
    Ochsner KN, Silvers JA, Buhle JT (2012) Functional imaging studies of emotion regulation: a synthetic review and evolving model of the cognitive control of emotion. Ann NY Acad Sci 1251:E1–E24. doi:10.1111/j.1749-6632.2012.06751.x
    OpenUrlCrossRefPubMed
  56. ↵
    Penny WD (2012) Comparing dynamic causal models using AIC, BIC and free energy. Neuroimage 59:319–330. doi:10.1016/j.neuroimage.2011.07.039 pmid:21864690
    OpenUrlCrossRefPubMed
  57. ↵
    Penny WD, Stephan KE, Mechelli A, Friston KJ (2004) Comparing dynamic causal models. Neuroimage 22:1157–1172. doi:10.1016/j.neuroimage.2004.03.026 pmid:15219588
    OpenUrlCrossRefPubMed
  58. ↵
    Penny WD, Stephan KE, Daunizeau J, Rosa MJ, Friston KJ, Schofield TM, Leff AP (2010) Comparing families of dynamic causal models. PLoS Comput Biol 6:e1000709. doi:10.1371/journal.pcbi.1000709 pmid:20300649
    OpenUrlCrossRefPubMed
  59. ↵
    Perlman G, Simmons AN, Wu J, Hahn KS, Tapert SF, Max JE, Paulus MP, Brown GG, Frank GK, Campbell-Sills L (2012) Amygdala response and functional connectivity during emotion regulation: a study of 14 depressed adolescents. J Affect Disord 139:75–84. doi:10.1016/j.jad.2012.01.044 pmid:22401827
    OpenUrlCrossRefPubMed
  60. ↵
    Pessoa L (2017) A network model of the emotional brain. Trends Cogn Sci 21:357–371. doi:10.1016/j.tics.2017.03.002 pmid:28363681
    OpenUrlCrossRefPubMed
  61. ↵
    Pessoa L, Adolphs R (2010) Emotion processing and the amygdala: from a ‘low road’ to ‘many roads’ of evaluating biological significance. Nat Rev Neurosci 11:773. doi:10.1038/nrn2920 pmid:20959860
    OpenUrlCrossRefPubMed
  62. ↵
    Phelps EA (2006) Emotion and cognition: insights from studies of the human amygdala. Annu Rev Psychol 57:27–53. doi:10.1146/annurev.psych.56.091103.070234 pmid:16318588
    OpenUrlCrossRefPubMed
  63. ↵
    Phillips ML, Ladouceur CD, Drevets WC (2008) A neural model of voluntary and automatic emotion regulation: implications for understanding the pathophysiology and neurodevelopment of bipolar disorder. Mol Psychiatry 13:829, 833–857. doi:10.1038/mp.2008.82
    OpenUrlCrossRef
  64. ↵
    Power JD, Barnes KA, Snyder AZ, Schlaggar BL, Petersen SE (2012) Spurious but systematic correlations in functional connectivity MRI networks arise from subject motion. Neuroimage 59:2142–2154. doi:10.1016/j.neuroimage.2011.10.018 pmid:22019881
    OpenUrlCrossRefPubMed
  65. ↵
    Price JL (2005) Free will versus survival: brain systems that underlie intrinsic constraints on behavior. J Comp Neurol 493:132–139. doi:10.1002/cne.20750 pmid:16255003
    OpenUrlCrossRefPubMed
  66. ↵
    Puce A, Allison T, Asgari M, Gore JC, McCarthy G (1996) Differential sensitivity of human visual cortex to faces, letterstrings, and textures: a functional magnetic resonance imaging study. J Neurosci 16:5205–5215. pmid:8756449
    OpenUrlAbstract/FREE Full Text
  67. ↵
    Quirk GJ, Beer JS (2006) Prefrontal involvement in the regulation of emotion: convergence of rat and human studies. Curr Opin Neurobiol 16:723–727. doi:10.1016/j.conb.2006.07.004 pmid:17084617
    OpenUrlCrossRefPubMed
  68. ↵
    Ray RD, Zald DH (2012) Anatomical insights into the interaction of emotion and cognition in the prefrontal cortex. Neurosci Biobehav Rev 36:479–501. doi:10.1016/j.neubiorev.2011.08.005 pmid:21889953
    OpenUrlCrossRefPubMed
  69. ↵
    Remmington NA, Fabrigar LR, Visser PS (2000) Reexamining the circumplex model of affect. J Pers Soc Psychol 79:286. pmid:10948981
    OpenUrlCrossRefPubMed
  70. ↵
    Roy M, Shohamy D, Wager TD (2012) Ventromedial prefrontal-subcortical systems and the generation of affective meaning. Trends Cogn Sci 16:147–156. doi:10.1016/j.tics.2012.01.005 pmid:22310704
    OpenUrlCrossRefPubMed
  71. ↵
    Sander D, Scherer K (2014) Surprise. In: Oxford companion to emotion and the affective sciences. New York, NY: Oxford University Press.
  72. ↵
    Sander D, Grafman J, Zalla T (2003) The human amygdala: an evolved system for relevance detection. Rev Neurosci 14:303–316. pmid:14640318
    OpenUrlCrossRefPubMed
  73. ↵
    Silvers JA, Shu J, Hubbard AD, Weber J, Ochsner KN (2015) Concurrent and lasting effects of emotion regulation on amygdala response in adolescence and young adulthood. Dev Sci 18:771–784. doi:10.1111/desc.12260
    OpenUrlCrossRefPubMed
  74. ↵
    Sladky R, Höflich A, Küblböck M, Kraus C, Baldinger P, Moser E, Lanzenberger R, Windischberger C (2015a) Disrupted effective connectivity between the amygdala and orbitofrontal cortex in social anxiety disorder during emotion discrimination revealed by dynamic causal modeling for FMRI. Cereb Cortex 25:895–903. doi:10.1093/cercor/bht279
    OpenUrlCrossRefPubMed
  75. ↵
    Sladky R, Spies M, Hoffmann A, Kranz G, Hummer A, Gryglewski G, Lanzenberger R, Windischberger C, Kasper S (2015b) (S)-citalopram influences amygdala modulation in healthy subjects: a randomized placebo-controlled double-blind fMRI study using dynamic causal modeling. Neuroimage 108:243–250. doi:10.1016/j.neuroimage.2014.12.044
    OpenUrlCrossRef
  76. ↵
    Smith DV, Hayden BY, Truong T-K, Song AW, Platt ML, Huettel SA (2010) Distinct value signals in anterior and posterior ventromedial prefrontal cortex. J Neurosci 30:2490–2495. doi:10.1523/JNEUROSCI.3319-09.2010 pmid:20164333
    OpenUrlAbstract/FREE Full Text
  77. ↵
    Soriano Salinas CM, Fontaine JR, Scherer KR (2015) Surprise in the GRID. Rev Cogn Ling 13:436–460.
    OpenUrl
  78. ↵
    Stephan KE, Penny WD, Moran RJ, den Ouden HE, Daunizeau J, Friston KJ (2010) Ten simple rules for dynamic causal modeling. Neuroimage 49:3099–3109. doi:10.1016/j.neuroimage.2009.11.015 pmid:19914382
    OpenUrlCrossRefPubMed
  79. ↵
    Trautmann SA, Fehr T, Herrmann M (2009) Emotions in motion: dynamic compared to static facial expressions of disgust and happiness reveal more widespread emotion-specific activations. Brain Res 1284:100–115. doi:10.1016/j.brainres.2009.05.075 pmid:19501062
    OpenUrlCrossRefPubMed
  80. ↵
    Urry HL, Van Reekum CM, Johnstone T, Kalin NH, Thurow ME, Schaefer HS, Jackson CA, Frye CJ, Greischar LL, Alexander AL (2006) Amygdala and ventromedial prefrontal cortex are inversely coupled during regulation of negative affect and predict the diurnal pattern of cortisol secretion among older adults. J Neurosci 26:4415–4425. doi:10.1523/JNEUROSCI.3215-05.2006 pmid:16624961
    OpenUrlAbstract/FREE Full Text
  81. ↵
    Vuilleumier P, Richardson MP, Armony JL, Driver J, Dolan RJ (2004) Distant influences of amygdala lesion on visual cortical activation during emotional face processing. Nat Neurosci 7:1271. doi:10.1038/nn1341 pmid:15494727
    OpenUrlCrossRefPubMed
  82. ↵
    Wagner DD, Heatherton TF (2013) Self-regulatory depletion increases emotional reactivity in the amygdala. Soc Cogn Affect Neurosci 8:410–417. doi:10.1093/scan/nss082 pmid:22842815
    OpenUrlCrossRefPubMed
  83. ↵
    Whalen PJ (1998) Fear, vigilance, and ambiguity: initial neuroimaging studies of the human amygdala. Curr Dir Psychol Sci 7:177–188. doi:10.1111/1467-8721.ep10836912
    OpenUrlCrossRef
  84. ↵
    Whalen PJ, Raila H, Bennett R, Mattek A, Brown A, Taylor J, van Tieghem M, Tanner A, Miner M, Palmer A (2013) Neuroscience and facial expressions of emotion: the role of amygdala–prefrontal interactions. Emot Rev 5:78–83. doi:10.1177/1754073912457231
    OpenUrlCrossRef
  85. ↵
    Winecoff A, Clithero JA, Carter RM, Bergman SR, Wang L, Huettel SA (2013) Ventromedial prefrontal cortex encodes emotional value. J Neurosci 33:11032–11039. doi:10.1523/JNEUROSCI.4317-12.2013 pmid:23825408
    OpenUrlAbstract/FREE Full Text
  86. ↵
    Zinchenko O, Yaple ZA, Arsalidou M (2018) Brain responses to dynamic facial expressions: a normative meta-analysis. Front Hum Neurosci 12:227.
    OpenUrl

Synthesis

Reviewing Editor: Daniela Schiller, Mount Sinai School of Medicine

Decisions are customarily a result of the Reviewing Editor and the peer reviewers coming together and discussing their recommendations until a consensus is reached. When revisions are invited, a fact-based synthesis statement explaining their decision and outlining what is needed to prepare a revision will be listed below. The following reviewer(s) agreed to reveal their identity: Dylan Gee, Xiaosi Gu.

This study used dynamic causal modeling of fMRI data collected during a face emotion task in healthy controls to examine prefrontal-amygdala circuitry and its modulation by emotional valence in healthy adults. The authors report that bidirectional prefrontal-amygdala coupling was increased by emotional valence, and that negative (but not positive) valence enhanced bottom-up connectivity from the amygdala to the mPFC. Strengths include the use of dynamic stimuli and the study conceptualization, which is well-grounded in the current literature. These findings are interesting and provide amongst the first insights into the directional relationship between amygdala and prefrontal regions under the modulation of emotion. Ultimately, these findings would be important for interpreting the emotional deficits found in many psychiatric disorders. If the methodological limitations raised can be addressed, the results of this study provide important information about the nature of prefrontal-amygdala interactions (e.g., the role of mPFC, modulation by valence, etc.) that have not been addressed in previous research.

Specific comments:

1) Motion is a known confound in studies of connectivity (e.g., Power et al., 2012), including task-based connectivity (e.g., Siegel et al., 2014). There does not seem to be any mention of criteria employed for exclusion based on motion in the current manuscript. How did the authors control for motion? Were any volumes or participants excluded due to excessive motion? Even if motion was very low, it should be characterized (e.g., mean motion in each direction, mean framewise displacement, etc.) so that the reader has a clearer picture of motion in this sample. Without this information, it is not possible to evaluate the current methods or results.

2) Given that the primary focus of the study is on valence, the operationalization of valence is particularly important. It is unclear how the authors selected which emotional faces to include in which blocks, and the results must be interpreted more specifically in the context of the faces that were included. There are two primary concerns- one is that surprised faces were included in both the positive and neutral blocks, and not in the negative blocks, while existing research shows strong evidence that the valence of surprised faces is largely dependent on individual differences (e.g., Neta et al., 2009; Neta et al., 2011; Neta et al., 2017). The other is that the negative valence block consists only of sad and disgusted faces, whereas the majority of extant literature (particularly in anxiety, which the authors reference) is based on angry and fearful faces.

3) The sample size (N=33) is relatively small and seems quite arbitrary. What was the target sample size? What was the authors' stopping rule?

4) How were psychiatric disorders evaluated for exclusion?

5) Were individual connectivity parameters associated with activation? One might expect stronger top-down connectivity to be associated with lower amygdala activation, and this dataset has the potential to inform this open question.

6) The whole brain GLM is not clearly described. How many actual events were included as regressors? In order to model driving and modulatory inputs, each input should be included as a separate event regressor (e.g. one for ‘all stimuli’, one for ‘all faces’, one for ‘all negative faces’ etc.)

7) The behavioral results are interesting and deserve being in the figure (e.g. as fig 1b).

8) The discussion on pg. 18 of parallels between the current findings and those in patients with anxiety is highly speculative - this should be emphasized given the absence of any associations with anxiety in the current sample. Alternatively, do the authors have a measure of trait anxiety that they could examine in this sample?

9) Line 191: “... the behavior of specific brain regions” sounds confusing - do the authors mean the activities of specific brain regions?

10) DCM analysis:

- The study could examine intrinsic connection between amygdala and LPFC - It does not require direct anatomical connections.

- Driving input - do the authors mean the driving input is all events (face and shape combined)? It is possible that using a visual ROI as the entrance of all visual stimuli would be more appropriate and/or change the results. Have the authors considered such possibility?

- To consider neutral as a valence seems counterintuitive. A more direct and comprehensive way to examine the impact of valence might be to construct DCMs with positive (against neutral) and negative (against neutral) as two modulatory inputs, while using both all visual stimuli and all faces as two driving inputs entering at the level of visual ROIs (e.g. V1 and FFA, respectively; resulting a total of 5 ROIs in the models). A slightly simpler version of this 5-region DCM would be to focus on face stimuli (while still including shape trials in the GLM): driving input - all faces, entering FFA; modulatory input 1 - negative faces; modulatory input 2 - positive faces, etc.

- The DCM result of equivalent impact of negative and neutral faces on AMY-MPFC connectivity is difficult to interpret and does not fit into the literature of the distinct neural effects of negatively valenced stimuli in general. It is unclear whether these results would still hold if the authors revise the DCM to include proper visual ROIs for driving inputs.

- Did the authors measure any traits related to mood or other disorders (e.g. BDI, anxiety etc.) and if so, are there any correlations between these measures with the DCM estimates?

Back to top

In this issue

eneuro: 6 (4)
eNeuro
Vol. 6, Issue 4
July/August 2019
  • Table of Contents
  • Index by author
  • Ed Board (PDF)
Email

Thank you for sharing this eNeuro article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
Valence-Dependent Coupling of Prefrontal-Amygdala Effective Connectivity during Facial Affect Processing
(Your Name) has forwarded a page to you from eNeuro
(Your Name) thought you would be interested in this article in eNeuro.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Print
View Full Page PDF
Citation Tools
Valence-Dependent Coupling of Prefrontal-Amygdala Effective Connectivity during Facial Affect Processing
David Willinger, Iliana I. Karipidis, Selina Beltrani, Sarah V. Di Pietro, Ronald Sladky, Susanne Walitza, Philipp Stämpfli, Silvia Brem
eNeuro 9 July 2019, 6 (4) ENEURO.0079-19.2019; DOI: 10.1523/ENEURO.0079-19.2019

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article
Share
Valence-Dependent Coupling of Prefrontal-Amygdala Effective Connectivity during Facial Affect Processing
David Willinger, Iliana I. Karipidis, Selina Beltrani, Sarah V. Di Pietro, Ronald Sladky, Susanne Walitza, Philipp Stämpfli, Silvia Brem
eNeuro 9 July 2019, 6 (4) ENEURO.0079-19.2019; DOI: 10.1523/ENEURO.0079-19.2019
Reddit logo Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • Abstract
    • Significance Statement
    • Introduction
    • Materials and Methods
    • Results
    • Discussion
    • Acknowledgments
    • Footnotes
    • References
    • Synthesis
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF

Keywords

  • amygdala
  • DCM
  • emotional valence
  • fMRI
  • prefrontal cortex

Responses to this article

Respond to this article

Jump to comment:

No eLetters have been published for this article.

Related Articles

Cited By...

More in this TOC Section

New Research

  • LIQ HD (Lick Instance Quantifier Home cage Device): An open-source tool for recording undisturbed two-bottle drinking behavior in a home cage environment
  • Insulin-like growth factor-1 supplementation promotes brain maturation in preterm pigs
  • SK and Kv4 channels limit spike timing perturbations in pacemaking dopamine neurons
Show more New Research

Cognition and Behavior

  • Deciding while acting - Mid-movement decisions are more strongly affected by action probability than reward amount
  • Environment Enrichment Facilitates Long-Term Memory Consolidation Through Behavioral Tagging
  • Effects of Cortical FoxP1 Knockdowns on Learned Song Preference in Female Zebra Finches
Show more Cognition and Behavior

Subjects

  • Cognition and Behavior

  • Home
  • Alerts
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Latest Articles
  • Issue Archive
  • Blog
  • Browse by Topic

Information

  • For Authors
  • For the Media

About

  • About the Journal
  • Editorial Board
  • Privacy Policy
  • Contact
  • Feedback
(eNeuro logo)
(SfN logo)

Copyright © 2023 by the Society for Neuroscience.
eNeuro eISSN: 2373-2822

The ideas and opinions expressed in eNeuro do not necessarily reflect those of SfN or the eNeuro Editorial Board. Publication of an advertisement or other product mention in eNeuro should not be construed as an endorsement of the manufacturer’s claims. SfN does not assume any responsibility for any injury and/or damage to persons or property arising from or related to any use of any material contained in eNeuro.