Skip to main content

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Blog
    • Collections
    • Podcast
  • TOPICS
    • Cognition and Behavior
    • Development
    • Disorders of the Nervous System
    • History, Teaching and Public Awareness
    • Integrative Systems
    • Neuronal Excitability
    • Novel Tools and Methods
    • Sensory and Motor Systems
  • ALERTS
  • FOR AUTHORS
  • ABOUT
    • Overview
    • Editorial Board
    • For the Media
    • Privacy Policy
    • Contact Us
    • Feedback
  • SUBMIT

User menu

Search

  • Advanced search
eNeuro

eNeuro

Advanced Search

 

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Blog
    • Collections
    • Podcast
  • TOPICS
    • Cognition and Behavior
    • Development
    • Disorders of the Nervous System
    • History, Teaching and Public Awareness
    • Integrative Systems
    • Neuronal Excitability
    • Novel Tools and Methods
    • Sensory and Motor Systems
  • ALERTS
  • FOR AUTHORS
  • ABOUT
    • Overview
    • Editorial Board
    • For the Media
    • Privacy Policy
    • Contact Us
    • Feedback
  • SUBMIT
PreviousNext
Research ArticleNew Research, Sensory and Motor Systems

The Neural Dynamics of Facial Identity Processing: Insights from EEG-Based Pattern Analysis and Image Reconstruction

Dan Nemrodov, Matthias Niemeier, Ashutosh Patel and Adrian Nestor
eNeuro 29 January 2018, 5 (1) ENEURO.0358-17.2018; DOI: https://doi.org/10.1523/ENEURO.0358-17.2018
Dan Nemrodov
1Department of Psychology, University of Toronto Scarborough, 1265 Military Trail, Toronto, Ontario M1C1A4, Canada
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Dan Nemrodov
Matthias Niemeier
1Department of Psychology, University of Toronto Scarborough, 1265 Military Trail, Toronto, Ontario M1C1A4, Canada
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Ashutosh Patel
1Department of Psychology, University of Toronto Scarborough, 1265 Military Trail, Toronto, Ontario M1C1A4, Canada
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Adrian Nestor
1Department of Psychology, University of Toronto Scarborough, 1265 Military Trail, Toronto, Ontario M1C1A4, Canada
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • Article
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF
Loading

Abstract

Uncovering the neural dynamics of facial identity processing along with its representational basis outlines a major endeavor in the study of visual processing. To this end, here, we record human electroencephalography (EEG) data associated with viewing face stimuli; then, we exploit spatiotemporal EEG information to determine the neural correlates of facial identity representations and to reconstruct the appearance of the corresponding stimuli. Our findings indicate that multiple temporal intervals support: facial identity classification, face space estimation, visual feature extraction and image reconstruction. In particular, we note that both classification and reconstruction accuracy peak in the proximity of the N170 component. Further, aggregate data from a larger interval (50–650 ms after stimulus onset) support robust reconstruction results, consistent with the availability of distinct visual information over time. Thus, theoretically, our findings shed light on the time course of face processing while, methodologically, they demonstrate the feasibility of EEG-based image reconstruction.

  • ERP
  • face space
  • image reconstruction
  • N170
  • pattern analysis
  • spatiotemporal dynamics

Significance Statement

Identifying a face is achieved through fast and efficient processing of visual information. Here, we investigate the nature of this information, its specific content and its availability at a fine-grained temporal scale. Notably, we provide a way to extract, to assess and to visualize such information from neural data associated with individual face processing. Thus, the present work accounts for the time course of face individuation through appeal to its underlying visual representations while, also, it provides a first demonstration regarding the ability to reconstruct the appearance of stimulus images from electroencephalography (EEG) data.

Introduction

Elucidating the dynamics of visual face processing is essential to understanding its underlying mechanisms. To this end, considerable efforts have been devoted to characterizing the time course of face processing especially through the use of electroencephalography (EEG) and magnetoencephalography (MEG) given the temporal resolution of these methods. Accordingly, much is known about the temporal profile of face processing as reflected by either traditional event-related potentials (ERPs; Bentin and Deouell, 2000; Itier and Taylor, 2002; Huddy et al., 2003; Tanaka et al., 2006; Rossion and Caharel, 2011; Zheng et al., 2012), or by spatiotemporal patterns (Liu et al., 2009; Cichy et al., 2014; Vida et al., 2017). Comparatively less is known about the visual representations underlying the dynamics of face processing, especially as related to facial identity. To shed light on this issue, the present work employs an image-reconstruction paradigm (Miyawaki et al., 2008; Naselaris et al., 2009; Nishimoto et al., 2011; Cowen et al., 2014; Chang and Tsao, 2017) seeking to approximate the visual appearance of individual faces from spatiotemporal EEG patterns. Concretely, this work aims to answer whether the visual information involved in face identification can be recovered from EEG signals and, further, whether such information can support the characterization of neural-based face space along with the reconstruction of individual face images.

Recent applications of pattern analysis have focused on the temporal profile of face discrimination at the category (Carlson et al., 2013; Van de Nieuwenhuijzen et al., 2013; Cauchoix et al., 2014; Cichy et al., 2014; Kaneshiro et al., 2015) and the exemplar level (Davidesco et al., 2014; Ghuman, et al., 2014). For instance, expression-invariant identity discrimination has been conducted using MEG (Vida et al., 2017), electrocorticography (Ghuman et al., 2014) and EEG (Nemrodov et al., 2016). These studies found multiple, distinct temporal windows sensitive to facial information and, consistent with results from monkey neurophysiology (Hung et al., 2005; Freiwald et al., 2009) and human psychophysics (Lehky, 2000; Tanaka and Curran, 2001; Crouzet et al., 2010), have estimated an early onset for such sensitivity (Ghuman et al., 2014; Nemrodov et al., 2016). Further, the cortical source of this information was attributed primarily to the fusiform gyrus (FG) in line with homologous investigations of face identification using functional magnetic resonance imaging (fMRI; Nestor et al., 2011; Goesaert and Op de Beeck, 2013; Anzellotti et al., 2014).

Yet, the representational basis of facial identity that allows successful discrimination from neural data remains to be elucidated. Arguably, neural patterns elicited by face perception speak to the properties of a representational face space (Valentine, 1991), or, more generally, of a representational similarity space (Kriegeskorte et al., 2008). In an effort to clarify the nature of such representations, recent fMRI work (Nestor et al., 2016) has combined the study of face space and neural-based image reconstruction. Specifically, this work has derived visual features from the structure of FG-based face space and, then, used such features for facial image reconstruction. However, this work did not consider the temporal aspects of face processing, neither did it assess the invariant structure of a facial identity space, surprisingly, we note that face space invariance over common image transformations (e.g., viewpoint, expression) has rarely been explicitly investigated (but see Newell et al., 1999; Blank and Yovel, 2011).

To address the issues above, the current work aims to derive face space constructs from the EEG signal associated with consecutive time windows separately for different facial expressions (i.e., neutral and happy). Then, it reconstructs the appearance of one set of faces (e.g., happy) based on the structure of the face space derived for the other faces (e.g., neutral). This demonstration provides evidence that the spatiotemporal information of EEG patterns is rich enough to support: (1) identity-level face discrimination; (2) neural-based face space estimation; (3) visual feature synthesis; and (4) facial image reconstruction. Further, this work characterizes the neural dynamics of expression-invariant face processing while, more generally, it provides proof of concept for the possibility of EEG-based image reconstruction.

Materials and Methods

Participants

Thirteen healthy adults (six males, seven females; age range: 18–27 years) with normal or corrected-to-normal vision were recruited from the University of Toronto community to participate in the study in exchange for monetary compensation. All participants provided informed consent and all experimental procedures were approved by the Research Ethics Board at the University of Toronto. The data of all participants were included in the analyses reported below.

Stimuli

A total of 140 face images of 70 individuals, each displaying a neutral and a happy expression were used as experimental stimuli. Out of these, 108 images of 54 unfamiliar males were selected from three databases: AR (Martinez and Benavente, 1998), Radboud (Langner et al., 2010), and FEI (Thomaz and Giraldi, 2010). The remaining 32 images displayed faces of six famous male and 10 female individuals selected from open access sources. To be clear, unfamiliar male face stimuli are the focus of the present investigation while female faces were used as go trials in a go/no-go gender recognition task (see Experimental design) and additional famous male faces were included to promote alertness (however, no results are reported for them below due to the smaller stimulus set that precluded a separate examination of famous face recognition).

All images featured young adult white individuals with frontal view, gaze and illumination. The stimuli were selected so that no facial accessories, hair or makeup obscured the internal features of the face and so that all happy expressions displayed an open-mouth smile. These images were: (1) scaled uniformly and aligned with roughly the same position of the eyes and the nose; (2) cropped to eliminate background; (3) normalized with the same mean and root mean square (RMS) contrast values separately for each color channel in CIEL*a*b* color space; and (4) reduced to the same size (95 × 64 pixels).

Experimental procedure

Before EEG testing participants were administered the Cambridge Face Memory Test (CFMT), (Duchaine and Nakayama, 2006) to confirm that their face processing abilities fall within the range of normal performance for young adults (Bowles et al., 2009). Participants also completed the Vividness of Visual Imagery Questionnaire 2 (VVIQ-2; Marks, 1995) along with a custom familiarity-rating famous face questionnaire.

During EEG sessions participants were seated in a dimly lit room at a viewing distance of 80 cm from an LCD monitor (resolution: 1920 × 1080, refresh rate: 60 Hz). The participants were instructed to perform a go/no-go gender recognition task by pressing a designated key every time they saw a female face, irrespective of expression. The experiment consisted of 32 blocks of stimulus presentation divided into two sessions conducted on different days. In each session, experimental blocks were preceded by one training block, subsequently discarded from all analyses. The blocks were separated by self-paced breaks.

Over the course of any given block, each image of a male face was presented twice and each image of a female face was presented once, for a total of 260 trials. Images were presented in a pseudorandom order under the constraint that no facial identity would appear consecutively. Each stimulus was presented in the center of the screen against a black background and subtended a visual angle of 3.2 × 4.9. A stimulus display lasted for 300 ms, and it was followed by a fixation cross for a duration ranging randomly between 925 and 1075 ms. Each session, including participant and equipment setup, lasted around 2.5 h. Stimulus presentation and response recording relied on Matlab 9.0 (Mathworks) and Psychtoolbox 3.0.8 (Brainard, 1997; Pelli, 1997).

EEG acquisition and preprocessing

EEG data were recorded using a 64-electrode Biosemi ActiveTwo EEG recording system (Biosemi B.V.). The electrodes were arranged according to the International 10/20 System. The electrode offset was kept below 40 mV. The EEG and EOG were low-pass filtered using a fifth order sinc filter with a half-power cutoff at 204.8 Hz and then digitized at 512 Hz with 24 bits of resolution. All data were digitally filtered offline (zero-phase 24 dB/octave Butterworth filter) with a bandpass of 0.1–40 Hz.

Next, data were separated into epochs, from 100 ms before stimulus presentation until 900 ms later, and baseline corrected. Epochs corresponding to go trials (i.e., female face stimuli) and epochs containing false alarms were discarded from further analysis. Further, noisy electrodes were interpolated if necessary (no more than two electrodes per subject) and epochs were rereferenced to the average reference. In addition, before univariate ERP analysis, data were cleaned of ocular artifacts using Infomax ICA (Delorme et al., 2007).

After removing trials containing artifacts and/or false alarms, an average of 96% of trials (range: 75–100% across participants) were selected for further analysis. In particular, we note that relatively few trials contained false alarms as participants performed the go/no-go recognition task at ceiling (accuracy range: 98.1–100% across participants).

All analyses were conducted using Letswave 6 (http://nocions.webnode.com/letswave; Mouraux and Iannetti, 2008), MATLAB 9.0 and the G*Power toolbox (Faul, et al., 2007).

Univariate ERP analyses

Twelve electrodes situated over homolog occipitotemporal areas (P5, P7, P9, PO3, PO7, and O1 on the left and P6, P8, P10, PO4, PO8, and O2 on the right) were selected for ERP analysis. Their selection was motivated by the relevance of these electrodes for face processing as revealed by ERP analysis (e.g., robust N170 amplitudes). Data from electrodes over the left and right sites were then averaged separately creating two sets of signals, one for each hemisphere.

For the purpose of univariate tests, the data were averaged for each unfamiliar face identity across expressions. Then, P1, N170 and N250 components were visually identified on a grand-average plot and a two-way repeated measures ANOVA over facial identities and hemispheres was conducted on maximum amplitudes in the 70–180 ms range for P1, and on minimum amplitudes in the 160–250 and 260–350 ms ranges for N170 and N250, respectively. Greenhouse–Geisser correction was applied in case of violation of the sphericity assumption.

Pattern classification analysis

Epochs were linearly detrended, z-scored across time and electrodes, and corrected for outliers (i.e., values exceeding 3SD from the mean were thresholded at ±3 to minimize the deleterious effect of extreme values on SVM-based pattern classification). Then, all epochs were normalized to the same range (0–1) and mean (i.e., 0.5). To boost the signal-to-noise ratio (SNR) of spatiotemporal patterns (Grootswagers et al., 2017) multiple epochs pertaining to the same condition were averaged into ERP traces. Specifically, all epochs corresponding to the same image stimulus across two consecutive blocks, for a maximum of four epochs, were averaged together resulting in 16 separate traces per stimulus. Further, this procedure was also instrumental in handling missing data following trial removal and in balancing the number of observations for pattern classification. Specifically, since it is possible that both trials associated with a given stimulus in a given block be removed (e.g., due to artifacts), averaging data across single blocks can lead to different numbers of observations for different stimuli. However, averaging data across pairs of consecutive blocks (i.e., one to four trials per stimulus following data removal) ensured that equal numbers of observations can be constructed for each pair of stimuli undergoing pattern classification.

Next, to increase the robustness of pattern analyses epochs were divided into temporal windows containing five consecutive bins (5 bins × 1.95 ms ≈ 10 ms; Blankertz et al., 2011). For each window, data were concatenated into observations, for instance, data across selected occipitotemporal electrodes (see univariate ERP analyses) were concatenated into 60-dimension observations (5 time bins × 12 electrodes). These observations were constructed for the purpose of pattern analyses in time, window by window. In addition, we constructed more inclusive observations that contain all time bins between 50 and 650 ms after stimulus presentation (3684-dimension vectors: 307 bins × 12 electrodes), and, thus, both early and late information relevant for face processing (Ghuman et al., 2014; Vida et al., 2017). These higher-dimensional observations were constructed for the purpose of temporally cumulative analyses able to exploit more extensive information over time.

Pairwise discrimination of facial identity was conducted with the aid of linear support vector machine (SVM) classification (c = 1; SVMLIB 3.22; Chang and Lin, 2011) and leave-one-out cross-validation (i.e., one out of 16 pairs of observations was systematically left out for testing while the remaining 15 were used for training). Concretely, 1431 pairs of stimuli were classified across 16 leave-one-out cross-validation iterations and then classification accuracy was obtained for each pair by averaging across these iterations. Classification was conducted for all pairs of facial identities in two ways: (1) within expression, the classifier was trained separately on one expression, happy or neutral, and tested on the same expression; and (2) across expression, the classifier was trained on one expression and tested on the other. Significance of classification accuracy was assessed via a permutation test, by randomly shuffling identity labels 103 times, and correcting for multiple comparisons across temporal windows using the false discovery rate (FDR).

The analyses above were conducted for data averaged across all participants (i.e., observations were constructed from averaged ERP traces). In addition, similar discrimination analyses were performed for data from individual participants. The significance of the overall classification accuracy for this analysis was assessed using one-sample two-tailed t tests against chance across participants separately for within- and across-expression discrimination.

Further, given that the three face databases used here for the purpose of stimulus selection and design sample different populations (i.e., individuals of different nationalities) we anticipated a possible effect of database. However, it is important to establish that any effects found here are not solely due to such differences. To assess this possibility, we also examined pairwise classification results for faces extracted from the same databases and, separately, for faces extracted from different databases, i.e., each face is classified only relative to faces from the same database, for a total of 472 pairs, or only relative to faces from different databases, for a total of 1918 pairs.

Relationship with behavioral performance

Average pairwise face discrimination, as indexed by temporally cumulative analysis, was correlated with behavioral markers of performance (from CFMT and VVIQ-2 tests) for each participant.

For a finer-grained analysis in the temporal domain, pairwise face discrimination was correlated, window by window, with behavioral data from a previous study (Nestor et al., 2013, experiment 1) in which participants rated the visual similarity of pairs of faces. Specifically, in this study participants viewed a larger set of faces, including the 108 unfamiliar face images used here, and rated the similarity of each face with every other face on a five-point scale across a change in expression (i.e., one face in a pair displayed a neutral expression and the other a happy expression). Ratings from 22 healthy adult participants were then averaged to deliver an estimate of behavioral face similarity. Here, this estimate was compared to its neural counterpart from across-expression classification conducted on group-averaged ERP traces. Concretely, behavioral and neural-based estimates were related via Pearson’s correlation and tested for significance via a permutation test (103 permutations for each temporal window; FDR correction across windows).

Estimation of EEG-based face space

Face space constructs were derived by applying multidimensional scaling (MDS) to EEG-based estimates of face discrimination. Specifically, classification accuracies for pairs of facial identities were organized into a dissimilarity matrix in which each cell estimates the discriminability of a pair of faces; then, all values were linearly scaled between 0 and 1 and metric MDS was applied to approximate a corresponding space. The dimensionality of such spaces was restricted to 20 since that was found sufficient to account for most variance in the data (e.g., over 90% for temporally cumulative analyses). This procedure was conducted on within-expression estimates of discrimination, separately for each expression, resulting in separate spaces for faces with neutral and happy expressions.

Next, we examined face space invariance to image changes introduced by emotional expression. To this end, the fit between neutral and happy face spaces was estimated by aligning one space to the other, via Procrustes transformation, and measuring the badness of fit as the sum of squared errors (SSEs) between the two spaces. Significance testing was then conducted through a permutation test for multidimensional spaces (Jackson, 1995): the labels of each point in one of the two spaces was randomly shuffled and the resulting space was fit to the intact one as above. This procedure was conducted for a total of 103 permutations, by leaving intact each of the two spaces half of the time while permuting the other space, and permutation-based SSE estimates were computed each time.

Reconstruction approach

The current procedure broadly follows a recently developed approach to facial image reconstruction designed to exploit spatial information in fMRI patterns (Nestor et al., 2016). This procedure capitalizes on the structure of neural-based face space for the purpose of feature derivation and image reconstruction. Here, we deployed this procedure to capture spatiotemporal information in EEG patterns and, further, to examine the ability of expression-invariant visual information to support image reconstructions of facial identity. This procedure was conducted in a sequence of steps as follows.

First, visual features accounting for face space topography were separately derived for each dimension of EEG-based face space. These features were computed as weighted sums of image stimuli following a strategy similar to reverse correlation/image classification (for review, see Murray, 2011; Smith et al., 2012) for applications to EEG data. Hence, here they are referred to as classification images (CIMs). Briefly, face stimuli, following their color conversion to CIEL*a*b*, were summed proportionally to their z-scored coordinates on a given dimension of face space. The resulting CIM (i.e., a triplet of images corresponding to L*, a*, and b* channels) amounts to a linear approximation of the visual information responsible for organizing faces along that specific dimension. Thus, for each expression this procedure delivered a total of 20 different CIMs, one for each corresponding dimension of face space.

Second, since not all dimensions may encode systematic visual information, feature/dimension selection was used to identify subspaces relevant for reconstruction purposes. To this end, CIMs corresponding to each dimension were assessed regarding the inclusion of significant information. Specifically, for each dimension, permutation-based CIMs were generated after randomly shuffling the coefficients associated with stimulus images. Then, pixel intensities in the true CIM were compared to the corresponding intensities of pixels in permutation-based CIMs 103 permutations; FDR correction across pixels and color channels) and only CIMs that contained pixel values significantly different from chance were considered for reconstruction purposes.

Third, the coordinates of a target face were estimated in an expression-specific face space. To be clear, the estimation of face space and its CIMs were conducted using all facial identities but one. Then, the left-out face was projected in this space based on its similarity with the other faces. Thus, the procedure ensured that features were not generated from the reconstruction target guarding against circularity.

Last, the target face was constructed through a linear combination of significant CIMs. That is, relevant CIMs, as identified through feature selection above, were summed proportionally with the target’s coordinates in face space and, then, added to an average face obtained from all remaining non-target faces and playing the role of a face prior. However, in contrast to previous work (Nestor et al., 2016), the current procedure capitalized on face space invariance for reconstruction purposes. Specifically, a happy version of face space was aligned to its neutral counterpart via Procrustes transformation using all but one face; then, the left-out face from the happy version of face space was projected into the neutral space using the parameters of the Procrustes mapping function found above. The resulting coordinates in neutral face space were next used to reconstruct the appearance of the target face, with a neutral expression, from neutral CIMs. Conversely, a happy version of the target face relied on aligning neutral face space to its happy counterpart.

Thus, reconstruction relied here on the presence of a robust visual-based structure shared by different face space estimates across expressions. More clearly, in the absence of a common space topography the target face would be projected to a non-informative location of face space and its subsequent reconstruction would resemble the target stimulus no better than expected at chance level (per the evaluation described below).

The procedure above, including face space estimation, was conducted for separate time windows of the ERP trace as well as for temporally cumulative data.

Evaluation of reconstruction results

Image reconstructions were compared to their target stimuli via two different methods. First, image-based accuracy was estimated as the percentage of instances for which a reconstructed image in CIEL*a*b* was more similar to its target, by a pixel-wise L2 metric, than to any other stimulus with the same expression. Average reconstruction accuracy was then compared against permutation-based chance estimates by shuffling reconstruction labels and by recomputing average accuracy across reconstructions each time (for a total of 103 permutations). This procedure was applied to all types of reconstruction (e.g., both window-based and temporally cumulative) separately for neutral and happy faces.

Second, a single set of reconstructions, based on temporally cumulative group-based data, was subjected to experimental evaluation in a separate behavioral test. To this end, 14 new participants (six males and eight females, age range: 20–28 years) were requested to match image reconstructions to their targets in a two-alternative forced choice (2AFC) task. Specifically, each of 108 unfamiliar face reconstructions, including both expressions, was presented in the company of two stimuli, one of which was the actual target and the other was a foil (another face image). Thus, on each trial, a display was shown containing a reconstructed image, at the top, and two stimuli side by side, at the bottom, all of which had the same expression and the same size (as specified in Experimental procedures). Each display was presented until participants made a response to decide which stimulus was more similar to the top image by pressing a designated left/right key. For each participant, any reconstructed image was presented four times in the company of different foils; thus, across participants, all 53 possible foils for a given reconstruction were exhausted. Stimulus order was pseudorandomized so that different reconstructed images appeared on consecutive trials while target stimuli appeared equally often on the left/right side. Each experimental session was completed over the course of 30 min.

Experimental-based estimates of reconstruction accuracy results were measured as the proportion of correct matches across participants and tested for significance tested against chance (50%) using a one-sample two-tailed t test. Last, experimental and homologous image-based estimates of reconstruction accuracy were compared to each other via Pearson’s correlation across images, separately for neutral and happy faces.

Results

Univariate ERP results

Three ERP components relevant for face processing, P1, N170, and N250, were each identified across occipitotemporal electrodes (Fig. 1) and examined by a two-way repeated measures ANOVA (54 unfamiliar face identities × two hemispheres). The analysis of the P1 component a–c (see letter-indexed rows of Table 1 for details of corresponding analyses) and N170d-f analyses found no significant effects. Last, N250 analysesg-i found a main effect of identity (F(53,636) = 3.69, p = 0.001, ηp 2 = 0.235) and a marginally significant interaction (F(53,636) = 1.32, p = 0.07, ηp 2 = 0.099), both hemispheresj,k showed a significant effect of identity (F(53,636) = 2.67, p = 0.009, ηp 2 = 0.182; F(53,636) = 3.214, p = 0.004, ηp 2 = 0.21 for the left and right hemispheres, respectively), but the effect was larger on the right side.

Figure 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 1.

Grand-averaged ERPs across (A) left hemisphere electrodes (P5, P7, P9, PO3, PO7, and O1) and (B) right hemisphere electrodes (P6, P8, P10, PO4, PO8, and O2) for 54 facial identities (averaged across expressions). Head maps show voltage distributions at (A) N170 (B) P1, N250.

View this table:
  • View inline
  • View popup
Table 1.

Statistical table

Pattern classification of facial identity

A total of 108 unfamiliar male faces (54 individuals × two emotional expressions) were classified based on ERP traces associated with their viewing. Specifically, spatiotemporal ERP patterns across bilateral occipitotemporal electrodes were averaged across participants and, then, evaluated for their ability to support facial identity discrimination. To assess the time course of individual face processing, classification was conducted across consecutive 10-ms temporal windows both within and across expression by training and testing the classifier on faces with the same or different expression.

This analysis found significant levels of discrimination across extensive intervalsl (permutation test; q < 0.01). Specifically, across-expression classification evinced above-chance accuracy from 152 ms after stimulus presentation until the end of the epoch, with two peaks at 170 and 295 ms (see Fig. 2A for the group-based results). Within-expression classification yielded a similar time course but consistently higher levels of accuracy and an earlier onset, at 140 ms, in agreement with the reliance on additional, lower-level image cues for this type of discrimination (In addition, an earlier interval of significance was found for across-expression classification between 0-5 ms; however, given its very early occurrence, its reduced amplitude and the absence of its replication by within-expression classification we treat this data point as a false positive). Of note, for both types of classification we found that discrimination accuracy was maximized in the vicinity of the N170 component as identified by univariate analyses of ERP data (Fig. 1).

Figure 2.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 2.

The time course of EEG-based classification accuracy for across- and within-expression discrimination of facial identity. A, Classification was conducted across consecutive 10-ms window patterns over 12 occipitotemporal electrodes for group-based ERP data. Both types of analysis exhibited above-chance discrimination across extensive temporal intervals (permutation test; FDR correction across time, q < 0.01); shaded areas mark intervals of better-than-chance discrimination for across-expression classification. B, The time course of EEG-based classification accuracy for across- and within-expression discrimination of facial identity for a single representative participant. Classification was conducted across consecutive 10-ms window patterns over 12 occipitotemporal electrodes. Both types of analysis exhibited above-chance discrimination across extensive temporal intervals (permutation test; FDR correction across time, q < 0.01); shaded areas mark intervals of better-than-chance discrimination for across-expression classification.

Following the approach described above group-based analyses were complemented next by single-participant analysesm. These analyses confirmed the feasibility of facial identity discrimination from the data of single participants (see Fig. 2B for the results of a representative participant). However, discrimination levels were lower than in the group-based analyses, likely due to the lower SNR of single-participant ERPs and its impact on classification (Grootswagers et al., 2017). Further, multiple intervals of discrimination emerged, as opposed to a single, uninterrupted one.

Next, pattern classification was applied to temporally cumulative data by concatenating data points from all time bins between 50 and 650 ms after stimulus onsetn-q. The aim of this analysis was to maximize discrimination performance by concomitantly exploiting relevant information from all potentially relevant time points. Specifically, while the ∼61-fold increase in pattern dimensionality (i.e., 12 electrodes × 307 time bins) would, by itself, reduce the effectiveness of classification, we considered the possibility that any ensuing classification decrement may be offset by the use of complementary sources of information from different time points (Blankertz et al., 2011).

Consistent with the hypothesis above, we found that this analysis yielded robust levels of discrimination for group-based data (Fig. 3A): 64% and 71% for across- and within-expression discrimination, respectively (permutation test; p = 0.001 for both types of discrimination and both expressions). Of note, these results outperform peak performance obtained with window-based analyses in the proximity of the N170 component. Further, single-participant estimates of discrimination with temporally cumulative data were also computedr-u and, then, averaged across participants (Fig. 3B). Again, performance was better than chance for both within-expression discrimination (two-tailed t test across participants against 50% chance-level discrimination; t(12) = 9.89 and 7.27, Cohen’s d = 2.1 and 2.86 for neutral and happy faces, respectively; p values = 0.001) and for across-expression discrimination (t(12) = 6.84 and 7; d = 2.02 and 1.97 for neutral and happy faces, respectively; p values < 0.001). Further, a two-way repeated measures ANOVAv-x (two discrimination types × two expressions) revealed a main effect of discrimination types (F(1,12) = 50.05, p < 0.001, ηp2 = 0.81), with higher accuracy for within than across-expression discrimination, but no effect of expression and no interaction.

Figure 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 3.

EEG-based classification accuracy for across- and within-expression discrimination of facial identity with temporally cumulative data (50–650 ms after stimulus onset). Accuracy corresponding to neutral and happy faces are separately shown for (A) group-based ERP data and (B) single-participant data (i.e., pattern classification was conducted individually for each participant and, then, its results averaged across participants). The plots display (A) the results of permutation tests (red solid and dash lines indicate average accuracy and 99% confidence intervals estimated with 103 permutations) and (B) the distribution of single-participant data (green and purple solid lines indicate medians, boxes represent 1st and 3rd quartiles, whiskers represent minimum and maximum accuracy values, points represent individual participants’ values and red solid lines indicate chance-level discrimination).

To examine possible effects of face database the analysis above was repeated while restricting pattern classification either to pairs of faces from the same database or from different databases. A two-way repeated measures ANOVAy-ab (two discrimination types: within/across expression × two pair types: within/across database) was conducted to this end – classification estimates were collapsed across neutral and happy faces given the absence of any expression effects above. This analysis revealed a main effect of discrimination types (F(1,12) = 45.92, p < 0.001, ηp2 = 0.79), with higher accuracy for within than across-expression discrimination, as well as a main effect of pair types (F(1,12) = 38.73, p < 0.001, ηp2 = 0.76), with higher accuracy for within than across-database classification. Critically though, all classification estimates were significantly above chance (two-tailed t tests against 50% chance-level discrimination)ac-af; mean accuracy = 56.3%, t(12) = 8.36, p < 0.001, Cohen’s d = 2.32 for within-expression, within-database discrimination; mean accuracy = 58.7%, t(12) = 8.38, p < 0.001, Cohen’s d = 2.32 for within-expression, across-database discrimination; mean accuracy = 53.9%, t(12) = 7.64, p < 0.001, Cohen’s d = 2.12 for across-expression, within-database discrimination; mean accuracy = 55.9%, t(12) = 6.88, p < 0.001, Cohen’s d = 1.91 for across-expression, across-database discrimination).

Last, for completeness, classification analyses including all face pairs within and across databases were repeated with all 64 electrodes, instead of the subset of 12 occipitotemporal electrodes noted above. However, no consistent boost in discrimination was found for any analysis in this case. Hence, for simplicity, the remainder of our analyses and results, as reported below, are based on occipitotemporal electrodes only.

Neural-based discrimination and behavioral performance

To assess and to quantify the relationship between behavioral and neural-based face similarity, group-level estimates of EEG-based discrimination were related to behavioral estimates of pairwise face similarity. Specifically, across-expression discrimination accuracy was compared with its behavioral counterpart, across identity pairs, through Pearson’s correlationag. This comparison revealed, first, a significant correlation for discrimination estimates from the temporally cumulative analysis (r = 0.245; p < 0.001). Then, behavioral estimates were correlated with their neural counterpart separately for each 10-ms temporal window (Fig. 4)ah. This analysis found multiple intervals of significant correlation, between 137 and 197, 236 and 335, and 340 and 385 ms, with peaks at 157 and 266 ms (q < 0.01).

Figure 4.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 4.

Correlation of EEG- and behavioral-based estimates of pairwise face similarity. EEG-based estimates are derived from across-expression discrimination of facial identity for consecutive 10-ms windows of group-based data. Multiple intervals, marked by shaded areas, exhibit significant levels of correlation (permutation test; FDR correction across time, q < 0.01).

Next, we sought to assess whether the different levels of discrimination achieved with different participants were related to their overall face processing skills. To this end, individual estimates of discrimination from the temporally cumulative analysis, averaged across identity pairs, were correlated with CFMTai scores of the same individuals. No significant correlations were found with these scores or with any other behavioral measures considered such as average familiarity ratings with famous faces or VVIQ-2aj scores.

Neural-based face space and expression-invariance

Face space estimates were derived through the application of MDS to within-expression face classification of temporally cumulative data. Specifically, MDS was applied to pairwise face discrimination values derived through pattern analysis of group-based data, separately for neutral and happy faces. Then, the resulting spaces were reduced to the first 20 dimensions and aligned with each other via Procrustes transformation.

An examination of the resulting spaces for neutral and happy faces, following their alignment, seemed consistent with the presence of a common topography across expressions (Fig. 5A). To assess their fit more rigorously a comparison with permutation-based alignment estimates (Fig. 5B)ak was computed next. This comparison indicated that the fit between the two spaces was considerably better than chance (p < 0.001). Beyond its theoretical implications, this finding is relevant here in that it may allow exploiting the structure of visual information invariant across expression for reconstruction purposes, as detailed next.

Figure 5.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 5.

Neutral and happy face space estimates along with their fit (after Procrustes alignment). Estimates were derived through MDS analysis of similarity matrices based on within-expression face discrimination of group-based temporally cumulative data. The two face space estimates exhibit a similar topography as found with (A) their visualization across multiple dimensions (red and green circles indicate neutral and happy faces, respectively; solid lines connect face images with the same identity with the thickness of the line proportionally reflecting shorter distances; the first four dimensions shown here account for 40% and 41% variance for neutral and happy face space); (B) badness of fit (SSEs) for the two spaces compared to their permutation-based counterpart (average fits and 95% confidence intervals estimated with 103 permutations).

Reconstruction results

Visual features were derived from the structure of face space, dimension by dimension, through a procedure akin to reverse correlation/image classification (Gosselin and Schyns, 2003; Sekuler et al., 2004; Martin-Malivel et al., 2006). Such features, or CIMs, were then assessed through a permutation test for the presence of significant information pixel by pixel separately for each CIEL*a*b* color channel.

An examination of CIMs containing significantal-aw information revealed global contrast patterns across multiple color channels, examples of such features derived from group-based temporally cumulative data are shown in Figure 6. The emergence of these features confirmed that neural-based face space is, at least partly, organized by visual information (as opposed, for instance, to higher-level semantic information). More relevantly here, it points to the potential value of CIMs as reconstruction features.

Figure 6.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 6.

Examples of CIMs extracted from EEG-based face space constructs for (A) neutral and (B) happy faces. Pairs of images show raw CIMs (odd columns) and their analysis (even columns) with a pixelwise permutation-based test (FDR-corrected across pixels; q < 0.05). Bright/dark, red/green, and yellow/blue regions in analyzed CIs mark areas of the face brighter (L*), redder (a*), or more yellow (b*) than chance in CIEL*a*b*. Results are shown separately for the first and fourth dimensions of face spaces derived from group-based temporally cumulative data.

Accordingly, significant CIMs were linearly combined to deliver an approximation of face appearance broadly following an image reconstruction approach recently used with fMRI data (Nestor et al., 2016). Specifically, image reconstruction was separately applied to neutral and happy expressions. As noted above, the common face space topography for the two expressions could, in theory, allow using the relative position of a given identity in one space to deliver a reconstruction of the same facial identity with the opposite expression.

Consistent with the hypothesis above, this procedure, as performed for separate time windows with group-based data, found evidence for multiple intervals capable of supporting above-chance reconstruction (image-based permutation test; q < 0.05), reconstruction accuracy was assessed via a pixelwise image-matching test across reconstructed images and stimulus images. The earliest interval had an onset at 160 and 170 ms for neutralax and happy ay faces, respectively, while accuracy peaked at 187 and 180 ms for the two expressions. Examples of reconstructions, converted from CIEL*a*b* back to RGB, are shown in Figure 7A for two different time points while average reconstruction accuracies are plotted in Figure 7B. Further, a representative example of image reconstruction for a single face, interval by interval, is shown in Movie 1 along with the temporal profile of its reconstruction accuracy.

Figure 7.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 7.

Reconstruction results for neutral and happy face images across consecutive 10-ms windows of group-based data. A, Examples of face stimuli along with their corresponding reconstructions at two different times (numbers in the upper left corner indicate image-based estimates of reconstruction accuracy). B, Time course of reconstruction accuracy. Both neutral and happy face images exhibit above-chance discrimination across multiple temporal intervals (permutation test; FDR correction across time, q < 0.05; shaded areas mark intervals of better-than-chance discrimination for neutral faces). Reconstruction accuracy is maximized in the vicinity of the N170 component (Fig. 1) and of the discrimination peak found with pattern classification (Fig. 2).

Movie 1.

Illustration of (top) neutral face stimulus and its reconstruction across 10-ms windows of group-based data along with (bottom) the time course of its reconstruction accuracy assessed with an image-based test. The last frame of the movie shows (top) facial image reconstruction achieved with temporally cumulative data and (B) its corresponding level of accuracy (dashed red line).

The application of the same procedure to temporally cumulative data led to more robust results: 69.46% image-based accuracy for neutral facesaz and 63.91% for happy facesba (permutation test; p values = 0.001). Figure 8A shows examples of reconstructions and Figure 8B displays average accuracy estimates and their permutation-based assessment. Experimental-based estimates of accuracy, obtained with a new group of participants, led to more modest levels of accuracy (Fig. 8C); however, both neutral and happy face reconstructions were still accurate above chance (two-tailed t test across participants against 50% chance-level discrimination: t(13) = 6.70, 4.38; Cohen’s d = 1.86, 1.22, for neutralbb and happybc, respectively; p < 0.001 for all), with no difference between neutral and happy faces. Further, reconstruction accuracies as estimated by the two tests, image-based and experimental-based, were compared with each other across facial identities and were found to significantly correlate with each other (r = 0.43 and 0.42; p = 0.001 and 0.002 for neutralbd and happybe faces, respectively), thus mutually reinforcing their validity.

Figure 8.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 8.

Reconstruction results for neutral and happy faces relying on temporally cumulative group-based data. (A) Examples of face stimuli along with their corresponding reconstructions (numbers in the upper left corner indicate image-based estimates of reconstruction accuracy; numbers in the upper right indicate experimental-based accuracy). B, Average image-based reconstruction accuracy (red solid and dash lines indicate average accuracy and 95% confidence intervals estimated with 103 permutations). C, Average experimental-based reconstruction accuracy (green and purple solid lines indicate medians, boxes represent 1st and 3rd quartiles, whiskers represent minimum and maximum accuracy values, points represent individual participants’ values and red solid lines indicate chance-level reconstruction).

Next, across-expression classification estimates obtained with temporally cumulative data were compared with their corresponding reconstruction accuracies averaged across expressions. Specifically, Pearson’s correlation across facial identities found a positive relationship between across-expression discrimination and image-based accuracybf (r = 0.87, p < 0.001). Thus, the more discriminable a facial identity is, the more accurately it can be reconstructed.

Last, reconstruction was performed for single-participant data and evaluated with the aid of the image-based test. Accuracy levels were still above chance (two-tailed t test across participants against 50% chance-level discrimination; mean accuracy = 53.1%, t(12) = 2.52, p = 0.027, Cohen’s d = 0.73 and mean accuracy = 53.3%, t(12) = 2.24, p = 0.045, Cohen’s d = 0.65 for neutralbg and happybh face reconstructions, respectively) while Pearson’s correlations between classification accuracy and reconstruction accuracy across participants were also found significant (r = 0.83 and r = 0.84, for neutralbi and happybj faces, respectively; p values < 0.001). Thus, participants who provided data supporting higher levels of face classification also provided more accurate reconstruction results.

Discussion

The current work investigates the neural basis of individual face processing and its temporal dynamics through the application of pattern analysis and image reconstruction to EEG data. This investigation yields several notable outcomes as follows.

First, we find that EEG data support facial identity discrimination. By and large, this finding confirms the possibility of EEG-based pattern classification of facial identity across changes in expression from EEG data (Nemrodov et al., 2016). Discrimination peaks were identified in the proximity of the N170 and N250 ERP components, consistent with univariate analyses pointing to the relevance of the former (Itier and Taylor, 2002; Heisz et al., 2006; Jacques et al., 2007; Caharel et al., 2009) and the latter (Schweinberger et al., 2002; Huddy et al., 2003; Tanaka et al., 2006) for face processing. The onset of discrimination, around 150 ms, was intermediary to early estimates in the vicinity of P1 (Nemrodov et al., 2016) and later estimates around 200 ms as reported with MEG (Vida et al., 2017). One possibility is that early estimates, along with higher levels of discrimination, can be triggered by the use of low-level image properties (Cauchoix et al., 2014; Ghuman et al., 2014). In line with this consideration, we found that within versus across-expression discrimination produced earlier and consistently higher levels of discrimination accuracy. Importantly though, across-expression classification, which aims to minimize reliance on low-level cues, exhibited robust levels of discrimination across an extensive interval (i.e., from ∼150 ms onwards) while its time course was also mirrored by that of neural-behavioral correlations in the context of pairwise face similarity.

Second, temporally cumulative analyses targeted identity discrimination across a broad interval between 50 and 650 ms after stimulus onset. Despite the increase in dimensionality for the classification patterns, these data supported even more robust levels of accuracy for both within and across-expression discrimination, consistent with the presence of relevant information at multiple time points. Moreover, the superior levels of discrimination obtained with temporally cumulative data, as opposed to 10-ms windows, agrees with the presence of distinct sources of information at different time points. That is, we relate the boost in classification accuracy with the ability to exploit complementary information about facial identity at different times. Interestingly, this conclusion echoes that based on the lack of temporal generalization found with cross-temporal object decoding of MEG data (Carlson et al., 2013; Isik et al., 2014), specifically, the lack of classification success with using training and testing data from distinct time intervals has been taken as evidence for the presence of different types of information over time (Grootswagers et al., 2017). Further, the boost in classification noted above is important for practical purposes: it suggests that investigations that place less emphasis on clarifying the time course of discrimination can be better served by exploiting patterns across larger temporal intervals. Accordingly, our subsequent investigations into face space structure and image reconstruction were conducted with both window-based and cumulative data.

Third, a neural-based estimate of face space was constructed from EEG data and its organization was explained by the presence of visual information captured by CIMs. This result is significant in that it confirms that pattern discrimination relies, at least partly, on relevant visual information (e.g., as opposed to higher-level semantic cues). More importantly, we note that neural-based face space has been examined in the context of fMRI (Loffler et al., 2005; Rotshtein et al., 2005; Gao and Wilson, 2013) and monkey neurophysiology (Leopold et al., 2006; Freiwald et al., 2009). Yet, many of its properties, as related to facial identity representation, remain to be clarified. For instance, its invariant structure across different types of image transformation remains to be assessed and quantified. Behavioral research suggests that face space topography is largely invariant across viewpoint and lighting (Blank and Yovel, 2011). Here, we reach a similar conclusion regarding the expression invariance of neural-based face space as derived from EEG data.

Fourth, image reconstruction was conducted with the aid of CIM features derived directly from the structure of EEG data (i.e., as opposed to predefined visual features selected due to their general biological plausibility). This endeavor builds on pattern classification while, critically, it validates its results by showcasing its reliance on relevant visual information encoded in the EEG signal. We found that multiple temporal intervals supported better-than-chance reconstruction for both neutral and happy faces with a peak in the proximity of the N170 component. Also, reconstruction accuracy was further boosted by considering temporally cumulative information, as used for pattern classification. More importantly, these results are notable in that, unlike previous work with fMRI-based facial image reconstruction (Cowen et al., 2014; Nestor et al., 2016), they exploit invariant face space information for reconstruction purposes. Thus, arguably the current findings speak to the visual nature of facial identity representations rather than just to lower-level pictorial aspects of face perception.

Further, the current work provides proof of principle for EEG-based image reconstruction. Importantly, not only does this demonstrate the applicability of image reconstruction to neuroimaging modalities other than fMRI but, critically, it shows that EEG-based reconstruction can compete in terms of overall accuracy with its fMRI counterpart (Nestor et al., 2016).

Thus, here we build on previous EEG investigations of face processing and on pattern analyses of neuroimaging data to address several theoretical and methodological issues. In particular, the current work capitalizes on previous attempts at clarifying the temporal profile of individual face processing via linear classification of spatiotemporal EEG patterns across facial expression (Nemrodov et al., 2016). In agreement with this previous work we find that individual faces can be discriminated from their corresponding EEG patterns, that their time course exhibits an extended interval of significant discrimination and that multiple discrimination peaks occur, including an early one in the vicinity of the N170 component. Unlike this previous work though, which only relied on a restricted set of eight male and female faces, we find that such discrimination can be performed even with a large, homogenous set of face images controlled for low and high-level face properties (e.g., through geometrical alignment and intensity normalization of 108 white male face images). Hence, differences in discrimination onset across studies (i.e., 70 ms in this previous work vs 152 ms here) are likely related to reliance on idiosyncratic image differences within a small stimulus set in this previous investigation. More importantly though, not only does the current work examine the time course of individual face classification in a more reliable and thorough manner but, critically, it utilizes its outcomes for the purpose of facial feature derivation and image reconstruction.

Naturally, boosting even further classification and reconstruction accuracy is an important future endeavor. Regarding classification, this could be achieved, for instance, through efficient techniques for feature selection, such as recursive feature elimination (Hanson and Halchenko, 2008; Nestor et al., 2011), aimed at reducing pattern dimensionality and optimizing discrimination performance. Since electrodes are likely to carry irrelevant or redundant information at multiple time points, eliminating this information from higher-dimensional spatiotemporal patterns (e.g., across all electrodes and time points) could benefit classification. Regarding reconstruction, more complex, biologically-plausible approaches can be developed, for instance, by considering shape and surface information separately within the reconstruction process. Since shape and surface provide complementary cues to face processing (Jiang et al., 2006; Andrews et al., 2016), it would be informative to derive separate types of CIMs corresponding to this distinction and to consider their separate contribution to facial image reconstruction.

Notably, however, beyond overall performance, EEG-based reconstruction stands out by its ability to clarify the dynamics of visual representations as they develop in response to a given stimulus. For instance, it can speak to how a percept evolves over time in response to a static stimulus, as attempted here, by inspecting image reconstruction across consecutive time windows. Alternatively, this method could be extended to recover fine-grained dynamic information as present in moving stimuli. While reconstruction of natural movies has been previously conducted with fMRI (Nishimoto et al., 2011), the superior temporal resolution of EEG could make this modality a more efficient choice for the recovery of dynamic visual information. Further, we note that the comparatively wide availability of EEG systems could also render this modality the preferred choice for the development of new types of image-reconstruction brain-computer interfaces.

Last, while the current investigation focuses on faces as a visual category of interest, we argue that the present methodological approach can inform individual object recognition more generally. This is theoretically suggested by the presence of common neurocomputational principles underlying face and object identification (Cowell and Cottrell, 2013; Wang et al., 2016) as well as, methodologically, by the ability to evaluate the dynamics of invariant object recognition (Isik, et al., 2014). Particularly encouraging in this sense is the success of efforts to construct and characterize object similarity spaces from MEG (Carlson et al., 2013) and EEG data (Kaneshiro et al., 2015).

To conclude, our investigation targets the neural dynamics of face processing as reflected by EEG patterns. Our findings shed new light on the time course of facial identity processing while providing a way to extract and to assess the underlying visual information. Last, from a methodological standpoint, our results establish the feasibility of EEG-based image reconstruction and, more generally, they confirm the rich informational content of spatiotemporal EEG patterns.

Footnotes

  • The authors declare no competing financial interests.

  • This work was supported by the Natural Sciences and Engineering Research Council of Canada (A.N. and M.N.) and by a Connaught New Researcher Award (A.N.).

This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International license, which permits unrestricted use, distribution and reproduction in any medium provided that the original work is properly attributed.

References

  1. ↵
    Andrews TJ, Baseler H, Jenkins R, Burton AM, Young AW (2016) Contributions of feature shapes and surface cues to the recognition and neural representation of facial identity. Cortex 83:280–291.
    OpenUrl
  2. ↵
    Anzellotti S, Fairhall SL, Caramazza A (2014) Decoding representations of face identity that are tolerant to rotation. Cereb Cortex 24:1988–1995. doi:10.1093/cercor/bht046 pmid:23463339
    OpenUrlCrossRefPubMed
  3. ↵
    Bentin S, Deouell LY (2000) Structural encoding and identification in face processing: ERP evidence for separate mechanisms. Cogn Neuropsychol 17:35–54.
    OpenUrlCrossRefPubMed
  4. ↵
    Blank I, Yovel G (2011) The structure of face–space is tolerant to lighting and viewpoint transformations. J Vis 11:1–13.
    OpenUrlAbstract/FREE Full Text
  5. ↵
    Blankertz B, Lemm S, Treder M, Haufe S, Müller KR (2011) Single-trial analysis and classification of ERP components—a tutorial. Neuroimage 56:814–825. doi:10.1016/j.neuroimage.2010.06.048 pmid:20600976
    OpenUrlCrossRefPubMed
  6. ↵
    Bowles DC, McKone E, Dawel A, Duchaine B, Palermo R, Schmalzl L, Rivolta D, Wilson CE, Yovel G (2009) Diagnosing prosopagnosia: effects of ageing, sex, and participant-stimulus ethnic match on the Cambridge Face Memory Test and Cambridge Face Perception Test. Cogn Neuropsychol 26:423–455.
    OpenUrlCrossRefPubMed
  7. ↵
    Brainard DH (1997) The psychophysics toolbox. Spat Vis 10:433–436. pmid:9176952
    OpenUrlCrossRefPubMed
  8. ↵
    Caharel S, D’Arripe O, Ramon M, Jacques C, Rossion B (2009) Early adaptation to repeated unfamiliar faces across viewpoint changes in the right hemisphere: evidence from the N170 ERP component. Neuropsychologia 47:639–643. doi:10.1016/j.neuropsychologia.2008.11.016
    OpenUrlCrossRefPubMed
  9. ↵
    Carlson TA, Tovar D, Alink A, Kriegeskorte N (2013) Representational dynamics of object vision: the first 1000 ms. J Vis 13:1–19.
    OpenUrlAbstract/FREE Full Text
  10. ↵
    Cauchoix M, Barragan-Jason G, Serre T, Barbeau EJ (2014) The neural dynamics of face detection in the wild revealed by MVPA. J Neurosci 34:846–854. doi:10.1523/JNEUROSCI.3030-13.2014 pmid:24431443
    OpenUrlAbstract/FREE Full Text
  11. ↵
    Chang C-C, Lin C-J (2011) LIBSVM. ACM Trans Intell Syst Technol 2:1–27.
    OpenUrlCrossRef
  12. ↵
    Chang L, Tsao DY (2017) The code for facial identity in the primate brain. Cell 169:1013–1028. doi:10.1016/j.cell.2017.05.011 pmid:28575666
    OpenUrlCrossRefPubMed
  13. ↵
    Cichy RM, Pantazis D, Oliva A (2014) Resolving human object recognition in space and time. Nat Neurosci 17:455–462. doi:10.1038/nn.3635 pmid:24464044
    OpenUrlCrossRefPubMed
  14. ↵
    Cowell RA, Cottrell GW (2013) What evidence supports special processing for faces? A cautionary tale for fMRI interpretation. J Cogn Neurosci 25:1777–1793.
    OpenUrl
  15. ↵
    Cowen AS, Chun MM, Kuhl BA (2014) Neural portraits of perception: reconstructing face images from evoked brain activity. Neuroimage 94:12–22. doi:10.1016/j.neuroimage.2014.03.018
    OpenUrlCrossRefPubMed
  16. ↵
    Crouzet SM, Kirchner H, Thorpe SJ (2010) Fast saccades toward faces: face detection in just 100 ms. J Vis 10:1–17.
    OpenUrlAbstract/FREE Full Text
  17. ↵
    Davidesco I, Zion-Golumbic E, Bickel S, Harel M, Groppe DM, Keller CJ, Schevon CA, McKhann GM, Goodman RR, Goelman G, Schroeder CE, Mehta AD, Malach R (2014) Exemplar selectivity reflects perceptual similarities in the human fusiform cortex. Cereb Cortex 24:1879–1893. doi:10.1093/cercor/bht038
    OpenUrlCrossRefPubMed
  18. ↵
    Delorme A, Sejnowski T, Makeig S (2007) Enhanced detection of artifacts in EEG data using higher-order statistics and independent component analysis. Neuroimage 34:1443–1449.
    OpenUrlCrossRefPubMed
  19. ↵
    Duchaine B, Nakayama K (2006) The Cambridge Face Memory Test: results for neurologically intact individuals and an investigation of its validity using inverted face stimuli and prosopagnosic participants. Neuropsychologia 44:576–585. doi:10.1016/j.neuropsychologia.2005.07.001
    OpenUrlCrossRefPubMed
  20. ↵
    Faul F, Erdfelder E, Lang A-G, Buchner A (2007) G*Power 3: a flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behav Res Methods 39:175–191.
    OpenUrlCrossRefPubMed
  21. ↵
    Freiwald WA, Tsao DY, Livingstone MS (2009) A face feature space in the macaque temporal lobe. Nat Neurosci 12:1187–1196. doi:10.1038/nn.2363 pmid:19668199
    OpenUrlCrossRefPubMed
  22. ↵
    Gao X, Wilson HR (2013) The neural representation of face space dimensions. Neuropsychologia 51:1787–1793. doi:10.1016/j.neuropsychologia.2013.07.001 pmid:23850598
    OpenUrlCrossRefPubMed
  23. ↵
    Ghuman AS, Brunet NM, Li Y, Konecky RO, Pyles JA, Walls SA, Destefinom V, Wang W, Richardson RM (2014) Dynamic encoding of face information in the human fusiform gyrus. Nat Commun 5:5672.
    OpenUrlCrossRefPubMed
  24. ↵
    Goesaert E, Op de Beeck HP (2013) Representations of facial identity information in the ventral visual stream investigated with multivoxel pattern analyses. J Neurosci 33:8549–8558.
    OpenUrlAbstract/FREE Full Text
  25. ↵
    Gosselin F, Schyns PG (2003) Superstitious perceptions reveal properties of internal representations. Psychol Sci 14:505–509.
    OpenUrlCrossRefPubMed
  26. ↵
    Grootswagers T, Wardle SG, Carlson TA (2017) Decoding dynamic brain patterns from evoked responses: a tutorial on multivariate pattern analysis applied to time-series neuroimaging data. J Cogn Neurosci 29:677–697. doi:10.1162/jocn_a_01068
    OpenUrlCrossRefPubMed
  27. ↵
    Hanson SJ, Halchenko YO (2008) Brain reading using full brain support vector machines for object recognition: there is no “face” identification area. Neural Comput 20:486–503. doi:10.1162/neco.2007.09-06-340
    OpenUrlCrossRefPubMed
  28. ↵
    Heisz JJ, Watter S, Shedden JM (2006) Progressive N170 habituation to unattended repeated faces. Vis Res 46:47–56. doi:10.1016/j.visres.2005.09.028 pmid:16289274
    OpenUrlCrossRefPubMed
  29. ↵
    Huddy V, Schweinberger SR, Jentzsch I, Burton AM (2003) Matching faces for semantic information and names: an event-related brain potentials study. Cogn Brain Res 17:314–326.
    OpenUrlCrossRefPubMed
  30. ↵
    Hung CP, Kreiman G, Poggio T, DiCarlo JJ (2005) Fast readout of object identity from macaque inferior temporal cortex. Science 310:863–866. doi:10.1126/science.1117593 pmid:16272124
    OpenUrlAbstract/FREE Full Text
  31. ↵
    Isik L, Meyers EM, Leibo JZ, Poggio TA, (2014) The dynamics of invariant object recognition in the human visual system. J Neurophysiol 111:91–102. doi:10.1152/jn.00394.2013 pmid:24089402
    OpenUrlCrossRefPubMed
  32. ↵
    Itier RJ, Taylor MJ (2002) Inversion and contrast polarity reversal affect both encoding and recognition processes of unfamiliar faces: a repetition study using ERPs. Neuroimage 15:353–372. doi:10.1006/nimg.2001.0982
    OpenUrlCrossRefPubMed
  33. ↵
    Jackson DA (1995) Protest: a procrustean randomization test of community environment concordance. Ecoscience 2:297–303.
    OpenUrl
  34. ↵
    Jacques C, D’Arripe O, Rossion B (2007) The time course of the inversion effect during individual face discrimination. J Vis 7:1–9.
    OpenUrlAbstract/FREE Full Text
  35. ↵
    Jiang F, Blanz V, O’Toole AJ (2006) Probing the visual representation of faces with adaptation: a view from the other side of the mean. Psychol Sci 17:493–500. doi:10.1111/j.1467-9280.2006.01734.x pmid:16771799
    OpenUrlCrossRefPubMed
  36. ↵
    Kaneshiro B, Perreau Guimaraes M, Kim H, Norcia AM, Suppes P, Guimaraes MP, Anthony M (2015) A representational similarity analysis of the dynamics of object processing using single-trial EEG classification. PLoS One 10:e0135697.
  37. ↵
    Kriegeskorte N, Mur M, Bandettini PA (2008) Representational similarity analysis - connecting the branches of systems neuroscience. Front Syst Neurosci 2:4.
    OpenUrlCrossRefPubMed
  38. ↵
    Langner O, Dotsch R, Bijlstra G, Wigboldus DHJ, Hawk ST, van Knippenberg A (2010) Presentation and validation of the Radboud Faces Database. Cogn Emot 24:1377–1388.
    OpenUrlCrossRef
  39. ↵
    Lehky SR (2000) Fine discrimination of faces can be performed rapidly. J Cogn Neurosci 12:848–855. pmid:11054926
    OpenUrlCrossRefPubMed
  40. ↵
    Leopold DA, Bondar IV, Giese MA (2006) Norm-based face encoding by single neurons in the monkey inferotemporal cortex. Nature 442:572–575. doi:10.1038/nature04951 pmid:16862123
    OpenUrlCrossRefPubMed
  41. ↵
    Liu H, Agam Y, Madsen JR, Kreiman G (2009) Timing, timing, timing: fast decoding of object information from intracranial field potentials in human visual cortex. Neuron 62:281–290. doi:10.1016/j.neuron.2009.02.025
    OpenUrlCrossRefPubMed
  42. ↵
    Loffler G, Yourganov G, Wilkinson F, Wilson HR (2005) fMRI evidence for the neural representation of faces. Nat Neurosci 8:1386–1390. doi:10.1038/nn1538 pmid:16136037
    OpenUrlCrossRefPubMed
  43. ↵
    Marks DF (1995) New directions for mental imagery research. J Mental Imagery 19:153–167.
    OpenUrl
  44. ↵
    Martin-Malivel J, Mangini MC, Fagot J, Biederman I (2006) Do humans and baboons use the same information when categorizing human and baboon faces? Psychol Sci 17:599–607. doi:10.1111/j.1467-9280.2006.01751.x
    OpenUrlCrossRefPubMed
  45. ↵
    Martinez A, Benavente R (1998) The AR Face CVC Technical Report #24. Available at www.cat.uab.cat/Public/Publications/1998/MaB1998/CVCReport24.pdf.
  46. ↵
    Miyawaki Y, Uchida H, Yamashita O, Sato MA, Morito Y, Tanabe HC, Sadato N, Kamitani Y (2008) Visual image reconstruction from human brain activity using a combination of multiscale local image decoders. Neuron 60:915–929.
    OpenUrlCrossRefPubMed
  47. ↵
    Mouraux A, Iannetti GD (2008) Across-trial averaging of event-related EEG responses and beyond. Magn Reson Imaging 26:1041–1054.
    OpenUrlCrossRefPubMed
  48. ↵
    Murray RF (2011) Classification images: a review. J Vis 11. doi:10.1167/11.5.2
    OpenUrlAbstract/FREE Full Text
  49. ↵
    Naselaris T, Prenger RJ, Kay KN, Oliver MD, Gallant JL (2009) Bayesian reconstruction of natural images from human brain activity. Neuron 63:902–915. doi:10.1016/j.neuron.2009.09.006
    OpenUrlCrossRefPubMed
  50. ↵
    Nemrodov D, Niemeier M, Mok JNY, Nestor A (2016) The time course of individual face recognition: a pattern analysis of ERP signals. Neuroimage 132:469–476. doi:10.1016/j.neuroimage.2016.03.006 pmid:26973169
    OpenUrlCrossRefPubMed
  51. ↵
    Nestor A, Plaut DC, Behrmann M (2011) Unraveling the distributed neural code of facial identity through spatiotemporal pattern analysis. Proc Natl Acad Sci USA 108:9998–10003.
    OpenUrlAbstract/FREE Full Text
  52. ↵
    Nestor A, Plaut DC, Behrmann M (2013) Face-space architectures: evidence for the use of independent color-based features. Psychol Sci 24:1294–1300. doi:10.1177/0956797612464889
    OpenUrlCrossRefPubMed
  53. ↵
    Nestor A, Plaut DC, Behrmann M (2016) Feature-based face representations and image reconstruction from behavioral and neural data. Proc Natl Acad Sci USA 113:416–421. doi:10.1073/pnas.1514551112 pmid:26711997
    OpenUrlAbstract/FREE Full Text
  54. ↵
    Newell FN, Chiroro P, Valentine T (1999) Recognizing unfamiliar faces: the effects of distinctiveness and view. Q J Exp Psychol A 52:509–534.
    OpenUrlCrossRefPubMed
  55. ↵
    Nishimoto S, Vu AT, Naselaris T, Benjamini Y, Yu B, Gallant JL (2011) Reconstructing visual experiences from brain activity evoked by natural movies. Curr Biol 21:1641–1646.
    OpenUrlCrossRefPubMed
  56. ↵
    Pelli D (1997) The VideoToolbox software for visual psychophysics: transforming numbers into movies. Spat Vis 10:437–442. doi:10.1163/156856897X00366
    OpenUrlCrossRefPubMed
  57. ↵
    Rossion B, Caharel S (2011) ERP evidence for the speed of face categorization in the human brain: disentangling the contribution of low-level visual cues from face perception. Vis Res 51:1297–1311. doi:10.1016/j.visres.2011.04.003
    OpenUrlCrossRefPubMed
  58. ↵
    Rotshtein P, Henson RN, Treves A, Driver J, Dolan RJ (2005) Morphing Marilyn into Maggie dissociates physical and identity face representations in the brain. Nat Neurosci 8:107–113. doi:10.1038/nn1370 pmid:15592463
    OpenUrlCrossRefPubMed
  59. ↵
    Schweinberger SR, Pickering EC, Burton AM, Kaufmann JM (2002) Human brain potential correlates of repetition priming in face and name recognition. Neuropsychologia 40:2057–2073.
    OpenUrlCrossRefPubMed
  60. ↵
    Sekuler AB, Gaspar CM, Gold JM, Bennett PJ (2004) Inversion leads to quantitative, not qualitative, changes in face processing. Curr Biol 14:391–396.
    OpenUrlCrossRefPubMed
  61. ↵
    Smith ML, Gosselin F, Schyns PG (2012) Measuring internal representations from behavioral and brain data. Current Biology 22:191–196. doi:10.1016/j.cub.2011.11.061
    OpenUrlCrossRefPubMed
  62. ↵
    Tanaka J, Curran T (2001) A neural basis for expert object recognition. Psychol Sci 12:43–47.
    OpenUrlCrossRefPubMed
  63. ↵
    Tanaka J, Curran T, Porterfield AL, Collins D (2006) Activation of preexisting and acquired face representations: the N250 event-related potential as an index of face familiarity. J Cogn Neurosci 18:1488–1497.
    OpenUrlCrossRefPubMed
  64. ↵
    Thomaz CE, Giraldi GA (2010) A new ranking method for principal components analysis and its application to face image analysis. Image Vis Comput 28:902–913.
    OpenUrlCrossRef
  65. ↵
    Valentine T (1991) A unified account of the effects of distinctiveness, inversion, and race in face recognition. Q J Exp Psychol A 43:161–204. doi:10.1080/14640749108400966
    OpenUrlCrossRefPubMed
  66. ↵
    Van de Nieuwenhuijzen ME, Backus AR, Bahramisharif A, Doeller CF, Jensen O, van Gerven MAJ (2013) MEG-based decoding of the spatiotemporal dynamics of visual category perception. Neuroimage 83:1063–1073.
    OpenUrl
  67. ↵
    Vida MD, Nestor A, Plaut DC, Behrmann M (2017) Spatiotemporal dynamics of similarity-based neural representations of facial identity. Proc Natl Acad Sci USA 114:388–393.
    OpenUrlAbstract/FREE Full Text
  68. ↵
    Wang P, Gauthier I, Cottrell GW (2016) Are face and object recognition independent? A neurocomputational modeling exploration. J Cogn Neurosci 28:558–574.
    OpenUrl
  69. ↵
    Zheng X, Mondloch CJ, Segalowitz SJ (2012) The timing of individual face recognition in the brain. Neuropsychologia 50:1451–1461. doi:10.1016/j.neuropsychologia.2012.02.030 pmid:22410414
    OpenUrlCrossRefPubMed

Synthesis

Reviewing Editor: Bradley Postle, University of Wisconsin

Decisions are customarily a result of the Reviewing Editor and the peer reviewers coming together and discussing their recommendations until a consensus is reached. When revisions are invited, a fact-based synthesis statement explaining their decision and outlining what is needed to prepare a revision will be listed below. The following reviewer(s) agreed to reveal their identity: Emmanuel Barbeau

Here are the two reviewers' comments, in full:

Reviewer #1

The authors expand on previous (presumably their) work by demonstrating that the information contained in EEG signals is not only sufficient for face identity classification but also robust enough for face reconstruction. Overall the authors did an excellent work and I only have minor comments:  

From Figures 2A, 7B and the supplemental movie it appears that classification/reconstruction spikes momentarily at around 0 (stimulus onset): at time t=-25ms (before stimulus onset), reconstruction accuracy is at 69% in the supplemental movie the authors provided. At the group level this is also evident but less so (at t=0 accuracy is around 54%).This is surprising and raises the question of what information the classification algorithm is picking up at that early stage. I understand that the authors controlled the luminance and RMS contrast of their stimuli but I am wondering if the classifiers are still picking up on lower-level differences between the faces. Perhaps the authors can address this comment in the discussion section.

Reviewer #2

In this study the authors use a combination of decoding and image reconstruction techniques to explore how EEG recorded from the scalp is related to face identity. The most interesting part of this study is the demonstration that it is possible to classify and reconstruct individual faces from EEG with methodological approaches that are well described. With some exceptions, most previous reported had focused on category decoding rather than on individual faces, which makes this study an interesting contribution. In contrast, the contribution of this study to the understanding of the time course of face processing is more minor. 

I have a few questions/suggestions: 

- the authors use in their title the term “facial identity processing” which is appropriate. However they also sometimes use in the article (and abstract) the term “face recognition” which is rather vague, and in fact confusing because there is no relation in this study with familiarity or semantic processing. 

- Have the authors included all subjects that were recorded (13) or have they excluded some? If yes, explain why. 

- Have the authors verified if there was an effect of the database from which the faces were chosen? 

- In the Methods, it is sometimes difficult to follow the exact number of trials/face that went into each analysis. The authors start from 64 trials/faces (this is not specifically written), then average trials to end up with 16 EEG trials/faces (which is something we understand two pages later). As this is critical for any future work that will aim at reducing the number of trials necessary for suitable decoding, all this could be more explicit or discussed. 

- A related question is why the authors chose to average EEG trials by group of 4. Did they make specific test about this? 

- The results about the accuracy levels for single-participants reconstructions are not provided (only the p values are). Please correct. I suppose the “d” mentioned in this section (p 22, line 45) refer to Cohen's d. Please clarify. 

- The Nemrodov et al (2016) study is related to the present one. It is briefly mentioned in the Discussion but should be discussed more extensively, both for its results and methodology. 

- Please change all histograms to boxplots and distribution of individual results. Those are much more informative. 

Typos and other: 

- I couldn't find the Naselaris in the references (p3, line 39). 

- p11, line 28, correct ‘data. and’. 

- p22, line 53, correct ‘evinces’ 

Author Response

We are grateful to our reviewers for the comments and suggestions that they have offered. We hereby submit a revised version of the paper, modified in accordance with these suggestions and we are of the opinion that the paper has been strengthened by these revisions. Below, we present a point-by-point response to the reviewers' comments and we reference the sections/figures of the manuscript that have been modified or added (all modifications to the manuscript have been marked with red font).

Reviewer 1

We thank the reviewer for his evaluation of our work.

One important issue is raised regarding the influence of lower-level image differences and their impact on our results. We agree that lower-level image differences can impact results and that it is critical to consider this impact carefully. Specifically for this reason we conducted classification both within expression (i.e., by training and testing classifiers across the same images) and across expression (i.e., by training on images displaying one expression and by testing classifiers across images of the same individuals displaying a different expression). In line with the expectation that within-expression classification exhibits more sensitivity to lower-level image differences we found higher levels of classification accuracy for the former relative to the latter. Critically though, cross-expression classification, which aims to tap into expression-invariant individual face processing, yielded significant results and, further, its time course mirrored that of within-expression classification. In conjunction with the ability to map neural-based face spaces across expression, these results argue in our view for the value of higher-level expression-invariant visual information in classification and reconstruction.

More specifically, the reviewer asks whether early spikes around stimulus onset in classification/reconstruction accuracy can be ascribed to lower-level image differences across our stimuli. Three points are relevant in this respect. First, such spikes do not occur consistently across participants and across different analyses - for instance, an early spike in group-based classification accuracy only occurs briefly for classification across expression in Fig 2a but not for its within-expression counterpart or for the single-participant results shown in Fig 2b. Second, the amplitude of these spikes is visibly lower than that of later ones during intervals of interest (e.g., in the vicinity of the N170 component). And third, these spikes only achieve statistical significance in a single instance: for classification across expression in Fig 2a. However, in this specific instance we reason that it is a lot more likely that within-expression classification would capture lower-level image differences at that time, which is not the case as shown in the same figure. Further, the timing of this spike, 0 to 5ms after stimulus onset, is arguably too early to capture any stimulus-driven effect related to visual processing. Hence, in the absence of robust and consistent results we treat this specific data point as a false positive. This point is currently summarized in the manuscript: “In addition, an earlier interval of significance was found for across-expression classification between 0-5ms; however, given its very early occurrence, its reduced amplitude and the absence of its replication by within-expression classification we treat this data point as a false positive” (p. 16).

Reviewer 2

We are grateful for the comments and recommendations that our reviewer has offered.

1. The reviewer recommends that the term ‘face recognition’ be removed from the manuscript as it does not accurately reflect the topic of research. Accordingly, this term was replaced with ‘facial identity processing’ which was deemed appropriate by our reviewer.

2. It is asked whether all participants reported in the manuscripts (i.e., 13) contributed data to the study or whether any were excluded. All participants were included, as currently mentioned in the manuscript...

3. The reviewer raises an interesting issue by inquiring whether there is any effect of the database from which faces were chosen.

Given that different databases sample individuals of different nationalities (FEI - Brazilian, AR - American, Radboud - Dutch) we anticipated a possible effect of database. However, it is important to establish that the results reported in the manuscript are not solely due to such differences. To assess this possibility we examined pairwise classification results for faces extracted from the same databases, for a total of 472 pairs, and, separately, for faces extracted from different databases, for a total of 1918 pairs (p. 10). In agreement with the hypothesis above we found that classification estimates were lower for the former than the latter yet, critically, classification results were above chance in all cases. Similar results were also found by recomputing and comparing reconstruction accuracies relative to faces within the same versus different databases; however, for convenience and, also, given that classification results form the basis of all other analyses, including reconstruction, in the manuscript we focused on the impact of database on classification.

For instance, the plots below display results for temporally cumulative analyses across participants for both (a) within- and (b) cross-expression classification - analyses follow the procedure described for Fig 3b in the manuscript, except that results are collapsed across neutral and happy faces.

Parametric tests associated with these results are currently summarized in the manuscript (p. 18): “To examine possible effects of face database the analysis above was repeated while restricting pattern classification either to pairs of faces from the same database or from different databases. A two-way repeated measures analysis of variance (2 discrimination types: within/across expression x 2 pair types: within/across database) was carried out to this end - classification estimates were collapsed across neutral and happy faces given the absence of any expression effects above. This analysis revealed a main effect of discrimination types (F(1,12) =45.92, p<0.001, p2 =0.79), with higher accuracy for within than across-expression discrimination, as well as a main effect of pair types (F(1,12) =38.73, p<0.001, p2 =0.76), with higher accuracy for within than across-database classification. Critically though, all classification estimates were significantly above chance (two-tailed t-tests against 50% chance-level discrimination); mean accuracy=56.3%, t(12)=8.36,p<0.001, Cohen's d=2.32 for within-expression, within-database discrimination; mean accuracy=58.7%, t(12)=8.38, p<0.001, Cohen's d=2.32 for within-expression, across-database discrimination; mean accuracy=53.9%, t(12)=7.64 ,p<0.001, Cohen's d=2.12 for across-expression, within- database discrimination; mean accuracy =55.9%, t(12)=6.88,p<0.001, Cohen's d=1.91 for across-expression, across- database discrimination).”

Last, an incorrect reference to an additional database (i.e., FERET) was eliminated from the manuscript.

4. The reviewer asks why EEG trials have been averaged in groups of 4 to deliver observations for pattern analyses.

As currently detailed in the manuscript “this procedure was also instrumental in handling missing data following trial removal and in balancing the number of observations for pattern classification. Specifically, since it is possible that both trials associated with a given stimulus in a given block be removed (e.g., due to artifacts), averaging data across single blocks can lead to different numbers of observations for different stimuli. However, averaging data across pairs of consecutive blocks (i.e., 1-4 trials per stimulus following data removal) ensured that equal numbers of observations can be constructed for each pair of stimuli undergoing pattern classification.”

5. The reviewer asks us to specify the accuracy levels for single-participant reconstructions and also to confirm whether the “d” notation refers to Cohen's d.

Accuracy levels have been specified accordingly and corresponding references to Cohen's d have been confirmed as suggested (p. 22,23).

6. The reviewer recommends that the results of Nemrodov et al (2016) be discussed and related to the current ones.

The discussion follows this recommendation accordingly: “Thus, here we build upon previous EEG investigations of face processing and on pattern analyses of neuroimaging data to address several theoretical and methodological issues. In particular, the current work capitalizes on previous attempts at clarifying the temporal profile of individual face processing via linear classification of spatiotemporal EEG patterns across facial expression (Nemrodov et al., 2016). In agreement with this previous work we find that individual faces can be discriminated from their corresponding EEG patterns, that their time course exhibits an extended interval of significant discrimination and that multiple discrimination peaks occur, including an early one in the vicinity of the N170 component. Unlike this previous work though, which only relied on a restricted set of eight male and female faces, we find that such discrimination can be performed even with a large, homogenous set of face images controlled for low and high-level face properties (e.g., through geometrical alignment and intensity normalization of 108 Caucasian male face images). Hence, differences in discrimination onset across studies (i.e., 70ms in this previous work versus 152ms here) are likely related to reliance on idiosyncratic image differences within a small stimulus set in this previous investigation. More importantly though, not only does the current work examine the time course of individual face classification in a more reliable and thorough manner but, critically, it utilizes its outcomes for the purpose of facial feature derivation and image reconstruction.” (p. 26).

7. It is suggested that certain plots should be changed to boxplots displaying the distribution of individual results.

Following this recommendation Figures 3b and 8c, displaying the results of individual participants, have been changed to boxplots - other plots displaying permutation-based analyses for single average estimates, such as Fig 3a, are still shown as bar graphs since the distribution of a large number of permutations may be less informative and more difficult to visualize.

All other suggestions for minor modifications and corrections (missing reference, spelling, etc.) have been implemented accordingly.

Back to top

In this issue

eneuro: 5 (1)
eNeuro
Vol. 5, Issue 1
January/February 2018
  • Table of Contents
  • Index by author
Email

Thank you for sharing this eNeuro article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
The Neural Dynamics of Facial Identity Processing: Insights from EEG-Based Pattern Analysis and Image Reconstruction
(Your Name) has forwarded a page to you from eNeuro
(Your Name) thought you would be interested in this article in eNeuro.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Print
View Full Page PDF
Citation Tools
The Neural Dynamics of Facial Identity Processing: Insights from EEG-Based Pattern Analysis and Image Reconstruction
Dan Nemrodov, Matthias Niemeier, Ashutosh Patel, Adrian Nestor
eNeuro 29 January 2018, 5 (1) ENEURO.0358-17.2018; DOI: 10.1523/ENEURO.0358-17.2018

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article
Share
The Neural Dynamics of Facial Identity Processing: Insights from EEG-Based Pattern Analysis and Image Reconstruction
Dan Nemrodov, Matthias Niemeier, Ashutosh Patel, Adrian Nestor
eNeuro 29 January 2018, 5 (1) ENEURO.0358-17.2018; DOI: 10.1523/ENEURO.0358-17.2018
Reddit logo Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • Abstract
    • Significance Statement
    • Introduction
    • Materials and Methods
    • Results
    • Discussion
    • Footnotes
    • References
    • Synthesis
    • Author Response
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF

Keywords

  • ERP
  • face space
  • image reconstruction
  • N170
  • pattern analysis
  • spatiotemporal dynamics

Responses to this article

Respond to this article

Jump to comment:

No eLetters have been published for this article.

Related Articles

Cited By...

More in this TOC Section

New Research

  • Heterozygous Dab1 null mutation disrupts neocortical and hippocampal development
  • The nasal solitary chemosensory cell signaling pathway triggers mouse avoidance behavior to inhaled nebulized irritants
  • Different control strategies drive interlimb differences in performance and adaptation during reaching movements in novel dynamics
Show more New Research

Sensory and Motor Systems

  • The nasal solitary chemosensory cell signaling pathway triggers mouse avoidance behavior to inhaled nebulized irritants
  • Different control strategies drive interlimb differences in performance and adaptation during reaching movements in novel dynamics
  • Taste-odor association learning alters the dynamics of intra-oral odor responses in the posterior piriform cortex of awake rats
Show more Sensory and Motor Systems

Subjects

  • Sensory and Motor Systems

  • Home
  • Alerts
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Latest Articles
  • Issue Archive
  • Blog
  • Browse by Topic

Information

  • For Authors
  • For the Media

About

  • About the Journal
  • Editorial Board
  • Privacy Policy
  • Contact
  • Feedback
(eNeuro logo)
(SfN logo)

Copyright © 2023 by the Society for Neuroscience.
eNeuro eISSN: 2373-2822

The ideas and opinions expressed in eNeuro do not necessarily reflect those of SfN or the eNeuro Editorial Board. Publication of an advertisement or other product mention in eNeuro should not be construed as an endorsement of the manufacturer’s claims. SfN does not assume any responsibility for any injury and/or damage to persons or property arising from or related to any use of any material contained in eNeuro.