- Open Access
Spatiotemporal dipole source localization of face processing ERPs in adolescents: a preliminary study
Behavioral and Brain Functions volume 5, Article number: 16 (2009)
Despite extensive investigation of the neural systems for face perception and emotion recognition in adults and young children in the past, the precise temporal activation of brain sources specific to the processing of emotional facial expressions in older children and adolescents is not well known. This preliminary study aims to trace the spatiotemporal dynamics of facial emotion processing during adolescence and provide a basis for future developmental studies and comparisons with patient populations that have social-emotional deficits such as autism.
We presented pictures showing happy, angry, fearful, or neutral facial expressions to healthy adolescents (aged 10–16 years) and recorded 128-channel event-related potentials (ERPs) while they performed an emotion discrimination task. ERP components were analyzed for effects of age and emotion on amplitude and latency. The underlying cortical sources of scalp ERP activity were modeled as multiple equivalent current dipoles using Brain Electrical Source Analysis (BESA).
Initial global/holistic processing of faces (P1) took place in the visual association cortex (lingual gyrus) around 120 ms post-stimulus. Next, structural encoding of facial features (N170) occurred between 160–200 ms in the inferior temporal/fusiform region, and perhaps early emotion processing (Vertex Positive Potential or VPP) in the amygdala and orbitofrontal cortex. Finally, cognitive analysis of facial expressions (P2) in the prefrontal cortex and emotional reactions in somatosensory areas were observed from about 230 ms onwards. The temporal sequence of cortical source activation in response to facial emotion processing was occipital, prefrontal, fusiform, parietal for young adolescents and occipital, limbic, inferior temporal, and prefrontal for older adolescents.
This is a first report of high-density ERP dipole source analysis in healthy adolescents which traces the sequence of neural activity within the first 500 ms of categorizing emotion from faces. Our spatio-temporal brain source models showed the presence of adult-like cortical networks for face processing in adolescents, whose functional specificity to different emotions appear to be not yet fully mature. Age-related differences in brain activation patterns illustrate the continued development and maturation of distinct neural systems for processing facial expressions during adolescence and possible changes in emotion perception, experience, and reaction with age.
Discerning facial emotions is fundamental for normal social interaction and commences in early infancy . Recognition and interpretation of emotion from facial expressions occurs almost instantaneously and is used in social referencing, for example, when an infant first encounters a novel object and looks towards the parent to direct their behavior. If the parent expresses a happy face, the child will understand this to be an encouragement to approach the object . On the other hand, if the parent displays an expression of fear or disgust, the child will tend to avoid the same object. This skill of "reading" emotions from faces becomes more proficient as children grow up and experience a multitude of social, emotional and cognitive changes. However, the neural circuitry that controls facial emotion processing during late childhood and adolescence has received scant attention to date . This period of cognitive and behavioral development is critical since heterochronous brain changes take place: cortical synaptic elimination occurs in the primary visual and auditory cortices during late childhood, then subsequently in the prefrontal cortex in mid-adolescence . Compatible with this is the region-specific nonlinear increase in cortical grey matter volume, which peaks in the frontal and parietal lobes around age 12, followed by the temporal and occipital lobes at about ages 16 and 20 respectively .
Neurobiological findings from lesion, functional magnetic resonance imaging (fMRI) and positron emission tomography (PET) studies have revealed that, just as distinct brain structures subserve verbal language, specialized neural networks also exist for processing the non-verbal language of facial expressions. These include the occipito-temporal cortices, amygdala, orbitofrontal cortex, as well as somatosensory-related cortices in the right hemisphere . Studies of pervasive developmental disorders such as autism have also identified abnormalities in the fusiform gyrus, superior temporal sulcus, and amygdale , suggesting that the innate inability to perceive and respond to complex emotional expressions of others may parallel early abnormal development of brain regions involved in face processing. Prosopagnosic patients who demonstrate impairments in recognizing familiar faces, but have normal abilities in identifying facial expressions of emotion , suggest a functional dissociation between facial identity and facial expression processing . In accordance with Bruce and Young's cognitive model for face perception , Haxby, Hoffman, and Gobbini's neuroanatomical model  posits that, after the initial stage of face perception has taken place in the inferior occipital gyri, (bidirectional) information proceeds along two distinct parallel neural pathways: one for coding facial identity (lateral fusiform gyrus for recognizing unique invariant facial features), and the other for coding facial expression (superior temporal sulcus region for analyzing the changeable aspects of faces e.g. eye gaze and lip movement). The models of Bruce and Young  and Haxby et al.  however, do not include the precise time frame in which the various components of face processing take place.
Such time-sensitive issues can be addressed using dense-array electroencephalography (EEG), event-related potentials (ERPs), and magnetoencephalography (MEG), since these techniques provide millisecond temporal resolution not accessible by hemodynamic measures. Recent work has localized sources of face-specific components to the occipital and inferotemporal cortices [10, 11], which overlap with regions identified in fMRI studies [12–14]. For example, an equivalent current dipole model for adult face recognition showed the following temporal dynamics of ERP source activation: bilateral lingual gyri at 120 ms, followed by the right and then left fusiform/hippocampal gyri around 150–170 ms, bilateral medial temporal gyri from 200 ms onwards, the right caudate nucleus peaking at 300 ms, and finally the anterior cingulate gyrus 300 ms after stimulus onset. However, such source localization has not yet been performed on ERP data from children and adolescents actively processing facial expressions of emotion, and only few studies to date have examined differential neural processing of emotional facial expressions in children [15, 16] and adolescents , let alone using high-density EEG.
Adult studies have suggested that emotionally expressive faces (especially high-arousal negative emotions like fear or anger) tend to evoke larger ERP responses than emotionally unexpressive (neutral) faces [15, 16, 18]. The "face-sensitive" N170 component over posterior temporal scalp regions is said to reflect the encoding of configural and relational features within a face  and appears not to be selective for any particular facial expression [17, 20–23]. However, others have found that negative (fear and disgust) emotions do indeed evoke larger N170 amplitudes in adults than positive (happy, surprise, neutral) ones [24, 25]. Similarly, the vertex positive potential (VPP), originating from the same pair of dipoles in the lateral inferior occipital cortex/posterior fusiform gyrus that generates the N170 in adults , is enhanced by facial expressions of fear [20, 21]. In addition, the timing of the neural response in adults seems also affected by emotion, such that the N170 appears at earlier latencies for positive than negative emotions. Krolak-Salmon et al.  reported that occipital ERP patterns first distinguished between emotional and neutral faces during the 250–550 ms time interval, and then differed between emotional expressions (fear and happiness in particular) from 550–750 ms over right occipito-temporal scalp regions. Activity specific to expressions of disgust occurred even later, between 700–900 ms in frontal and right temporal areas .
Several studies of young children (3–7 years of age) have shown differential ERP activity to facial emotion at latencies longer than about 300 ms [27–29], but earlier emotion effects on the face-elicited P1, N170, or P2 components as seen in adults were not present in children between 9–15 years . In contrast, Batty and Taylor  detected an effect of emotion on the P1 latency in children as young as four years old. This early emotion effect, indexing rapid holistic processing of faces , however, disappeared with age (absent after 7 years old) along with age-related decreases in P1 amplitude and latency. By mid-teens, global processing of emotions was replaced by detailed configural processing, reflected by an emotion effect on the N170 amplitude in 14- to 15-year-olds [19, 30]. This change in the spatio-temporal profile of face-processing ERPs with development  suggests that as cortical circuits mature and become more specialized for processing faces, dipole generators of ERP components may shift in location or orientation from childhood to adolescence . For example, it is thought that two functionally distinct cortical generators of the N170a and N170b subcomponents in children are fused into a single source responsible for the adult N170 . One model of adult face encoding and recognition  explained the posterior P1 and P2 components with bilateral dipole generators in the parietooccipital cortex, the N170 component by a pair of posterior ventral dipoles, and the VPP component by an orbitofrontal regional dipole source in the right hemisphere. However, no study has yet investigated the dipole source location and time course of processing facial expressions of emotion in the adolescent years.
The main purpose of this study was to construct a dipole model of brain sources for healthy adolescents to show the temporal unfolding of neural events that occur during explicit processing of different facial emotions. In addition, we examined the effects of emotion on ERP amplitudes, latencies, and scalp distribution, as well as developmental changes in ERPs to faces from early (age 10 to 13 years) to late (age 14 to 16 years) adolescence. We hypothesized that facial-emotional processing would occur rapidly in time, sequentially activating brain regions similar to those seen in adults subserving face detection, structural encoding of facial features, and emotion recognition.
Twenty healthy right-handed Chinese adolescents (12 male, 8 female) aged between 10 and 16 years (mean 14.2 ± 1.8 years) took part in this study. In order to examine developmental changes in face-specific ERPs and their sources , participants were divided into two groups: 10 young adolescents aged 10–13 years, and 10 older adolescents aged 14–16 years. However, due to severe artifact contamination in four young participants, we analysed 6 datasets (3 male) in the young group and 10 datasets (5 male) in the old group.
The research protocol was approved by the Institutional Review Board of the hospital where the study took place and all participants were compensated for travel and inconvenience. Parents of participating children signed informed consent and every participant gave their assent to participate in the project. All were screened for psychiatric illness using the parental Chinese Diagnostic Interview Schedule for Children for DSMIV  and had no history of headache, loss of consciousness, nor any neurological disorder. All attended normal school and had normal or corrected-to-normal vision.
Stimuli and procedure
ERP stimuli consisted of 32 color photographs selected from the standardized set of Japanese and Caucasian Facial Expressions of Emotion (JACFEE) and Neutral Faces (JACNeuF) . The images of four males and four females each depicting happy, angry, fearful, and neutral facial expressions were randomized six times to generate a sequence of 192 stimulus trials. Participants were instructed to press one of four buttons corresponding to 'happy', 'angry', 'fearful', or 'neutral' expressions, in response to the face stimuli presented. As shown in Figure 1, each face was presented for 500 ms on the full screen of an 11- by 13-inch monitor situated about one meter in front of the participant. A central fixation '+' was displayed during inter-stimulus intervals of lengths varying between 1500–2000 ms. E-Prime v.1.0 beta (Psychology Software Tools, Inc.) was used to present the visual stimuli and to record reaction times and button-press responses. During EEG recording, the attentive state of the participant was monitored and short breaks were permitted after each 2-minute block of 48 trials to allow the participant to rest.
EEG recording and signal processing
Scalp EEG was recorded using a 128-channel Geodesic Sensor Net (Electrical Geodesics, Inc., Eugene, OR, USA) that matched the head size of the participant. Electrode impedances were adjusted to stabilize under 50 kΩ before recording, and continuous EEG was sampled at 500 Hz with filters 0.1 to 100 Hz using the vertex (electrode 129) as reference. Offline, notch filters were employed to remove electrical noise from the mains power supply (50 Hz) and the stimulus display monitor (60 Hz refresh rate). Subsequent data processing was done using Brain Electrical Source Analysis (BESA 5.1, MEGIS Software GmbH, Munich, Germany). Eye-blink patterns were marked and averaged for each individual and the spatial topography of the blink artifact was used for correction of individual EEG data by the surrogate model approach . The surrogate model for brain activity consisted of 15 regional sources covering all brain regions. After blink correction, trials with remaining artifacts between -100 ms and 500 ms (amplitude > 100 uV, gradient > 75) were rejected and a 1 Hz high-pass forward filter was applied to reduce slow channel drift as well as to provide implicit baseline correction. EEG epochs, starting 200 ms prior to stimulus onset until 800 ms post-stimulus, were averaged for each emotion category. The resulting individual ERPs were transformed to an average reference and low-pass filtered at 30 Hz for ERP analysis and dipole source modeling.
A mean of 37 ± 9 (SD) clean trials were averaged for each emotion per participant. ERPscore  was used to extract the individual peak amplitude and latency values for each ERP component in the four emotion conditions. Time windows for scoring ERP peaks were determined from plots of global field power of individual participant averages, and amplitudes were measured from electrodes where the component was largest. The posterior component P1 was scored at occipital sites 76 (Oz), 72, and 77, and the anterior N1, VPP, and N2 components were measured from fronto-central sites 6 (FCz), 7, and 107. The N170 component was measured from bilateral temporal sites 57, 58 (T5), 64, and 96, 97 (T6), 101, and the P2 from bilateral parietal sites 59, 60 (P3), 66, and 85, 86 (P4), 92.
Our selected scalp regions for ERP analysis (see Figure 2) correspond to electrode sites commonly used in similar high-density ERP studies [17, 18, 39]. Instead of peak amplitude, average amplitude was used to measure the slower P2 and N2 components across seven 30-ms time windows between 220 and 430 ms . Repeated measures analyses of variance (ANOVA, using SPSS 14.0 for Windows) on ERP amplitudes and latencies were performed using Emotion (happy, angry, fearful, neutral), Hemisphere (left, right for N170 and P2 components), and Electrode (3) as within-subject factors and Age Group (10–13 years, 14–16 years) as the between-subject factor. Greenhouse-Geisser corrected degrees of freedom were used whenever the sphericity assumption was violated .
Dipole source analysis
Brain Electrical Source Analysis (BESA v.5.1, MEGIS Software GmbH, Munich, Germany) was used to model cortical sources of ERPs as equivalent spatiotemporal current dipoles with a certain orientation and time-varying dipole moments . Separate source models were created for the 10–13-year-olds and the 14–16-year-olds in realistic isotropic head models with conductivity ratio of 60 for the young adolescents and 80 for the older group [42–44]. Visual inspection of scalp topographies at latencies of peak ERP activity suggested that placing a symmetrical constraint on the location of bilateral pairs of sources was appropriate in order to limit the number of parameters in the inverse problem. Principal component analysis (PCA) decomposition showed that four pairs of dipole sources would sufficiently explain the majority of the variance in the ERP waveform.
Four pairs of sources were fit sequentially over time at latencies of the P1, N170/VPP, and P2 components, then adjusted in an iterative manner until they localized to stable locations and orientations, with the best temporal separation of source waveforms and as little interaction between sources as possible. The fit interval assigned to each pair of sources was defined such that the decomposition of the residual data (unexplained by previously fitted sources) was dominated by a single PCA component. Since preliminary source analysis on the grand average ERPs to each of the four different facial expressions did not show qualitative differences in source location between emotions, the source model for each age group was constructed based on the combined-conditions (all facial expressions) grand average ERP [45, 46]. The final source solutions required a residual variance of less than 10% [37, 47], i.e. a goodness of fit over 90%.
Significant main effects of Emotion on both accuracy [F(3,39) = 24.328, p < 0.001] and reaction time [F(3,39) = 39.004, p < 0.001] showed that happy faces were recognized most quickly (mean reaction time 761.5 ms) and accurately (mean 97.0% correct), while angry expressions took the longest to discriminate (mean reaction time 968.1 ms) and were more likely to be misclassified (mean 75.4% correct). The older group were overall more accurate (mean 92% correct) than the younger group (mean 83% correct) in identifying facial expressions [F(1,13) = 6.099, p = 0.028], but their mean reaction times did not differ significantly. A significant Emotion by Age Group interaction [F(3,39) = 7.339, p = 0.001] revealed that younger adolescents misclassified angry expressions more often than any of the other expressions (all p < 0.05) and recognized happy faces more easily than fearful faces (p = 0.024). Older adolescents were equally proficient at identifying happy, fearful, and neutral faces but also found angry expressions most difficult to discriminate (p = 0.018).
Figure 2 shows ERP waveforms averaged over all four expressions and electrodes of interest for the two age groups.
P1: The occipital P1 peaked at a mean latency of 122 ms in 10- to 13-year-olds and 114 ms in 14- to 16-year-olds, but effects of Emotion, Electrode, and Age Group on amplitude and latency did not reach statistical significance.
N1: The fronto-central N1, with mean latencies of 118 and 115 ms in the younger and older age groups respectively, was maximal over electrode FCz [F(2,28) = 3.568, p = 0.042] but did not show significant effects of Emotion or Age Group.
N170: There were no significant effects of Emotion, Hemisphere, Electrode, or Age Group on the N170 latency, which averaged 189 ms in 10- to 13-year-olds and 173 ms in 14- to 16-year-olds. All effects on N170 amplitude were also insignificant. Although Figure 3 shows a more widespread occipital N170 distribution in 10- to 13-year-olds and a more bilateral temporal distribution in 14- to 16-year-olds, this difference in N170 topography did not reach statistical significance (no significant Electrode by Age Group interaction was found when a larger group of temporal-occipital electrodes were analysed).
VPP: The VPP showed a trend of earlier latencies in the older adolescents compared to the younger ones [mean latency 175 ms versus 196 ms respectively, F(1,14) = 4.095, p = 0.063].
P2: No significant effects of Emotion, Hemisphere, Electrode, or Age Group were observed for the parietal P2 mean amplitudes across the seven 30-ms time windows between 220 and 430 ms. However, significant Electrode by Age Group interactions were found in time intervals 250–280 ms and 340–430 ms (p < 0.05) of the P2 wave, as illustrated by the age-related differences in P2–N2 topographies shown in Figure 3.
N2: The fronto-central N2 was also unaffected by Emotion or Electrode, and Age Group effects were present only in the last two time windows between 370–430 ms, reflecting the delayed N2 peak in young relative to older adolescents (see Figure 2).
Dipole sources of ERPs
Figure 4 illustrates the time-varying cortical activity that explains the scalp ERP components and the slightly different dipole locations and orientations within the three-dimensional head model of the two groups. Talairach coordinates of the dipole locations are listed in Tables 1 and 2, together with an estimation of their nearest corresponding anatomical structure and latencies of peak source activation. With the caveat that discrete equivalent current dipoles only represent idealized point sources on an active patch of cortex , each source waveform is likely to describe activity of the named structure and/or integrated activity from the 'center of mass' of several neighboring structures within a one centimeter radius. Nevertheless, our two source models explained 93% and 91% of variance in the group averages of 10- to 13-year-olds and 14- to 16-year-olds respectively within the analysis time window of 90 to 400 ms.
The posterior P1 scalp component in all participants was characterized by a pair of occipital sources localized in the lingual gyrus bilaterally. Activation of this first pair of sources in the visual cortex was particularly strong in the young adolescents at their P1 peak latency (122 ms), reflecting their larger P1 amplitude relative to the older group, whose dipoles were situated more superiorly. Within this early time frame, the 10- to 13-year-olds also activated sources in the frontopolar region near the superior frontal gyrus symmetrically, which helped to explain their relatively larger fronto-central N1 peak.
Next, a pair of sources in the temporal lobes around the fusiform gyrus accounted for the 10- to 13-year-olds' N170 just before 200 ms, and bilateral sources in the superior parietal lobe contributed to their P2 and N2 components around 300 ms. In contrast, the source model for 14- to 16-year-olds showed limbic activity originating near the parahippocampal gyrus at about 160 ms (N170) after the initial visual (P1) response in the lingual gyrus at 114 ms. Generators of the subsequent P2 and N2 components were estimated in the inferior temporal and prefrontal regions with greater source strengths seen in the right hemisphere than the left.
This study examined the precise temporal dynamics of adolescent neural processing in response to judging emotion from facial expressions and provided a first report of multiple dipole source analysis on healthy adolescent data. In 10- to 13-year-olds, sequential activation of occipital, prefrontal, fusiform, and parietal brain sources were observed, whereas in 14- to 16-year-olds, neural sources in the occipital, limbic, inferior temporal, and finally prefrontal region were successively activated in response to face processing. The similarity in source locations between our adolescent models and previous reports of dipole source analyses on adult data [11, 19, 49–51] suggest that cortical networks for face-specific processing are present before 16 years of age [52, 53], but that their functional specificity to different emotional expressions are not yet fully mature. Differences in brain sources and activation patterns between young and older adolescents detected in this study provide preliminary evidence for the continued development and maturation of distinct neural systems for processing emotional facial expressions during adolescence and possible changes in emotion perception, experience, and reaction with age.
Our behavioral data confirmed the improvement of facial emotion recognition with age , and that happiness is universally the most accurately recognized facial emotion . Angry expressions on the other hand, are likely perceived as more complex than other emotions  and hence took the longest to discriminate and were misclassified most often by adolescents, while neutral expressions might have been emotionally ambiguous. The lower accuracy rate in 10- to 13-year-olds could also substantiate the temporary performance dip in face encoding/recognition [56, 57] and cognitive/emotional processing  during this period of synaptic reorganization at the onset of puberty.
Our ERP results support the hypothesis that the P1 component reflects an initial response to visual stimuli and the early global/holistic processing of faces , while the N170 represents a subsequent encoding stage of configural and relational features within a face . Consistent with developmental ERP studies of facial expression processing [30, 32], we observed a decrease in ERP peak latencies as well as a reduction in the occipital P1 amplitude from the young to old adolescent groups, reflecting increasing cortical efficiency in face processing with age. The N170 voltage topography showed a shift from a widespread occipital negativity in 10- to 13-year-olds to a more focal lateral temporal distribution in 14- to 16-year-olds, providing support for a change in location and/or orientation of its dipole generator(s) as neural circuits mature and become more specialized for processing faces . Continuing maturation of the face-specific N170 through adolescence is compatible with previous work suggesting that between the ages of 10 and 13 years, two functionally distinct cortical generators of the N170a and N170b subcomponents in children begin to fuse into a single source responsible for the adult N170 . Developmental fMRI studies have also supported differences in the engagement and modulation of neural systems involved in face processing between adolescents and adults [59, 60]. Our finding that early ERP components in adolescents were insensitive to emotion agrees with a similar study  in which emotion effects were present only in adults but not in children. We presume that face-specific ERP components which seem to distinguish between emotional expressions in adults [16, 21, 24] are still immature in early adolescence, and that emotion effects and hemisphere asymmetry may only become visible when the adult N170 morphology develops fully sometime after the mid-teens . However, due to the limited sample size of this study and our attempt to explore developmental differences in ERPs and their sources during the adolescent period, significant effects of emotion and age may possibly have been masked by large inter-subject variability in brain responses during pubertal development. Further confirmatory studies using larger participant numbers and narrower age bands will be important for more in-depth examination of developmental changes in ERP sources as well as sex-related differences in brain responses to different facial emotions .
Using BESA, we constructed two equivalent current dipole models to explain ERP activity in the two adolescent age groups. Both models contained sources in the visual association cortex (lingual gyrus, LG) and around the inferior temporal/fusiform gyrus (FG) that explained the P1/N1 and N170/VPP dipolar complexes respectively. Occipito-temporal sources including the LG and FG have been previously identified in adult studies as dipole generators of face-sensitive ERPs [11, 19, 49–51] and MEG signals [62, 63]. Although the accuracy of source localization is limited to the spatial sampling of scalp signals as well as the statistical and geometrical properties of the brain and head model used , our source locations and activation time courses corresponded well with adult intracortical ERP recordings [65, 66], MEG sources [62, 63, 67], and fMRI activations [12, 13, 68, 69] to face stimuli. Halgren et al.  located an intracortical N75-P105 response (cf. our P1) to faces in the LG (visual areas 17 and 18), followed by an N130-P180-N240 complex (cf. our scalp N170-VPP-N2) in the FG (Brodmann areas 19 and 37) and the posterior superior and middle temporal gyrus. They also showed a wide transmission of the face-specific potential (P180) from the basal temporo-occipital cortex to the superior temporal sulcal, parietotemporal, and dorsolateral prefrontal cortices . With extensive reciprocal connections between temporal visual regions and the prefrontal cortex, it is postulated that activity in the temporal cortex can modulate input to the orbitofrontal cortex during facial emotion processing .
Indeed, adult studies using fMRI and PET have demonstrated significant involvement of frontal brain regions for explicit identification of facial emotion [13, 68–70]. Furthermore, adults have been found to produce greater activation in the orbitofrontal cortex than adolescents regardless of whether attention is directed to emotional or nonemotional aspects of a face . According to structural MRI studies, structures in the frontal lobe are still developing post-adolescence , as evidenced by significant loss of frontal gray matter due to synaptic pruning between adolescence and adulthood and continued axonal myelination that increases white matter density with age . We believe that structural changes in the prefrontal cortex during adolescent brain development may account for the difference in activation patterns of prefrontal sources observed in our two dipole models. Our adolescent source locations are consistent with a previous model of adult face encoding and recognition  that explained the VPP component by an orbitofrontal regional dipole source in the right hemisphere, together with bilateral dipole generators of the posterior P1 and P2 components in the parietooccipital cortex and a pair of posterior ventral sources accounting for the N170. Similarly, Sabbagh et al.  localized the frontal N270–400 elicited during mental state decoding from pictures of eyes to the orbitofrontal and medial frontal cortices in adults using low-resolution electromagnetic tomography (LORETA-KEY). They contended that the right orbitofrontal region is used for understanding others' mental states, while the left medial frontal regions are engaged in theory-of-mind reasoning about mental states . Strong right-lateralized activation of the frontopolar region in our 14- to 16-year-old adolescents likely reflects the conscious evaluation and judgment of emotion category [69, 70] and is also suggestive of recognition of previously seen faces . Recruitment of more mature frontal networks for social cognition and executive function in older adolescents  is expected to facilitate emotional processing and explains their superior performance in discriminating facial expressions compared to the younger group.
In addition to occipital, temporal, and prefrontal sources of ERPs, we also detected parahippocampal activity in the older adolescents around 160 ms. The parahippocampal gyrus has been speculated to be part of a ventro-medial pathway that receives input from the lingual and fusiform gyri and transmits perceptual information to the hippocampal and medial frontal regions for memory and face recognition [52, 75]. Limbic responses are known to be greater during explicit (emotion discrimination) rather than implicit (e.g. age discrimination) facial emotion processing . Moreover, since the amygdala plays such an important role in emotional responses and has structural and functional connections to the prefrontal and temporal cortices [77, 78], frontal and temporal source activity in our adolescent dipole models possibly reflected underlying amygdala activity triggered by emotional facial expressions.
Finally, we suspect that parietal source activity in young adolescents at roughly 280 ms post-stimulus could represent an engagement of the parietal somatosensory cortices when participants mentally simulated the perceived facial expression in order to achieve an emotional response for a better understanding of the emotion [5, 72, 79]. It has been shown that a single presentation of a face displaying an emotion for 500 ms is sufficient to elicit an emotional reaction in the observer . We speculate that expressions of happiness, sadness, and fear for example, are likely to have evoked the corresponding emotions, more so in the younger group than in the older adolescents, while angry faces may have triggered feelings of fear and disgust .
Although the small sample size in this study may have limited the power to detect significant effects of emotion and age on scalp ERP components and ERP brain source activity, our findings will form the basis for future developmental ERP studies employing spatiotemporal dipole source analysis on a wider age range of children, adolescents, and adults. In particular, frontal and parietal sources of later ERP components (those after 200–300 ms post-stimulus) deserve to be studied in greater detail, as emotion-specific activities are likely to occur at longer latencies in children and adolescents than in adults. Incorporating local maxima of fMRI activations from the same participants as constraints for ERP dipole sources can further improve localization accuracy and optimize the spatial and temporal resolution these complementary modalities provide . It would also be useful to extend this work to populations with social-emotional deficits or abnormalities in processing of faces and emotions, such as autism, depression, and schizophrenia, in order to examine differences in the engagement of neural systems and distinct perceptual/cognitive styles of facial emotion processing.
This is a first report of spatiotemporal dipole source analysis on healthy adolescent ERP data tracing the sequence of neural activity within the first 500 ms of categorizing emotion from faces. The similarity in source locations between our adolescent models and previous adult findings [11, 19] suggest that cortical networks for face processing are present before 16 years of age , but their functional specificity to different emotions are not yet fully mature. Differences in brain sources and activation patterns between young and older adolescents illustrate the continuing development and maturation of distinct neural systems for processing emotional facial expressions during adolescence and possible changes in emotion perception, experience, and reaction with age.
Baron-Cohen S: How to build a baby that can read minds: Cognitive mechanisms in mindreading. Curr Psychol Cogn. 1994, 13: 513-552.
Blair RJ: Facial expressions, their communicatory functions and neuro-cognitive substrates. Philos Trans R Soc Lond B Biol Sci. 2003, 358: 561-572. 10.1098/rstb.2002.1220.
Huttenlocher PR, Dabholkar AS: Regional differences in synaptogenesis in human cerebral cortex. J Comp Neurol. 1997, 387: 167-178. 10.1002/(SICI)1096-9861(19971020)387:2<167::AID-CNE1>3.0.CO;2-Z.
Giedd JN, Blumenthal J, Jeffries NO, Castellanos FX, Liu H, Zijdenbos A, Paus T, Evans AC, Rapoport JL: Brain development during childhood and adolescence: a longitudinal MRI study. Nat Neurosci. 1999, 2: 861-863. 10.1038/13158.
Adolphs R: Neural systems for recognizing emotion. Curr Opin Neurobiol. 2002, 12: 169-177. 10.1016/S0959-4388(02)00301-X.
Pelphrey K, Adolphs R, Morris JP: Neuroanatomical substrates of social cognition dysfunction in autism. Ment Retard Dev Disabil Res Rev. 2004, 10: 259-271. 10.1002/mrdd.20040.
De Renzi E: Prosopagnosia. Behavioral neurology and neuropsychology. Edited by: Feinberg TE, Farah MJ. 2003, New York: McGraw-Hill Medical Pub. Division, 245-255. 2
Bruce V, Young A: Understanding Face Recognition. Br J Psychol. 1986, 77: 305-327.
Haxby JV, Hoffman EA, Gobbini MI: The distributed human neural system for face perception. Trends Cogn Sci. 2000, 4: 223-233. 10.1016/S1364-6613(00)01482-0.
Halgren E, Raij T, Marinkovic K, Jousmaki V, Hari R: Cognitive response profile of the human fusiform face area as determined by MEG. Cereb Cortex. 2000, 10: 69-81. 10.1093/cercor/10.1.69.
Mnatsakanian EV, Tarkka IM: Familiar-face recognition and comparison: source analysis of scalp-recorded event-related potentials. Clin Neurophysiol. 2004, 115: 880-886. 10.1016/j.clinph.2003.11.027.
Critchley H, Daly E, Phillips M, Brammer M, Bullmore E, Williams S, Van Amelsvoort T, Robertson D, David A, Murphy D: Explicit and implicit neural mechanisms for processing of social information from facial expressions: A functional magnetic resonance imaging study. Hum Brain Mapp. 2000, 9: 93-105. 10.1002/(SICI)1097-0193(200002)9:2<93::AID-HBM4>3.0.CO;2-Z.
Kesler/West ML, Andersen AH, Smith CD, Avison MJ, Davis CE, Kryscio RJ, Blonder LX: Neural substrates of facial emotion processing using fMRI. Cogn Brain Res. 2001, 11: 213-226. 10.1016/S0926-6410(00)00073-2.
Puce A, Allison T, Bentin S, Gore JC, McCarthy G: Temporal cortex activation in humans viewing eye and mouth movements. J Neurosci. 1998, 18: 2188-2199.
Balconi M, Pozzili U: Face-selective processing and the effect of pleasant and unpleasant emotional expressions on ERP correlates. Int J Psychophysiol. 2003, 49: 67-74. 10.1016/S0167-8760(03)00081-3.
Sato W, Kochiyama T, Yoshikawa S, Matsumura M: Emotional expression boosts early visual processing of the face: ERP recording and its decomposition by independent component analysis. Neuroreport. 2001, 12: 709-714. 10.1097/00001756-200103260-00019.
O'Connor K, Hamm JP, Kirk IJ: The neurophysiological correlates of face processing in adults and children with Asperger's syndrome. Brain Cogn. 2005, 59: 82-95. 10.1016/j.bandc.2005.05.004.
Schupp HT, Ohman A, Junghofer M, Weike AI, Stockburger J, Hamm AO: The facilitated processing of threatening faces: an ERP analysis. Emotion. 2004, 4: 189-200. 10.1037/1528-35126.96.36.199.
Itier RJ, Taylor MJ: Inversion and contrast polarity reversal affect both encoding and recognition processes of unfamiliar faces: A repetition study using ERPs. Neuroimage. 2002, 15: 353-372. 10.1006/nimg.2001.0982.
Ashley V, Vuilleumier P, Swick D: Time course and specificity of event-related potentials to emotional expressions. Neuroreport. 2004, 15: 211-216. 10.1097/00001756-200401190-00041.
Eimer M, Holmes A: An ERP study on the time course of emotional face processing. Neuroreport. 2002, 13: 427-431. 10.1097/00001756-200203250-00013.
Eimer M, Holmes A, McGlone FP: The role of spatial attention in the processing of facial expression: an ERP study of rapid brain responses to six basic emotions. Cogn Affect Behav Neurosci. 2003, 3: 97-110. 10.3758/CABN.3.2.97.
Krolak-Salmon P, Fischer C, Vighetto A, Mauguiere F: Processing of facial emotional expression: spatio-temporal data as assessed by scalp event-related potentials. Eur J Neurosci. 2001, 13: 987-994. 10.1046/j.0953-816x.2001.01454.x.
Batty M, Taylor MJ: Early processing of the six basic facial emotional expressions. Brain Res Cogn Brain Res. 2003, 17: 613-620. 10.1016/S0926-6410(03)00174-5.
Caharel S, Courtay N, Bernard C, Lalonde R, Rebai M: Familiarity and emotional expression influence an early stage of face processing: An electrophysiological study. Brain Cognition. 2005, 59: 96-100. 10.1016/j.bandc.2005.05.005.
Rossion B, Joyce CA, Cottrell GW, Tarr MJ: Early lateralization and orientation tuning for face, word, and object processing in the visual cortex. Neuroimage. 2003, 20: 1609-1624. 10.1016/j.neuroimage.2003.07.010.
Dawson G, Webb SJ, Carver L, Panagiotides H, McPartland J: Young children with autism show atypical brain responses to fearful versus neutral facial expressions of emotion. Dev Sci. 2004, 7: 340-359. 10.1111/j.1467-7687.2004.00352.x.
Nelson CA, Nugent KM: Recognition Memory and Resource-Allocation as Revealed by Childrens Event-Related Potential Responses to Happy and Angry Faces. Dev Psychol. 1990, 26: 171-179. 10.1037/0012-16188.8.131.52.
Kestenbaum R, Nelson CA: Neural and behavioral correlates of emotion recognition in children and adults. J Exp Child Psychol. 1992, 54: 1-18. 10.1016/0022-0965(92)90014-W.
Batty M, Taylor MJ: The development of emotional face processing during childhood. Developmental Sci. 2006, 9: 207-220. 10.1111/j.1467-7687.2006.00480.x.
Taylor MJ: Non-spatial attentional effects on P1. Clin Neurophysiol. 2002, 113: 1903-1908. 10.1016/S1388-2457(02)00309-7.
Taylor MJ, Batty M, Itier RJ: The faces of development: A review of early face processing over childhood. J Cogn Neurosci. 2004, 16: 1426-1442. 10.1162/0898929042304732.
de Haan M, Johnson MH, Halit H: Development of face-sensitive event-related potentials during infancy: a review. Int J Psychophysiol. 2003, 51: 45-58. 10.1016/S0167-8760(03)00152-1.
Taylor MJ, Baldeweg T: Application of EEG, ERP and intracranial recordings to the investigation of cognitive functions in children. Developmental Sci. 2002, 5: 318-334. 10.1111/1467-7687.00372.
Ho TP, Leung PWL, Lee CC, Tang CP, Hung SF, Kwong SL, Lucas CP, Lieh-Mak F, Shaffer D: Test-retest reliability of the Chinese version of the Diagnostic Interview Schedule for Children-Version 4 (DISC-IV). J Child Psychol Psyc. 2005, 46: 1135-1138. 10.1111/j.1469-7610.2005.01435.x.
Ekman PMD: Japanese and Caucasian Facial Expressions of Emotion (JACFEE) and Neutral Faces (JACNeuF). 1988
Berg P, Scherg M: A Multiple Source Approach to the Correction of Eye Artifacts. Electroencephalogr Clin Neurophysiol. 1994, 90: 229-241. 10.1016/0013-4694(94)90094-9.
Segalowitz SJ: ERPScore program: Peak and area analysis of event-related potentials. St. Catharines, Ontario: Brock University (Available from author). 1999
McPartland J, Dawson G, Webb SJ, Panagiotides H, Carver LJ: Event-related brain potentials reveal anomalies in temporal processing of faces in autism spectrum disorder. J Child Psychol Psychiatry. 2004, 45: 1235-1245. 10.1111/j.1469-7610.2004.00318.x.
Picton TW, Bentin S, Berg P, Donchin E, Hillyard SA, Johnson R, Miller GA, Ritter W, Ruchkin DS, Rugg MD, Taylor MJ: Guidelines for using human event-related potentials to study cognition: Recording standards and publication criteria. Psychophysiology. 2000, 37: 127-152. 10.1017/S0048577200000305.
Scherg M, Berg P: New concepts of brain source imaging and localization. Electroencephalogr Clin Neurophysiol Suppl. 1996, 46: 127-137.
Lamm C, Zelazo PD, Lewis MD: Neural correlates of cognitive control in childhood and adolescence: Disentangling the contributions of age and executive function. Neuropsychologia. 2006, 44: 2139-2148. 10.1016/j.neuropsychologia.2005.10.013.
Lewis MD, Lamm C, Segalowitz SJ, Stieben J, Zelazo PD: Neurophysiological correlates of emotion regulation in children and adolescents. J Cogn Neurosci. 2006, 18: 430-443. 10.1162/jocn.2006.18.3.430.
Jonkman LM, Sniedt FLF, Kemner C: Source localization of the Nogo-N2: A developmental study. Clin Neurophysiol. 2007, 118: 1069-1077. 10.1016/j.clinph.2007.01.017.
Hoechstetter K, Rupp A, Meinck HM, Weckesser D, Bornfleth H, Stippich C, Berg P, Scherg M: Magnetic source imaging of tactile input shows task-independent attention effects in SII. Neuroreport. 2000, 11: 2461-2465. 10.1097/00001756-200008030-00024.
Hoechstetter K, Rupp A, Stancak A, Meinck HM, Stippich C, Berg P, Scherg M: Interaction of tactile input in the human primary and secondary somatosensory cortex – A magnetoencephalographic study. Neuroimage. 2001, 14: 759-767. 10.1006/nimg.2001.0855.
Tarkka IM, Mnatsakanian EV: Functional Specialization in the Human Frontal Cortex Observed during Task Anticipation. 2003, 48: 102-110.
Darvas F, Pantazis D, Kucukaltun-Yildirim E, Leahy RM: Mapping human brain function with MEG and EEG: methods and validation. Neuroimage. 2004, 23: S289-S299. 10.1016/j.neuroimage.2004.07.014.
Shibata T, Nishijo H, Tamura R, Miyamoto K, Eifuku S, Endo S, Ono T: Generators of visual evoked potentials for faces and eyes in the human brain as determined by dipole localization. Brain Topogr. 2002, 15: 51-63. 10.1023/A:1019944607316.
Caldara R, Thut G, Servoir P, Michel CM, Bovet P, Renault B: Face versus non-face object perception and the 'other-race' effect: a spatio-temporal event-related potential study. Clin Neurophysiol. 2003, 114: 515-528. 10.1016/S1388-2457(02)00407-8.
Werheid K, Alpay G, Jentzsch I, Sommer W: Priming emotional facial expressions as evidenced by event-related brain potentials. Int J Psychophysiol. 2005, 55: 209-219. 10.1016/j.ijpsycho.2004.07.006.
Sergent J, Ohta S, Macdonald B: Functional Neuroanatomy of Face and Object Processing – a Positron Emission Tomography Study. Brain. 1992, 115: 15-36. 10.1093/brain/115.1.15.
Lobaugh NJ, Gibson E, Taylor MJ: Children recruit distinct neural systems for implicit emotional face processing. Neuroreport. 2006, 17: 215-219. 10.1097/01.wnr.0000198946.00445.2f.
Gross AL, Ballif B: Childrens Understanding of Emotion from Facial Expressions and Situations – a Review. Dev Rev. 1991, 11: 368-398. 10.1016/0273-2297(91)90019-K.
Elfenbein HA, Ambady N: On the universality and cultural specificity of emotion recognition: A meta-analysis. Psychol Bull. 2002, 128: 203-235. 10.1037/0033-2909.128.2.203.
Carey S, Diamond R, Woods B: Developmental of face recognition – A maturational component?. Dev Psychol. 1980, 16: 257-269. 10.1037/0012-16184.108.40.2067.
Ellis HD: Developmental trends in face recognition. Psychologist. 1990, 3: 114-119.
McGivern RF, Andersen J, Byrd D, Mutter KL, Reilly J: Cognitive efficiency on a match to sample task decreases at the onset of puberty in children. Brain Cogn. 2002, 50: 73-89. 10.1016/S0278-2626(02)00012-X.
Monk CS, McClure EB, Nelson EE, Zarahn E, Bilder RM, Leibenluft E, Charney DS, Ernst M, Pine DS: Adolescent immaturity in attention-related brain engagement to emotional facial expressions. Neuroimage. 2003, 20: 420-428. 10.1016/S1053-8119(03)00355-0.
Passarotti AM, Paul BM, Russiere JR, Buxton RB, Wong EC, Stiles J: The development of face and location processing: an fMRI study. Dev Sci. 2003, 6: 100-117. 10.1111/1467-7687.00259.
Everhart DE, Shucard JL, Quatrin T, Shucard DW: Sex-related differences in event-related potentials, face recognition, and facial affect processing in prepubertal children. Neuropsychology. 2001, 15: 329-341. 10.1037/0894-4220.127.116.119.
Liu L, Ioannides AA, Streit M: Single trial analysis of neurophysiological correlates of the recognition of complex objects and facial expressions of emotion. Brain Topogr. 1999, 11: 291-303. 10.1023/A:1022258620435.
Watanabe S, Kakigi R, Koyama S, Kirino E: Human face perception traced by magneto- and electro-encephalography. Brain Res Cogn Brain Res. 1999, 8: 125-142. 10.1016/S0926-6410(99)00013-0.
Handy TC: Event-related potentials: a methods handbook. 2005, Cambridge, Mass: MIT Press
Halgren E, Baudena P, Heit G, Clarke M, Marinkovic K, Chauvel P: Spatiotemporal Stages in Face and Word-Processing .2. Depth-Recorded Potentials in the Human Frontal and Rolandic Cortices. J Physiol Paris. 1994, 88: 51-80. 10.1016/0928-4257(94)90093-0.
Ojemann JG, Ojemann GA, Lettich E: Neuronal-Activity Related to Faces and Matching in Human Right Nondominant Temporal Cortex. Brain. 1992, 115: 1-13. 10.1093/brain/115.1.1.
Lewis S, Thoma RJ, Lanoue MD, Miller GA, Heller W, Edgar C, Huang M, Weisend MP, Irwin J, Paulson K, Canive JM: Visual processing of facial affect. Neuroreport. 2003, 14: 1841-1845. 10.1097/00001756-200310060-00017.
Gorno-Tempini ML, Pradelli S, Serafini M, Pagnoni G, Baraldi P, Porro C, Nicoletti R, Umita C, Nichelli P: Explicit and incidental facial expression processing: An fMRI study. Neuroimage. 2001, 14: 465-473. 10.1006/nimg.2001.0811.
Narumoto J, Yamada H, Iidaka T, Sadato N, Fukui K, Itoh H, Yonekura Y: Brain regions involved in verbal or non-verbal aspects of facial emotion recognition. Neuroreport. 2000, 11: 2571-2576. 10.1097/00001756-200008030-00044.
Nakamura K, Kawashima R, Ito K, Sugiura M, Kato T, Nakamura A, Hatano K, Nagumo S, Kubota K, Fukuda H, Kojima S: Activation of the right inferior frontal cortex during assessment of facial emotion. J Neurophysiol. 1999, 82: 1610-1614.
Sowell ER, Thompson PM, Holmes CJ, Jernigan TL, Toga AW: In vivo evidence for post-adolescent brain maturation in frontal and striatal regions. Nat Neurosci. 1999, 2: 859-861. 10.1038/13154.
Blakemore SJ, Choudhury S: Development of the adolescent brain: implications for executive function and social cognition. J Child Psychol Psychiatry. 2006, 47: 296-312. 10.1111/j.1469-7610.2006.01611.x.
Sabbagh MA, Moulson MC, Harkness KL: Neural correlates of mental state decoding in human adults: An event-related potential study. J Cogn Neurosci. 2004, 16: 415-426. 10.1162/089892904322926755.
Wiser AK, Andreasen NC, O'Leary DS, Crespo-Facorro B, Boles-Ponto LL, Watkins GL, Hichwa RD: Novel vs. well-learned memory for faces: A positron emission tomography study. J Cogn Neurosci. 2000, 12: 255-266. 10.1162/089892900562084.
Elgar K, Campbell R: Annotation: The cognitive neuroscience of face recognition: Implications for developmental disorders. J Child Psychol Psychiatry. 2001, 42: 705-717. 10.1111/1469-7610.00767.
Gur RC, Schroeder L, Turner T, McGrath C, Chan RM, Turetsky BI, Alsop D, Maldjian J, Gur RE: Brain activation during facial emotion processing. Neuroimage. 2002, 16: 651-662. 10.1006/nimg.2002.1097.
Aggleton JP: The Amygdala: neurobiological aspects of emotion, memory, and mental dysfunction. 1992, New York: Wiley-Liss
Iidaka T, Omori M, Murata T, Kosaka H, Yonekura Y, Okada T, Sadato N: Neural interaction of the amygdala with the prefrontal and temporal cortices in the processing of facial expressions as revealed by fMRI. J Cogn Neurosci. 2001, 13: 1035-1047. 10.1162/089892901753294338.
Wild B, Erb M, Bartels M: Are emotions contagious? Evoked emotions while viewing emotionally expressive faces: quality, quantity, time course and gender differences. Psychiatry Res. 2001, 102: 109-124. 10.1016/S0165-1781(01)00225-6.
Horwitz B, Poeppel D: How can EEG/MEG and fMRI/PET data be combined?. Hum Brain Mapp. 2002, 17: 1-3. 10.1002/hbm.10057.
PCWF was supported by the NSFC/RGC grant no. N_HKU 703/03. The study was also supported by a University of Hong Kong Research grant to SEC. We are most grateful to Drs. Charlton Cheung and Vinci Cheung for assisting in data collection, Mr Howard Li for assisting with ERP scoring, and Prof. Pak Sham for statistical advice. We also thank Dr. Sid Segalowitz for supplying the ERPScore program, Dr. Karsten Hoechstetter for supporting BESA problems, and Drs. Margot Taylor and Peter Boord for their valuable comments on the manuscript. Last but not least, we warmly thank all the participants and their families for their contribution to this study.
The authors declare that they have no competing interests.
SEC, GMM and PCWF conceived of, acquired funds for, and coordinated the study. TKWW acquired, processed, and analyzed the data, as well as drafted the manuscript. All authors were involved in the design of the paradigm, interpretation of data, and revision of the manuscript. All authors read and approved the final manuscript.
About this article
Cite this article
Wong, T.K.W., Fung, P.C.W., McAlonan, G.M. et al. Spatiotemporal dipole source localization of face processing ERPs in adolescents: a preliminary study. Behav Brain Funct 5, 16 (2009). https://doi.org/10.1186/1744-9081-5-16
- Facial Expression
- Face Processing
- Facial Emotion
- Dipole Source
- Emotional Facial Expression