- Open Access
Spatiotemporal dipole source localization of face processing ERPs in adolescents: a preliminary study
© Wong et al; licensee BioMed Central Ltd. 2009
- Received: 16 June 2008
- Accepted: 12 March 2009
- Published: 12 March 2009
Despite extensive investigation of the neural systems for face perception and emotion recognition in adults and young children in the past, the precise temporal activation of brain sources specific to the processing of emotional facial expressions in older children and adolescents is not well known. This preliminary study aims to trace the spatiotemporal dynamics of facial emotion processing during adolescence and provide a basis for future developmental studies and comparisons with patient populations that have social-emotional deficits such as autism.
We presented pictures showing happy, angry, fearful, or neutral facial expressions to healthy adolescents (aged 10–16 years) and recorded 128-channel event-related potentials (ERPs) while they performed an emotion discrimination task. ERP components were analyzed for effects of age and emotion on amplitude and latency. The underlying cortical sources of scalp ERP activity were modeled as multiple equivalent current dipoles using Brain Electrical Source Analysis (BESA).
Initial global/holistic processing of faces (P1) took place in the visual association cortex (lingual gyrus) around 120 ms post-stimulus. Next, structural encoding of facial features (N170) occurred between 160–200 ms in the inferior temporal/fusiform region, and perhaps early emotion processing (Vertex Positive Potential or VPP) in the amygdala and orbitofrontal cortex. Finally, cognitive analysis of facial expressions (P2) in the prefrontal cortex and emotional reactions in somatosensory areas were observed from about 230 ms onwards. The temporal sequence of cortical source activation in response to facial emotion processing was occipital, prefrontal, fusiform, parietal for young adolescents and occipital, limbic, inferior temporal, and prefrontal for older adolescents.
This is a first report of high-density ERP dipole source analysis in healthy adolescents which traces the sequence of neural activity within the first 500 ms of categorizing emotion from faces. Our spatio-temporal brain source models showed the presence of adult-like cortical networks for face processing in adolescents, whose functional specificity to different emotions appear to be not yet fully mature. Age-related differences in brain activation patterns illustrate the continued development and maturation of distinct neural systems for processing facial expressions during adolescence and possible changes in emotion perception, experience, and reaction with age.
- Facial Expression
- Face Processing
- Facial Emotion
- Dipole Source
- Emotional Facial Expression
Discerning facial emotions is fundamental for normal social interaction and commences in early infancy . Recognition and interpretation of emotion from facial expressions occurs almost instantaneously and is used in social referencing, for example, when an infant first encounters a novel object and looks towards the parent to direct their behavior. If the parent expresses a happy face, the child will understand this to be an encouragement to approach the object . On the other hand, if the parent displays an expression of fear or disgust, the child will tend to avoid the same object. This skill of "reading" emotions from faces becomes more proficient as children grow up and experience a multitude of social, emotional and cognitive changes. However, the neural circuitry that controls facial emotion processing during late childhood and adolescence has received scant attention to date . This period of cognitive and behavioral development is critical since heterochronous brain changes take place: cortical synaptic elimination occurs in the primary visual and auditory cortices during late childhood, then subsequently in the prefrontal cortex in mid-adolescence . Compatible with this is the region-specific nonlinear increase in cortical grey matter volume, which peaks in the frontal and parietal lobes around age 12, followed by the temporal and occipital lobes at about ages 16 and 20 respectively .
Neurobiological findings from lesion, functional magnetic resonance imaging (fMRI) and positron emission tomography (PET) studies have revealed that, just as distinct brain structures subserve verbal language, specialized neural networks also exist for processing the non-verbal language of facial expressions. These include the occipito-temporal cortices, amygdala, orbitofrontal cortex, as well as somatosensory-related cortices in the right hemisphere . Studies of pervasive developmental disorders such as autism have also identified abnormalities in the fusiform gyrus, superior temporal sulcus, and amygdale , suggesting that the innate inability to perceive and respond to complex emotional expressions of others may parallel early abnormal development of brain regions involved in face processing. Prosopagnosic patients who demonstrate impairments in recognizing familiar faces, but have normal abilities in identifying facial expressions of emotion , suggest a functional dissociation between facial identity and facial expression processing . In accordance with Bruce and Young's cognitive model for face perception , Haxby, Hoffman, and Gobbini's neuroanatomical model  posits that, after the initial stage of face perception has taken place in the inferior occipital gyri, (bidirectional) information proceeds along two distinct parallel neural pathways: one for coding facial identity (lateral fusiform gyrus for recognizing unique invariant facial features), and the other for coding facial expression (superior temporal sulcus region for analyzing the changeable aspects of faces e.g. eye gaze and lip movement). The models of Bruce and Young  and Haxby et al.  however, do not include the precise time frame in which the various components of face processing take place.
Such time-sensitive issues can be addressed using dense-array electroencephalography (EEG), event-related potentials (ERPs), and magnetoencephalography (MEG), since these techniques provide millisecond temporal resolution not accessible by hemodynamic measures. Recent work has localized sources of face-specific components to the occipital and inferotemporal cortices [10, 11], which overlap with regions identified in fMRI studies [12–14]. For example, an equivalent current dipole model for adult face recognition showed the following temporal dynamics of ERP source activation: bilateral lingual gyri at 120 ms, followed by the right and then left fusiform/hippocampal gyri around 150–170 ms, bilateral medial temporal gyri from 200 ms onwards, the right caudate nucleus peaking at 300 ms, and finally the anterior cingulate gyrus 300 ms after stimulus onset. However, such source localization has not yet been performed on ERP data from children and adolescents actively processing facial expressions of emotion, and only few studies to date have examined differential neural processing of emotional facial expressions in children [15, 16] and adolescents , let alone using high-density EEG.
Adult studies have suggested that emotionally expressive faces (especially high-arousal negative emotions like fear or anger) tend to evoke larger ERP responses than emotionally unexpressive (neutral) faces [15, 16, 18]. The "face-sensitive" N170 component over posterior temporal scalp regions is said to reflect the encoding of configural and relational features within a face  and appears not to be selective for any particular facial expression [17, 20–23]. However, others have found that negative (fear and disgust) emotions do indeed evoke larger N170 amplitudes in adults than positive (happy, surprise, neutral) ones [24, 25]. Similarly, the vertex positive potential (VPP), originating from the same pair of dipoles in the lateral inferior occipital cortex/posterior fusiform gyrus that generates the N170 in adults , is enhanced by facial expressions of fear [20, 21]. In addition, the timing of the neural response in adults seems also affected by emotion, such that the N170 appears at earlier latencies for positive than negative emotions. Krolak-Salmon et al.  reported that occipital ERP patterns first distinguished between emotional and neutral faces during the 250–550 ms time interval, and then differed between emotional expressions (fear and happiness in particular) from 550–750 ms over right occipito-temporal scalp regions. Activity specific to expressions of disgust occurred even later, between 700–900 ms in frontal and right temporal areas .
Several studies of young children (3–7 years of age) have shown differential ERP activity to facial emotion at latencies longer than about 300 ms [27–29], but earlier emotion effects on the face-elicited P1, N170, or P2 components as seen in adults were not present in children between 9–15 years . In contrast, Batty and Taylor  detected an effect of emotion on the P1 latency in children as young as four years old. This early emotion effect, indexing rapid holistic processing of faces , however, disappeared with age (absent after 7 years old) along with age-related decreases in P1 amplitude and latency. By mid-teens, global processing of emotions was replaced by detailed configural processing, reflected by an emotion effect on the N170 amplitude in 14- to 15-year-olds [19, 30]. This change in the spatio-temporal profile of face-processing ERPs with development  suggests that as cortical circuits mature and become more specialized for processing faces, dipole generators of ERP components may shift in location or orientation from childhood to adolescence . For example, it is thought that two functionally distinct cortical generators of the N170a and N170b subcomponents in children are fused into a single source responsible for the adult N170 . One model of adult face encoding and recognition  explained the posterior P1 and P2 components with bilateral dipole generators in the parietooccipital cortex, the N170 component by a pair of posterior ventral dipoles, and the VPP component by an orbitofrontal regional dipole source in the right hemisphere. However, no study has yet investigated the dipole source location and time course of processing facial expressions of emotion in the adolescent years.
The main purpose of this study was to construct a dipole model of brain sources for healthy adolescents to show the temporal unfolding of neural events that occur during explicit processing of different facial emotions. In addition, we examined the effects of emotion on ERP amplitudes, latencies, and scalp distribution, as well as developmental changes in ERPs to faces from early (age 10 to 13 years) to late (age 14 to 16 years) adolescence. We hypothesized that facial-emotional processing would occur rapidly in time, sequentially activating brain regions similar to those seen in adults subserving face detection, structural encoding of facial features, and emotion recognition.
Twenty healthy right-handed Chinese adolescents (12 male, 8 female) aged between 10 and 16 years (mean 14.2 ± 1.8 years) took part in this study. In order to examine developmental changes in face-specific ERPs and their sources , participants were divided into two groups: 10 young adolescents aged 10–13 years, and 10 older adolescents aged 14–16 years. However, due to severe artifact contamination in four young participants, we analysed 6 datasets (3 male) in the young group and 10 datasets (5 male) in the old group.
The research protocol was approved by the Institutional Review Board of the hospital where the study took place and all participants were compensated for travel and inconvenience. Parents of participating children signed informed consent and every participant gave their assent to participate in the project. All were screened for psychiatric illness using the parental Chinese Diagnostic Interview Schedule for Children for DSMIV  and had no history of headache, loss of consciousness, nor any neurological disorder. All attended normal school and had normal or corrected-to-normal vision.
Stimuli and procedure
EEG recording and signal processing
Scalp EEG was recorded using a 128-channel Geodesic Sensor Net (Electrical Geodesics, Inc., Eugene, OR, USA) that matched the head size of the participant. Electrode impedances were adjusted to stabilize under 50 kΩ before recording, and continuous EEG was sampled at 500 Hz with filters 0.1 to 100 Hz using the vertex (electrode 129) as reference. Offline, notch filters were employed to remove electrical noise from the mains power supply (50 Hz) and the stimulus display monitor (60 Hz refresh rate). Subsequent data processing was done using Brain Electrical Source Analysis (BESA 5.1, MEGIS Software GmbH, Munich, Germany). Eye-blink patterns were marked and averaged for each individual and the spatial topography of the blink artifact was used for correction of individual EEG data by the surrogate model approach . The surrogate model for brain activity consisted of 15 regional sources covering all brain regions. After blink correction, trials with remaining artifacts between -100 ms and 500 ms (amplitude > 100 uV, gradient > 75) were rejected and a 1 Hz high-pass forward filter was applied to reduce slow channel drift as well as to provide implicit baseline correction. EEG epochs, starting 200 ms prior to stimulus onset until 800 ms post-stimulus, were averaged for each emotion category. The resulting individual ERPs were transformed to an average reference and low-pass filtered at 30 Hz for ERP analysis and dipole source modeling.
A mean of 37 ± 9 (SD) clean trials were averaged for each emotion per participant. ERPscore  was used to extract the individual peak amplitude and latency values for each ERP component in the four emotion conditions. Time windows for scoring ERP peaks were determined from plots of global field power of individual participant averages, and amplitudes were measured from electrodes where the component was largest. The posterior component P1 was scored at occipital sites 76 (Oz), 72, and 77, and the anterior N1, VPP, and N2 components were measured from fronto-central sites 6 (FCz), 7, and 107. The N170 component was measured from bilateral temporal sites 57, 58 (T5), 64, and 96, 97 (T6), 101, and the P2 from bilateral parietal sites 59, 60 (P3), 66, and 85, 86 (P4), 92.
Dipole source analysis
Brain Electrical Source Analysis (BESA v.5.1, MEGIS Software GmbH, Munich, Germany) was used to model cortical sources of ERPs as equivalent spatiotemporal current dipoles with a certain orientation and time-varying dipole moments . Separate source models were created for the 10–13-year-olds and the 14–16-year-olds in realistic isotropic head models with conductivity ratio of 60 for the young adolescents and 80 for the older group [42–44]. Visual inspection of scalp topographies at latencies of peak ERP activity suggested that placing a symmetrical constraint on the location of bilateral pairs of sources was appropriate in order to limit the number of parameters in the inverse problem. Principal component analysis (PCA) decomposition showed that four pairs of dipole sources would sufficiently explain the majority of the variance in the ERP waveform.
Four pairs of sources were fit sequentially over time at latencies of the P1, N170/VPP, and P2 components, then adjusted in an iterative manner until they localized to stable locations and orientations, with the best temporal separation of source waveforms and as little interaction between sources as possible. The fit interval assigned to each pair of sources was defined such that the decomposition of the residual data (unexplained by previously fitted sources) was dominated by a single PCA component. Since preliminary source analysis on the grand average ERPs to each of the four different facial expressions did not show qualitative differences in source location between emotions, the source model for each age group was constructed based on the combined-conditions (all facial expressions) grand average ERP [45, 46]. The final source solutions required a residual variance of less than 10% [37, 47], i.e. a goodness of fit over 90%.
Significant main effects of Emotion on both accuracy [F(3,39) = 24.328, p < 0.001] and reaction time [F(3,39) = 39.004, p < 0.001] showed that happy faces were recognized most quickly (mean reaction time 761.5 ms) and accurately (mean 97.0% correct), while angry expressions took the longest to discriminate (mean reaction time 968.1 ms) and were more likely to be misclassified (mean 75.4% correct). The older group were overall more accurate (mean 92% correct) than the younger group (mean 83% correct) in identifying facial expressions [F(1,13) = 6.099, p = 0.028], but their mean reaction times did not differ significantly. A significant Emotion by Age Group interaction [F(3,39) = 7.339, p = 0.001] revealed that younger adolescents misclassified angry expressions more often than any of the other expressions (all p < 0.05) and recognized happy faces more easily than fearful faces (p = 0.024). Older adolescents were equally proficient at identifying happy, fearful, and neutral faces but also found angry expressions most difficult to discriminate (p = 0.018).
Figure 2 shows ERP waveforms averaged over all four expressions and electrodes of interest for the two age groups.
P1: The occipital P1 peaked at a mean latency of 122 ms in 10- to 13-year-olds and 114 ms in 14- to 16-year-olds, but effects of Emotion, Electrode, and Age Group on amplitude and latency did not reach statistical significance.
N1: The fronto-central N1, with mean latencies of 118 and 115 ms in the younger and older age groups respectively, was maximal over electrode FCz [F(2,28) = 3.568, p = 0.042] but did not show significant effects of Emotion or Age Group.
VPP: The VPP showed a trend of earlier latencies in the older adolescents compared to the younger ones [mean latency 175 ms versus 196 ms respectively, F(1,14) = 4.095, p = 0.063].
P2: No significant effects of Emotion, Hemisphere, Electrode, or Age Group were observed for the parietal P2 mean amplitudes across the seven 30-ms time windows between 220 and 430 ms. However, significant Electrode by Age Group interactions were found in time intervals 250–280 ms and 340–430 ms (p < 0.05) of the P2 wave, as illustrated by the age-related differences in P2–N2 topographies shown in Figure 3.
N2: The fronto-central N2 was also unaffected by Emotion or Electrode, and Age Group effects were present only in the last two time windows between 370–430 ms, reflecting the delayed N2 peak in young relative to older adolescents (see Figure 2).
Dipole sources of ERPs
Dipole locations and latencies of maximum dipole moments in source model for 10- to 13-year-olds
Talairach coordinates (mm)
Nearest anatomical structure
Dipole moment (nAm)
Right lingual gyrus (BA18)
Left lingual gyrus (BA18)
Right superior frontal gyrus (BA10)
Left superior frontal gyrus (BA10)
Right fusiform gyrus (BA37)
Left fusiform gyrus (BA37)
Right paracentral lobule (BA5)
Left paracentral lobule (BA5)
Dipole locations and latencies of maximum dipole moments in source model for 14- to 16-year-olds
Talairach coordinates (mm)
Nearest anatomical structure
Dipole moment (nAm)
Right lingual gyrus (BA18)
Left lingual gyrus (BA18)
Right parahippocampal gyrus (BA19)
Left parahippocampal gyrus (BA19)
Right inferior temporal gyrus (BA20)
Left inferior temporal gyrus (BA20)
Right superior frontal lobule (BA10)
Left superior frontal lobule (BA10)
The posterior P1 scalp component in all participants was characterized by a pair of occipital sources localized in the lingual gyrus bilaterally. Activation of this first pair of sources in the visual cortex was particularly strong in the young adolescents at their P1 peak latency (122 ms), reflecting their larger P1 amplitude relative to the older group, whose dipoles were situated more superiorly. Within this early time frame, the 10- to 13-year-olds also activated sources in the frontopolar region near the superior frontal gyrus symmetrically, which helped to explain their relatively larger fronto-central N1 peak.
Next, a pair of sources in the temporal lobes around the fusiform gyrus accounted for the 10- to 13-year-olds' N170 just before 200 ms, and bilateral sources in the superior parietal lobe contributed to their P2 and N2 components around 300 ms. In contrast, the source model for 14- to 16-year-olds showed limbic activity originating near the parahippocampal gyrus at about 160 ms (N170) after the initial visual (P1) response in the lingual gyrus at 114 ms. Generators of the subsequent P2 and N2 components were estimated in the inferior temporal and prefrontal regions with greater source strengths seen in the right hemisphere than the left.
This study examined the precise temporal dynamics of adolescent neural processing in response to judging emotion from facial expressions and provided a first report of multiple dipole source analysis on healthy adolescent data. In 10- to 13-year-olds, sequential activation of occipital, prefrontal, fusiform, and parietal brain sources were observed, whereas in 14- to 16-year-olds, neural sources in the occipital, limbic, inferior temporal, and finally prefrontal region were successively activated in response to face processing. The similarity in source locations between our adolescent models and previous reports of dipole source analyses on adult data [11, 19, 49–51] suggest that cortical networks for face-specific processing are present before 16 years of age [52, 53], but that their functional specificity to different emotional expressions are not yet fully mature. Differences in brain sources and activation patterns between young and older adolescents detected in this study provide preliminary evidence for the continued development and maturation of distinct neural systems for processing emotional facial expressions during adolescence and possible changes in emotion perception, experience, and reaction with age.
Our behavioral data confirmed the improvement of facial emotion recognition with age , and that happiness is universally the most accurately recognized facial emotion . Angry expressions on the other hand, are likely perceived as more complex than other emotions  and hence took the longest to discriminate and were misclassified most often by adolescents, while neutral expressions might have been emotionally ambiguous. The lower accuracy rate in 10- to 13-year-olds could also substantiate the temporary performance dip in face encoding/recognition [56, 57] and cognitive/emotional processing  during this period of synaptic reorganization at the onset of puberty.
Our ERP results support the hypothesis that the P1 component reflects an initial response to visual stimuli and the early global/holistic processing of faces , while the N170 represents a subsequent encoding stage of configural and relational features within a face . Consistent with developmental ERP studies of facial expression processing [30, 32], we observed a decrease in ERP peak latencies as well as a reduction in the occipital P1 amplitude from the young to old adolescent groups, reflecting increasing cortical efficiency in face processing with age. The N170 voltage topography showed a shift from a widespread occipital negativity in 10- to 13-year-olds to a more focal lateral temporal distribution in 14- to 16-year-olds, providing support for a change in location and/or orientation of its dipole generator(s) as neural circuits mature and become more specialized for processing faces . Continuing maturation of the face-specific N170 through adolescence is compatible with previous work suggesting that between the ages of 10 and 13 years, two functionally distinct cortical generators of the N170a and N170b subcomponents in children begin to fuse into a single source responsible for the adult N170 . Developmental fMRI studies have also supported differences in the engagement and modulation of neural systems involved in face processing between adolescents and adults [59, 60]. Our finding that early ERP components in adolescents were insensitive to emotion agrees with a similar study  in which emotion effects were present only in adults but not in children. We presume that face-specific ERP components which seem to distinguish between emotional expressions in adults [16, 21, 24] are still immature in early adolescence, and that emotion effects and hemisphere asymmetry may only become visible when the adult N170 morphology develops fully sometime after the mid-teens . However, due to the limited sample size of this study and our attempt to explore developmental differences in ERPs and their sources during the adolescent period, significant effects of emotion and age may possibly have been masked by large inter-subject variability in brain responses during pubertal development. Further confirmatory studies using larger participant numbers and narrower age bands will be important for more in-depth examination of developmental changes in ERP sources as well as sex-related differences in brain responses to different facial emotions .
Using BESA, we constructed two equivalent current dipole models to explain ERP activity in the two adolescent age groups. Both models contained sources in the visual association cortex (lingual gyrus, LG) and around the inferior temporal/fusiform gyrus (FG) that explained the P1/N1 and N170/VPP dipolar complexes respectively. Occipito-temporal sources including the LG and FG have been previously identified in adult studies as dipole generators of face-sensitive ERPs [11, 19, 49–51] and MEG signals [62, 63]. Although the accuracy of source localization is limited to the spatial sampling of scalp signals as well as the statistical and geometrical properties of the brain and head model used , our source locations and activation time courses corresponded well with adult intracortical ERP recordings [65, 66], MEG sources [62, 63, 67], and fMRI activations [12, 13, 68, 69] to face stimuli. Halgren et al.  located an intracortical N75-P105 response (cf. our P1) to faces in the LG (visual areas 17 and 18), followed by an N130-P180-N240 complex (cf. our scalp N170-VPP-N2) in the FG (Brodmann areas 19 and 37) and the posterior superior and middle temporal gyrus. They also showed a wide transmission of the face-specific potential (P180) from the basal temporo-occipital cortex to the superior temporal sulcal, parietotemporal, and dorsolateral prefrontal cortices . With extensive reciprocal connections between temporal visual regions and the prefrontal cortex, it is postulated that activity in the temporal cortex can modulate input to the orbitofrontal cortex during facial emotion processing .
Indeed, adult studies using fMRI and PET have demonstrated significant involvement of frontal brain regions for explicit identification of facial emotion [13, 68–70]. Furthermore, adults have been found to produce greater activation in the orbitofrontal cortex than adolescents regardless of whether attention is directed to emotional or nonemotional aspects of a face . According to structural MRI studies, structures in the frontal lobe are still developing post-adolescence , as evidenced by significant loss of frontal gray matter due to synaptic pruning between adolescence and adulthood and continued axonal myelination that increases white matter density with age . We believe that structural changes in the prefrontal cortex during adolescent brain development may account for the difference in activation patterns of prefrontal sources observed in our two dipole models. Our adolescent source locations are consistent with a previous model of adult face encoding and recognition  that explained the VPP component by an orbitofrontal regional dipole source in the right hemisphere, together with bilateral dipole generators of the posterior P1 and P2 components in the parietooccipital cortex and a pair of posterior ventral sources accounting for the N170. Similarly, Sabbagh et al.  localized the frontal N270–400 elicited during mental state decoding from pictures of eyes to the orbitofrontal and medial frontal cortices in adults using low-resolution electromagnetic tomography (LORETA-KEY). They contended that the right orbitofrontal region is used for understanding others' mental states, while the left medial frontal regions are engaged in theory-of-mind reasoning about mental states . Strong right-lateralized activation of the frontopolar region in our 14- to 16-year-old adolescents likely reflects the conscious evaluation and judgment of emotion category [69, 70] and is also suggestive of recognition of previously seen faces . Recruitment of more mature frontal networks for social cognition and executive function in older adolescents  is expected to facilitate emotional processing and explains their superior performance in discriminating facial expressions compared to the younger group.
In addition to occipital, temporal, and prefrontal sources of ERPs, we also detected parahippocampal activity in the older adolescents around 160 ms. The parahippocampal gyrus has been speculated to be part of a ventro-medial pathway that receives input from the lingual and fusiform gyri and transmits perceptual information to the hippocampal and medial frontal regions for memory and face recognition [52, 75]. Limbic responses are known to be greater during explicit (emotion discrimination) rather than implicit (e.g. age discrimination) facial emotion processing . Moreover, since the amygdala plays such an important role in emotional responses and has structural and functional connections to the prefrontal and temporal cortices [77, 78], frontal and temporal source activity in our adolescent dipole models possibly reflected underlying amygdala activity triggered by emotional facial expressions.
Finally, we suspect that parietal source activity in young adolescents at roughly 280 ms post-stimulus could represent an engagement of the parietal somatosensory cortices when participants mentally simulated the perceived facial expression in order to achieve an emotional response for a better understanding of the emotion [5, 72, 79]. It has been shown that a single presentation of a face displaying an emotion for 500 ms is sufficient to elicit an emotional reaction in the observer . We speculate that expressions of happiness, sadness, and fear for example, are likely to have evoked the corresponding emotions, more so in the younger group than in the older adolescents, while angry faces may have triggered feelings of fear and disgust .
Although the small sample size in this study may have limited the power to detect significant effects of emotion and age on scalp ERP components and ERP brain source activity, our findings will form the basis for future developmental ERP studies employing spatiotemporal dipole source analysis on a wider age range of children, adolescents, and adults. In particular, frontal and parietal sources of later ERP components (those after 200–300 ms post-stimulus) deserve to be studied in greater detail, as emotion-specific activities are likely to occur at longer latencies in children and adolescents than in adults. Incorporating local maxima of fMRI activations from the same participants as constraints for ERP dipole sources can further improve localization accuracy and optimize the spatial and temporal resolution these complementary modalities provide . It would also be useful to extend this work to populations with social-emotional deficits or abnormalities in processing of faces and emotions, such as autism, depression, and schizophrenia, in order to examine differences in the engagement of neural systems and distinct perceptual/cognitive styles of facial emotion processing.
This is a first report of spatiotemporal dipole source analysis on healthy adolescent ERP data tracing the sequence of neural activity within the first 500 ms of categorizing emotion from faces. The similarity in source locations between our adolescent models and previous adult findings [11, 19] suggest that cortical networks for face processing are present before 16 years of age , but their functional specificity to different emotions are not yet fully mature. Differences in brain sources and activation patterns between young and older adolescents illustrate the continuing development and maturation of distinct neural systems for processing emotional facial expressions during adolescence and possible changes in emotion perception, experience, and reaction with age.
PCWF was supported by the NSFC/RGC grant no. N_HKU 703/03. The study was also supported by a University of Hong Kong Research grant to SEC. We are most grateful to Drs. Charlton Cheung and Vinci Cheung for assisting in data collection, Mr Howard Li for assisting with ERP scoring, and Prof. Pak Sham for statistical advice. We also thank Dr. Sid Segalowitz for supplying the ERPScore program, Dr. Karsten Hoechstetter for supporting BESA problems, and Drs. Margot Taylor and Peter Boord for their valuable comments on the manuscript. Last but not least, we warmly thank all the participants and their families for their contribution to this study.
- Baron-Cohen S: How to build a baby that can read minds: Cognitive mechanisms in mindreading. Curr Psychol Cogn. 1994, 13: 513-552.Google Scholar
- Blair RJ: Facial expressions, their communicatory functions and neuro-cognitive substrates. Philos Trans R Soc Lond B Biol Sci. 2003, 358: 561-572. 10.1098/rstb.2002.1220.PubMed CentralView ArticlePubMedGoogle Scholar
- Huttenlocher PR, Dabholkar AS: Regional differences in synaptogenesis in human cerebral cortex. J Comp Neurol. 1997, 387: 167-178. 10.1002/(SICI)1096-9861(19971020)387:2<167::AID-CNE1>3.0.CO;2-Z.View ArticlePubMedGoogle Scholar
- Giedd JN, Blumenthal J, Jeffries NO, Castellanos FX, Liu H, Zijdenbos A, Paus T, Evans AC, Rapoport JL: Brain development during childhood and adolescence: a longitudinal MRI study. Nat Neurosci. 1999, 2: 861-863. 10.1038/13158.View ArticlePubMedGoogle Scholar
- Adolphs R: Neural systems for recognizing emotion. Curr Opin Neurobiol. 2002, 12: 169-177. 10.1016/S0959-4388(02)00301-X.View ArticlePubMedGoogle Scholar
- Pelphrey K, Adolphs R, Morris JP: Neuroanatomical substrates of social cognition dysfunction in autism. Ment Retard Dev Disabil Res Rev. 2004, 10: 259-271. 10.1002/mrdd.20040.View ArticlePubMedGoogle Scholar
- De Renzi E: Prosopagnosia. Behavioral neurology and neuropsychology. Edited by: Feinberg TE, Farah MJ. 2003, New York: McGraw-Hill Medical Pub. Division, 245-255. 2Google Scholar
- Bruce V, Young A: Understanding Face Recognition. Br J Psychol. 1986, 77: 305-327.View ArticlePubMedGoogle Scholar
- Haxby JV, Hoffman EA, Gobbini MI: The distributed human neural system for face perception. Trends Cogn Sci. 2000, 4: 223-233. 10.1016/S1364-6613(00)01482-0.View ArticlePubMedGoogle Scholar
- Halgren E, Raij T, Marinkovic K, Jousmaki V, Hari R: Cognitive response profile of the human fusiform face area as determined by MEG. Cereb Cortex. 2000, 10: 69-81. 10.1093/cercor/10.1.69.View ArticlePubMedGoogle Scholar
- Mnatsakanian EV, Tarkka IM: Familiar-face recognition and comparison: source analysis of scalp-recorded event-related potentials. Clin Neurophysiol. 2004, 115: 880-886. 10.1016/j.clinph.2003.11.027.View ArticlePubMedGoogle Scholar
- Critchley H, Daly E, Phillips M, Brammer M, Bullmore E, Williams S, Van Amelsvoort T, Robertson D, David A, Murphy D: Explicit and implicit neural mechanisms for processing of social information from facial expressions: A functional magnetic resonance imaging study. Hum Brain Mapp. 2000, 9: 93-105. 10.1002/(SICI)1097-0193(200002)9:2<93::AID-HBM4>3.0.CO;2-Z.View ArticlePubMedGoogle Scholar
- Kesler/West ML, Andersen AH, Smith CD, Avison MJ, Davis CE, Kryscio RJ, Blonder LX: Neural substrates of facial emotion processing using fMRI. Cogn Brain Res. 2001, 11: 213-226. 10.1016/S0926-6410(00)00073-2.View ArticleGoogle Scholar
- Puce A, Allison T, Bentin S, Gore JC, McCarthy G: Temporal cortex activation in humans viewing eye and mouth movements. J Neurosci. 1998, 18: 2188-2199.PubMedGoogle Scholar
- Balconi M, Pozzili U: Face-selective processing and the effect of pleasant and unpleasant emotional expressions on ERP correlates. Int J Psychophysiol. 2003, 49: 67-74. 10.1016/S0167-8760(03)00081-3.View ArticlePubMedGoogle Scholar
- Sato W, Kochiyama T, Yoshikawa S, Matsumura M: Emotional expression boosts early visual processing of the face: ERP recording and its decomposition by independent component analysis. Neuroreport. 2001, 12: 709-714. 10.1097/00001756-200103260-00019.View ArticlePubMedGoogle Scholar
- O'Connor K, Hamm JP, Kirk IJ: The neurophysiological correlates of face processing in adults and children with Asperger's syndrome. Brain Cogn. 2005, 59: 82-95. 10.1016/j.bandc.2005.05.004.View ArticlePubMedGoogle Scholar
- Schupp HT, Ohman A, Junghofer M, Weike AI, Stockburger J, Hamm AO: The facilitated processing of threatening faces: an ERP analysis. Emotion. 2004, 4: 189-200. 10.1037/1528-35126.96.36.199.View ArticlePubMedGoogle Scholar
- Itier RJ, Taylor MJ: Inversion and contrast polarity reversal affect both encoding and recognition processes of unfamiliar faces: A repetition study using ERPs. Neuroimage. 2002, 15: 353-372. 10.1006/nimg.2001.0982.View ArticlePubMedGoogle Scholar
- Ashley V, Vuilleumier P, Swick D: Time course and specificity of event-related potentials to emotional expressions. Neuroreport. 2004, 15: 211-216. 10.1097/00001756-200401190-00041.View ArticlePubMedGoogle Scholar
- Eimer M, Holmes A: An ERP study on the time course of emotional face processing. Neuroreport. 2002, 13: 427-431. 10.1097/00001756-200203250-00013.View ArticlePubMedGoogle Scholar
- Eimer M, Holmes A, McGlone FP: The role of spatial attention in the processing of facial expression: an ERP study of rapid brain responses to six basic emotions. Cogn Affect Behav Neurosci. 2003, 3: 97-110. 10.3758/CABN.3.2.97.View ArticlePubMedGoogle Scholar
- Krolak-Salmon P, Fischer C, Vighetto A, Mauguiere F: Processing of facial emotional expression: spatio-temporal data as assessed by scalp event-related potentials. Eur J Neurosci. 2001, 13: 987-994. 10.1046/j.0953-816x.2001.01454.x.View ArticlePubMedGoogle Scholar
- Batty M, Taylor MJ: Early processing of the six basic facial emotional expressions. Brain Res Cogn Brain Res. 2003, 17: 613-620. 10.1016/S0926-6410(03)00174-5.View ArticlePubMedGoogle Scholar
- Caharel S, Courtay N, Bernard C, Lalonde R, Rebai M: Familiarity and emotional expression influence an early stage of face processing: An electrophysiological study. Brain Cognition. 2005, 59: 96-100. 10.1016/j.bandc.2005.05.005.View ArticlePubMedGoogle Scholar
- Rossion B, Joyce CA, Cottrell GW, Tarr MJ: Early lateralization and orientation tuning for face, word, and object processing in the visual cortex. Neuroimage. 2003, 20: 1609-1624. 10.1016/j.neuroimage.2003.07.010.View ArticlePubMedGoogle Scholar
- Dawson G, Webb SJ, Carver L, Panagiotides H, McPartland J: Young children with autism show atypical brain responses to fearful versus neutral facial expressions of emotion. Dev Sci. 2004, 7: 340-359. 10.1111/j.1467-7687.2004.00352.x.View ArticlePubMedGoogle Scholar
- Nelson CA, Nugent KM: Recognition Memory and Resource-Allocation as Revealed by Childrens Event-Related Potential Responses to Happy and Angry Faces. Dev Psychol. 1990, 26: 171-179. 10.1037/0012-16188.8.131.52.View ArticleGoogle Scholar
- Kestenbaum R, Nelson CA: Neural and behavioral correlates of emotion recognition in children and adults. J Exp Child Psychol. 1992, 54: 1-18. 10.1016/0022-0965(92)90014-W.View ArticlePubMedGoogle Scholar
- Batty M, Taylor MJ: The development of emotional face processing during childhood. Developmental Sci. 2006, 9: 207-220. 10.1111/j.1467-7687.2006.00480.x.View ArticleGoogle Scholar
- Taylor MJ: Non-spatial attentional effects on P1. Clin Neurophysiol. 2002, 113: 1903-1908. 10.1016/S1388-2457(02)00309-7.View ArticlePubMedGoogle Scholar
- Taylor MJ, Batty M, Itier RJ: The faces of development: A review of early face processing over childhood. J Cogn Neurosci. 2004, 16: 1426-1442. 10.1162/0898929042304732.View ArticlePubMedGoogle Scholar
- de Haan M, Johnson MH, Halit H: Development of face-sensitive event-related potentials during infancy: a review. Int J Psychophysiol. 2003, 51: 45-58. 10.1016/S0167-8760(03)00152-1.View ArticlePubMedGoogle Scholar
- Taylor MJ, Baldeweg T: Application of EEG, ERP and intracranial recordings to the investigation of cognitive functions in children. Developmental Sci. 2002, 5: 318-334. 10.1111/1467-7687.00372.View ArticleGoogle Scholar
- Ho TP, Leung PWL, Lee CC, Tang CP, Hung SF, Kwong SL, Lucas CP, Lieh-Mak F, Shaffer D: Test-retest reliability of the Chinese version of the Diagnostic Interview Schedule for Children-Version 4 (DISC-IV). J Child Psychol Psyc. 2005, 46: 1135-1138. 10.1111/j.1469-7610.2005.01435.x.View ArticleGoogle Scholar
- Ekman PMD: Japanese and Caucasian Facial Expressions of Emotion (JACFEE) and Neutral Faces (JACNeuF). 1988Google Scholar
- Berg P, Scherg M: A Multiple Source Approach to the Correction of Eye Artifacts. Electroencephalogr Clin Neurophysiol. 1994, 90: 229-241. 10.1016/0013-4694(94)90094-9.View ArticlePubMedGoogle Scholar
- Segalowitz SJ: ERPScore program: Peak and area analysis of event-related potentials. St. Catharines, Ontario: Brock University (Available from author). 1999Google Scholar
- McPartland J, Dawson G, Webb SJ, Panagiotides H, Carver LJ: Event-related brain potentials reveal anomalies in temporal processing of faces in autism spectrum disorder. J Child Psychol Psychiatry. 2004, 45: 1235-1245. 10.1111/j.1469-7610.2004.00318.x.View ArticlePubMedGoogle Scholar
- Picton TW, Bentin S, Berg P, Donchin E, Hillyard SA, Johnson R, Miller GA, Ritter W, Ruchkin DS, Rugg MD, Taylor MJ: Guidelines for using human event-related potentials to study cognition: Recording standards and publication criteria. Psychophysiology. 2000, 37: 127-152. 10.1017/S0048577200000305.View ArticlePubMedGoogle Scholar
- Scherg M, Berg P: New concepts of brain source imaging and localization. Electroencephalogr Clin Neurophysiol Suppl. 1996, 46: 127-137.PubMedGoogle Scholar
- Lamm C, Zelazo PD, Lewis MD: Neural correlates of cognitive control in childhood and adolescence: Disentangling the contributions of age and executive function. Neuropsychologia. 2006, 44: 2139-2148. 10.1016/j.neuropsychologia.2005.10.013.View ArticlePubMedGoogle Scholar
- Lewis MD, Lamm C, Segalowitz SJ, Stieben J, Zelazo PD: Neurophysiological correlates of emotion regulation in children and adolescents. J Cogn Neurosci. 2006, 18: 430-443. 10.1162/jocn.2006.18.3.430.View ArticlePubMedGoogle Scholar
- Jonkman LM, Sniedt FLF, Kemner C: Source localization of the Nogo-N2: A developmental study. Clin Neurophysiol. 2007, 118: 1069-1077. 10.1016/j.clinph.2007.01.017.View ArticlePubMedGoogle Scholar
- Hoechstetter K, Rupp A, Meinck HM, Weckesser D, Bornfleth H, Stippich C, Berg P, Scherg M: Magnetic source imaging of tactile input shows task-independent attention effects in SII. Neuroreport. 2000, 11: 2461-2465. 10.1097/00001756-200008030-00024.View ArticlePubMedGoogle Scholar
- Hoechstetter K, Rupp A, Stancak A, Meinck HM, Stippich C, Berg P, Scherg M: Interaction of tactile input in the human primary and secondary somatosensory cortex – A magnetoencephalographic study. Neuroimage. 2001, 14: 759-767. 10.1006/nimg.2001.0855.View ArticlePubMedGoogle Scholar
- Tarkka IM, Mnatsakanian EV: Functional Specialization in the Human Frontal Cortex Observed during Task Anticipation. 2003, 48: 102-110.Google Scholar
- Darvas F, Pantazis D, Kucukaltun-Yildirim E, Leahy RM: Mapping human brain function with MEG and EEG: methods and validation. Neuroimage. 2004, 23: S289-S299. 10.1016/j.neuroimage.2004.07.014.View ArticlePubMedGoogle Scholar
- Shibata T, Nishijo H, Tamura R, Miyamoto K, Eifuku S, Endo S, Ono T: Generators of visual evoked potentials for faces and eyes in the human brain as determined by dipole localization. Brain Topogr. 2002, 15: 51-63. 10.1023/A:1019944607316.View ArticlePubMedGoogle Scholar
- Caldara R, Thut G, Servoir P, Michel CM, Bovet P, Renault B: Face versus non-face object perception and the 'other-race' effect: a spatio-temporal event-related potential study. Clin Neurophysiol. 2003, 114: 515-528. 10.1016/S1388-2457(02)00407-8.View ArticlePubMedGoogle Scholar
- Werheid K, Alpay G, Jentzsch I, Sommer W: Priming emotional facial expressions as evidenced by event-related brain potentials. Int J Psychophysiol. 2005, 55: 209-219. 10.1016/j.ijpsycho.2004.07.006.View ArticlePubMedGoogle Scholar
- Sergent J, Ohta S, Macdonald B: Functional Neuroanatomy of Face and Object Processing – a Positron Emission Tomography Study. Brain. 1992, 115: 15-36. 10.1093/brain/115.1.15.View ArticlePubMedGoogle Scholar
- Lobaugh NJ, Gibson E, Taylor MJ: Children recruit distinct neural systems for implicit emotional face processing. Neuroreport. 2006, 17: 215-219. 10.1097/01.wnr.0000198946.00445.2f.View ArticlePubMedGoogle Scholar
- Gross AL, Ballif B: Childrens Understanding of Emotion from Facial Expressions and Situations – a Review. Dev Rev. 1991, 11: 368-398. 10.1016/0273-2297(91)90019-K.View ArticleGoogle Scholar
- Elfenbein HA, Ambady N: On the universality and cultural specificity of emotion recognition: A meta-analysis. Psychol Bull. 2002, 128: 203-235. 10.1037/0033-2909.128.2.203.View ArticlePubMedGoogle Scholar
- Carey S, Diamond R, Woods B: Developmental of face recognition – A maturational component?. Dev Psychol. 1980, 16: 257-269. 10.1037/0012-16184.108.40.2067.View ArticleGoogle Scholar
- Ellis HD: Developmental trends in face recognition. Psychologist. 1990, 3: 114-119.Google Scholar
- McGivern RF, Andersen J, Byrd D, Mutter KL, Reilly J: Cognitive efficiency on a match to sample task decreases at the onset of puberty in children. Brain Cogn. 2002, 50: 73-89. 10.1016/S0278-2626(02)00012-X.View ArticlePubMedGoogle Scholar
- Monk CS, McClure EB, Nelson EE, Zarahn E, Bilder RM, Leibenluft E, Charney DS, Ernst M, Pine DS: Adolescent immaturity in attention-related brain engagement to emotional facial expressions. Neuroimage. 2003, 20: 420-428. 10.1016/S1053-8119(03)00355-0.View ArticlePubMedGoogle Scholar
- Passarotti AM, Paul BM, Russiere JR, Buxton RB, Wong EC, Stiles J: The development of face and location processing: an fMRI study. Dev Sci. 2003, 6: 100-117. 10.1111/1467-7687.00259.View ArticleGoogle Scholar
- Everhart DE, Shucard JL, Quatrin T, Shucard DW: Sex-related differences in event-related potentials, face recognition, and facial affect processing in prepubertal children. Neuropsychology. 2001, 15: 329-341. 10.1037/0894-4220.127.116.119.View ArticlePubMedGoogle Scholar
- Liu L, Ioannides AA, Streit M: Single trial analysis of neurophysiological correlates of the recognition of complex objects and facial expressions of emotion. Brain Topogr. 1999, 11: 291-303. 10.1023/A:1022258620435.View ArticlePubMedGoogle Scholar
- Watanabe S, Kakigi R, Koyama S, Kirino E: Human face perception traced by magneto- and electro-encephalography. Brain Res Cogn Brain Res. 1999, 8: 125-142. 10.1016/S0926-6410(99)00013-0.View ArticlePubMedGoogle Scholar
- Handy TC: Event-related potentials: a methods handbook. 2005, Cambridge, Mass: MIT PressGoogle Scholar
- Halgren E, Baudena P, Heit G, Clarke M, Marinkovic K, Chauvel P: Spatiotemporal Stages in Face and Word-Processing .2. Depth-Recorded Potentials in the Human Frontal and Rolandic Cortices. J Physiol Paris. 1994, 88: 51-80. 10.1016/0928-4257(94)90093-0.View ArticlePubMedGoogle Scholar
- Ojemann JG, Ojemann GA, Lettich E: Neuronal-Activity Related to Faces and Matching in Human Right Nondominant Temporal Cortex. Brain. 1992, 115: 1-13. 10.1093/brain/115.1.1.View ArticlePubMedGoogle Scholar
- Lewis S, Thoma RJ, Lanoue MD, Miller GA, Heller W, Edgar C, Huang M, Weisend MP, Irwin J, Paulson K, Canive JM: Visual processing of facial affect. Neuroreport. 2003, 14: 1841-1845. 10.1097/00001756-200310060-00017.View ArticlePubMedGoogle Scholar
- Gorno-Tempini ML, Pradelli S, Serafini M, Pagnoni G, Baraldi P, Porro C, Nicoletti R, Umita C, Nichelli P: Explicit and incidental facial expression processing: An fMRI study. Neuroimage. 2001, 14: 465-473. 10.1006/nimg.2001.0811.View ArticlePubMedGoogle Scholar
- Narumoto J, Yamada H, Iidaka T, Sadato N, Fukui K, Itoh H, Yonekura Y: Brain regions involved in verbal or non-verbal aspects of facial emotion recognition. Neuroreport. 2000, 11: 2571-2576. 10.1097/00001756-200008030-00044.View ArticlePubMedGoogle Scholar
- Nakamura K, Kawashima R, Ito K, Sugiura M, Kato T, Nakamura A, Hatano K, Nagumo S, Kubota K, Fukuda H, Kojima S: Activation of the right inferior frontal cortex during assessment of facial emotion. J Neurophysiol. 1999, 82: 1610-1614.PubMedGoogle Scholar
- Sowell ER, Thompson PM, Holmes CJ, Jernigan TL, Toga AW: In vivo evidence for post-adolescent brain maturation in frontal and striatal regions. Nat Neurosci. 1999, 2: 859-861. 10.1038/13154.View ArticlePubMedGoogle Scholar
- Blakemore SJ, Choudhury S: Development of the adolescent brain: implications for executive function and social cognition. J Child Psychol Psychiatry. 2006, 47: 296-312. 10.1111/j.1469-7610.2006.01611.x.View ArticlePubMedGoogle Scholar
- Sabbagh MA, Moulson MC, Harkness KL: Neural correlates of mental state decoding in human adults: An event-related potential study. J Cogn Neurosci. 2004, 16: 415-426. 10.1162/089892904322926755.View ArticlePubMedGoogle Scholar
- Wiser AK, Andreasen NC, O'Leary DS, Crespo-Facorro B, Boles-Ponto LL, Watkins GL, Hichwa RD: Novel vs. well-learned memory for faces: A positron emission tomography study. J Cogn Neurosci. 2000, 12: 255-266. 10.1162/089892900562084.View ArticlePubMedGoogle Scholar
- Elgar K, Campbell R: Annotation: The cognitive neuroscience of face recognition: Implications for developmental disorders. J Child Psychol Psychiatry. 2001, 42: 705-717. 10.1111/1469-7610.00767.View ArticlePubMedGoogle Scholar
- Gur RC, Schroeder L, Turner T, McGrath C, Chan RM, Turetsky BI, Alsop D, Maldjian J, Gur RE: Brain activation during facial emotion processing. Neuroimage. 2002, 16: 651-662. 10.1006/nimg.2002.1097.View ArticlePubMedGoogle Scholar
- Aggleton JP: The Amygdala: neurobiological aspects of emotion, memory, and mental dysfunction. 1992, New York: Wiley-LissGoogle Scholar
- Iidaka T, Omori M, Murata T, Kosaka H, Yonekura Y, Okada T, Sadato N: Neural interaction of the amygdala with the prefrontal and temporal cortices in the processing of facial expressions as revealed by fMRI. J Cogn Neurosci. 2001, 13: 1035-1047. 10.1162/089892901753294338.View ArticlePubMedGoogle Scholar
- Wild B, Erb M, Bartels M: Are emotions contagious? Evoked emotions while viewing emotionally expressive faces: quality, quantity, time course and gender differences. Psychiatry Res. 2001, 102: 109-124. 10.1016/S0165-1781(01)00225-6.View ArticlePubMedGoogle Scholar
- Horwitz B, Poeppel D: How can EEG/MEG and fMRI/PET data be combined?. Hum Brain Mapp. 2002, 17: 1-3. 10.1002/hbm.10057.View ArticlePubMedGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.