Skip to main content

Event-related potentials to task-irrelevant changes in facial expressions



Numerous previous experiments have used oddball paradigm to study change detection. This paradigm is applied here to study change detection of facial expressions in a context which demands abstraction of the emotional expression-related facial features among other changing facial features.


Event-related potentials (ERPs) were recorded in adult humans engaged in a demanding auditory task. In an oddball paradigm, repeated pictures of faces with a neutral expression ('standard', p = .9) were rarely replaced by pictures with a fearful ('fearful deviant', p = .05) or happy ('happy deviant', p = .05) expression. Importantly, facial identities changed from picture to picture. Thus, change detection required abstraction of facial expression from changes in several low-level visual features.


ERPs to both types of deviants differed from those to standards. At occipital electrode sites, ERPs to deviants were more negative than ERPs to standards at 150–180 ms and 280–320 ms post-stimulus. A positive shift to deviants at fronto-central electrode sites in the analysis window of 130–170 ms post-stimulus was also found. Waveform analysis computed as point-wise comparisons between the amplitudes elicited by standards and deviants revealed that the occipital negativity emerged earlier to happy deviants than to fearful deviants (after 140 ms versus 160 ms post-stimulus, respectively). In turn, the anterior positivity was earlier to fearful deviants than to happy deviants (110 ms versus 120 ms post-stimulus, respectively).


ERP amplitude differences between emotional and neutral expressions indicated pre-attentive change detection of facial expressions among neutral faces. The posterior negative difference at 150–180 ms latency resembled visual mismatch negativity (vMMN) – an index of pre-attentive change detection previously studied only to changes in low-level features in vision. The positive anterior difference in ERPs at 130–170 ms post-stimulus probably indexed pre-attentive attention orienting towards emotionally significant changes. The results show that the human brain can abstract emotion related features of faces while engaged to a demanding task in another sensory modality.


The ability to detect changes in one's environment is important to survival. Not surprisingly, perhaps, change detection has been shown to work pre-attentively (for a review, see e.g. [1]). In a multitude of psychophysiological studies, change detection has been explored by recording event-related potentials (ERPs) to serially presented stimuli in a so called oddball paradigm. In this condition, frequently presented (standard) stimuli are randomly replaced by infrequent (deviant) ones that differ from them in one or more aspects. In audition, these changes elicit the mismatch negativity (MMN) component at 100–200 ms from stimulus onset, even if the subjects are not attending to the stimulation but concentrating on another task (for a review, see e.g. [2]). MMN is usually accompanied by the P3 (or P300) component, which has been interpreted to indicate attention switching to a change in the stimulus. P3 is modality non-specific and has been observed in response to infrequent (unattended) deviant stimuli (P3a, e.g. [3]) and to target stimuli (P3b, e.g. [4]).

MMN was originally established in the auditory modality [5], but there is growing evidence of its visual counterpart, visual MMN (vMMN, for a review, see [6]). For example, changes in color (e.g. [7]), line orientation (e.g. [8]), motion direction (e.g. [9]), and spatial frequency (e.g. [10]) elicit vMMN.

In audition, MMN is found not only to changes in single physical features but also to changes in feature combinations and even in abstract stimulus features (for a review, see [11]). The same seems to apply to vision. vMMN has been reported for combinations of two features, color and grating pattern orientation [12]. Interestingly, a possible vMMN to changes in a complex and socially relevant visual stimulus, facial expression, has been reported. In the experiment by Zhao and Li [13], neutral (standards), happy and sad faces (deviants) were presented during intervals between two attended tones. The paradigm involved a tone discrimination task and the participants were instructed to ignore the faces. The results showed that ERPs to both types of deviant expressive faces were more negative than those to standard neutral faces. Difference ERPs started around 110–120 ms post-stimulus, spanned approximately 300 ms and they were elicited in a large posterior area. Zhao and Li [13] named their finding as expression mismatch negativity (EMMN). However, the facial stimuli in their experiment contained pictures of one person only. It is, thus, possible that the differential ERPs they found to expression changes reflected only differences in low-level features in these pictures. The same holds true also with another study applying neutral expressions as deviants and happy faces as standards while the subjects were instructed to detect target faces with glasses [14]. The authors reported a vMMN with a maximum amplitude around 280 ms post-stimulus in lateral posterior electrode sites.

In order to test whether the brain detects emotional expression-related changes when a multitude of low-level features forming facial identity simultaneously varies, we used pictures of faces from four different models. In the oddball paradigm, pictures of happy or fearful faces as rare deviants were presented among neutral standard faces. The identity of the faces changed from trial to trial. In the auditory MMN studies, the primary task has usually been in the visual modality. Analogously, in the present study, the subjects were attending to an auditory task. Because the task was presented asynchronously with the visual deviants, it required participants' ongoing attention. The present paradigm presumably left very little room for attentional processing of the visual deviants unlike stimulus conditions in which the attended stimuli are never overlapping with the stimuli instructed to be ignored (as was the case, for example, in the previous studies of vMMN to facial stimuli [13, 14]). A demanding primary task is important in MMN experiments, because the attention switches towards oddball stimuli may elicit attention-related ERP components which may cover the MMN (for the N2/P3a responses to changes in attended emotional faces, see [15]).

We hypothesized that both deviants would elicit differential ERPs compared to standards. Furthermore, we expected that fearful faces would elicit vMMN of a larger amplitude as compared to vMMN elicited by happy faces. This was expected because previously a possible vMMN to negative (i.e. sad) faces elicited enhanced ERPs compared to happy faces [13]. Also, apart from oddball paradigm, fearful faces have elicited enhanced ERP amplitudes compared to neutral or positive faces in other stimulus conditions also (e.g. [16, 17]).

The comparison of the time course between the processing of fear and happiness is also interesting because there are two, somewhat contradictory, lines of evidence related to the speed of processing of these emotional expressions. Namely, on the one hand, some behavioral and electrophysiological studies have shown a so called happy face advantage (e.g. [16, 1820], that is, faster responses to happy expression relative to negative expressions. On the other hand, several behavioral studies have reported faster responses for threat-related versus happy facial stimuli (e.g. [2123]). Given this, we did not make exact predictions regarding whether differential ERPs would emerge more rapidly for happy or for fearful faces.



Fourteen (four male and ten female) native Finnish-speaking volunteers (age-range 20–29 years, mean age 23.8 years) participated in the study. The data obtained from two subjects were abandoned (the recording of one subject was interrupted because of headache and the data of another subject contained a large number of movement artefacts). The data were thus analyzed for twelve subjects. The participants were all right-handed and had normal or corrected-to-normal vision and reported no neurological impairments. An informed consent was obtained from the subjects before their participation. The experiment was undertaken in accordance with the Declaration of Helsinki.

Stimuli and procedure

During the recordings, participants were seated in a chair and attended to a radio play presented via ear-phones. The play consisted of an oral narrative in Finnish including speech and non-speech sounds as well as musical elements. The subjects were instructed to count the number of words in the story beginning with the sound /y/ and to fix their gaze on a computer screen (20-inch, 1024 × 768-pixel, 75 Hz display) 100 cm away from them. During the recordings, the subjects were observed via a video monitor.

The visual stimuli were pictures of faces of four different models (male actors PE and JJ, female actors MF and NR) from Pictures of Facial Affect [24]. Pictures of a neutral, fearful and happy expression from each model were used. The pictures were digitized (Kodak Photo CD™) and Adobe Photoshop 4.0 was used to convert the pictures to 256 (bits) grayscale images. The stimulus presentation was controlled with STIM program (NeuroScan, Inc.).

The facial pictures, occupying a visual angle of 4 × 5°, were presented at fixation for 200 ms. The stimulus-onset asynchrony (SOA) was 500 ms. Fig. 1 illustrates the stimulus sequence. In a modified oddball paradigm, two different deviant stimulus types were infrequently interspersed between frequently presented standard stimuli. Standards and deviants differed from each other in emotional expression. In the standard pictures, the models wore a neutral facial expression, while in the deviant pictures either a fearful (fearful deviant) or happy (happy deviant) expression was present. Standards and deviants were presented pseudo-randomly with the restriction that there were no less than two standards between consecutive deviants. Of 1200 stimuli, the probability for the standards was 90% and the probability for the deviants was 10% (5% for happy deviants and 5% for fearful deviants). Importantly, the facial identity of the faces changed from trial to trial.

Figure 1
figure 1

Illustration of the stimulus paradigm applied.

ERP recordings

An electroencephalogram (EEG) was continuously recorded with Ag/AgCl electrodes at Fz, Cz, Pz, O1, Oz, and O2 according to the international 10–20 system. Horizontal eye movements were measured by two bipolar electrodes placed lateral to the left and right orbits. Eye blinks and vertical eye movements were measured from bipolar electrodes placed below and above the right eye. A reference electrode was placed at the tip of the nose. The impedance of all the electrodes was maintained at < 5 kΩ. The signals from the electrodes were amplified, digitally band-bass filtered from 0.05 to 100 Hz, and stored on a computer disk at a sample rate of 1000 Hz (Syn-Amps 4.3, NeuroScan, Inc.).

Data analysis

Single EEG sweeps (epochs from 100 ms before to 500 ms after stimulus onset) were corrected by their baseline (mean amplitude of the 100-ms pre-stimulus period) and digitally filtered (1 to 30 Hz, 12 dB per octave roll off). Prior to EEG averaging, computerized artefact rejection was performed to discard epochs whose amplitude exceeded ± 50 μV in any recording electrode. Discarded epochs were rejected from the analysis. The recording sessions for four subjects, which would have contained more than 1/3 discarded epochs, were not treated in this way. Instead, the continuous EEG-signal was corrected for blink artefact using an eye movement reduction algorithm [25].

The sweeps were averaged separately for both types of deviants and standards. Moreover, in order to obtain an equal number of standards and deviants, only the standards immediately preceding both types of deviants were used in the analysis.

Time windows for the statistical analyses were extracted according to the literature and to visual inspection of the waveforms. The vMMN is usually found occipitally at 100–200 ms from stimulus onset [6]. Consistently, in our data the vMMN-like deflection for the fearful deviants would appear to be starting after 150 ms post-stimulus (Fig. 2), and the vMMN for the happy deviants would appear to be ending at around 180 ms post-stimulus (Fig. 3). Therefore, for the amplitude analysis, the mean amplitude values for the standard and deviant ERPs at O1, Oz, and O2 were extracted from the time window of 150–180 ms post-stimulus.

Figure 2
figure 2

Grand averaged ERPs to the fearful deviants (blue lines) and the neutral standards (red lines) immediately preceding them. The difference ERPs (deviant minus standard) are drawn with the black lines. The time windows for extracting the mean values for the repeated measures MANOVA are marked with rectangles. The x-axis shows stimulus onset.

Figure 3
figure 3

Grand averaged ERPs to the happy deviants (blue lines) and the neutral standards (red lines) immediately preceding them. The difference ERPs (deviant minus standard) are drawn with the black lines. The time windows for extracting the mean values for the repeated measures MANOVA are marked with rectangles. The x-axis shows stimulus onset.

Visual inspection of grand average ERPs also revealed a later posterior negativity peaking at a latency of approximately 300 ms (Fig. 2 and 3). The mean amplitude values of this deflection were extracted in the time window of 280–320 ms post-stimulus. No P3a deflection, which is usually elicited fronto-centro-parietally at approximately 300–500 ms after stimulus onset by the deviants (if they catch the subjects' involuntary attention, e.g. [26]), was observed in the present study (Fig. 2 and 3).

In addition to the occipital negativities, an anterior positivity at around 130–170 ms post-stimulus was observed (Fig. 2 and 3). Accordingly, the mean amplitude values for the standard and deviant ERPs were extracted also at Cz and Fz.

The resultant mean amplitudes concerning the three deflections (separately for the deflections at latencies of 130–170 ms, 150–180 ms and 280–320 ms) were analyzed by repeated measures multivariate analyses of variance (MANOVA) with Stimulus type (standard, deviant), Expression (fearful, happy) and Electrode site (Fz, Cz for the deflection at 130–170 ms post-stimulus, and O1, Oz, O2 for the deflections at 150–180 ms and 280–320 ms post-stimulus) as factors. The F statistics were exact.

Because the analysis windows with fixed latencies do not take into account the time-course of ERP differences, we also studied possible differences between the detection of fearful and happy faces against neutral standards by applying a global waveform analysis. This approach preserves the temporal resolution of the EEG (here, sampling frequency of 1000 Hz) (for a similar data analysis, see e.g. [27, 28]). Note that even if this approach does not qualify as a reliable procedure for identifying ERP effects, it does qualify as a method for reliably revealing the temporal dynamics of the EEG by estimating the onsets and offsets of ERP effects. We ran point-by-point paired t-tests between responses to deviants and standards separately for each electrode and both deviant types. To counteract the likelihood of the increased number of statistically significant results associated with multiple t-tests, an alpha level smaller than .05 in at least 10 consecutive data points (10 ms) was required for the difference to be considered significant (see e.g. [29]). All the applied t-tests were two-tailed.


In the primary task, the mean number of identified target words was 30 (S.E. = 2.29) out of 49. In the ERPs, two well-known responses to visual stimuli were identified: the occipital P1 (e.g. [30]) and parieto-occipital P2 (e.g. [31]) (Fig. 2 and 3). However, no P3a (e.g. [26]), which is associated with involuntary attention switches towards changes, was found. Note also that face-sensitive N170 (e.g. [32]) was unobservable, as it is usually recorded maximally from lateral parietal/temporal electrodes (i.e. P7, P8, T5, T6), and we did not have these sites in our electrode montage.

ERP amplitudes

For posterior negativity at 150–180 ms post-stimulus (Fig. 2 and 3), a 3-way MANOVA, Stimulus type (standard, deviant) * Expression (happy, fearful) * Electrode site (O1, Oz, O2), showed a significant main effect of Stimulus type, F(1,11) = 14.7, p = .003, and Expression, F(1,11) = 6.6, p = .026. Electrode site or any of the interaction effects were not significant. Notably, the Stimulus type * Expression interaction was not significant, indicating that the ERP amplitudes to the fearful and happy deviants relative to the amplitudes of their respective standards were equal. The mean amplitude values for the ERPs to fearful deviants and standards were 0.55 μV and 1.38 μV, respectively (averaged across O1, Oz, and O2). The mean difference in amplitudes (deviant – standard) was, thus, -0.83 μV. Correspondingly, the mean ERP amplitudes to happy deviants was 0.03 μV and to standards 0.94 μV, the difference being -0.91 μV.

For posterior negativity at 280–320 ms post-stimulus (Fig. 2 and 3), a 3-way MANOVA, Stimulus type (standard, deviant) * Expression (happy, fearful) * Electrode site (O1, Oz, O2,) revealed a significant main effect of Stimulus type, F(1,11) = 18.2, p = .001. The main effects of neither Expression or Electrode site nor any of the interaction effects involving the main factors were not significant. The mean amplitudes of the ERPs for the deviants and standards were 0.08 μV and 1.21 μV, respectively (averaged over expressions and electrodes). Thus, the mean difference was -1.13 μV.

For anterior positivity at 130–170 ms after stimulus onset (Fig. 2 and 3), a MANOVA with factors Stimulus type (standard, deviant) * Expression (happy, fearful) * Electrode site (Fz, Cz) revealed a main effect of Stimulus type, F(1,11) = 16.4, p = .002, and Electrode site, F(1,11) = 19.2, p = .001. In addition, an interaction effect of Expression * Electrode site was found, F(1,11) = 6.4, p = .028. The main effect of Stimulus type indicated that the anterior ERP response was significantly more positive to deviants (1.52 μV) than to standards (0.26 μV). Because of the Expression * Electrode site interaction, the effect of expression was analyzed separately for both electrode sites. At Fz, there was no difference in the mean amplitudes (averaged over deviants and standards) of the ERPs to happy and fearful stimuli, 0.53 μV and 0.52 μV, respectively. Instead, at Cz, the mean amplitude was larger for fearful than happy standard-deviant -pairs, 1.48 μV and 1.02 μV, respectively, t(1,11) = -2.2, p = .048.


In order to analyze possible differences in the time course of the vMMN for the fearful and happy deviants, the ERPs to them were compared to the ERPs to the respective standards preceding them by point-wise paired t-tests (Table 1). These analyses revealed that, at the occipital sites, ERPs to happy deviants differed from ERPs to standards earlier than ERPs to fearful deviants differed from ERPs to standards. Differential ERPs between happy deviants and standards emerged at approximately 140 ms from stimulus onset while differential ERPs between fearful deviants and standards were found after 160 ms from stimulus onset. At Fz and Cz, this difference emerged, in turn, earlier to fearful than to happy deviants (approximately at a latency of 110 ms and 120 ms, respectively). At Pz, ERPs to fearful deviants did not differ from ERPs to standards (according to the criterion used; see Methods section), but differential ERPs between happy deviants and standards emerged at 280 ms from stimulus onset.

Table 1 Significant differences in ERP latencies.


In an oddball paradigm, task-irrelevant changes in facial expression, i.e. fearful and happy deviants, elicited differential ERPs relative to neutral standard expressions in three analysis windows. Negative shifts were observed occipitally at 150–180 and 280–320 ms post-stimulus and a positive shift fronto-centrally at 130–170 ms post-stimulus.

The occipital negativity at the latency of 150–180 ms may correspond to visual mismatch negativity, vMMN. This is supported by the experimental condition applied (oddball paradigm), scalp distribution (occipital), polarity (a negative difference between ERPs to deviants and those to standards), and the latency range of the deflection found. Regarding the latency of the vMMN, most of the studies have reported it occurring before 200 ms (e.g. color deviance: 120–200 ms, [7, 33]; orientation deviance: 160–205 ms, [8, 34]; spatial frequency: 120–200 ms, [10]; conjunction of color and orientation: < 150 ms, [12]). However, some data have suggested later latencies for the vMMN (i.e. 210–400 ms; [35, 36]). A possible vMMN to emotional expressions (EMMN) was found at a wide latency range including the modulation of N170 and P250 [13]. The relationship of N170 and vMMN to facial expressions is puzzling, however. N170 is a face-specific component which is traditionally associated to the structural encoding of the facial stimuli [32], but not to encoding of emotional expressions (e.g. [14, 3742]). However, there are also studies which have shown a modulation of N170 to the emotional faces [16, 17, 43, 44], but this modulation has been observed for threatening (fearful/angry) faces only. Unlike in the present study, in the study by Zhao & Li [13] electrode montage covered occipital-temporal sites, and a clear N170 which was enhanced to both happy and sad deviants as compared to neutral standards was found. Interestingly, the authors interpreted that the modulation of N170 was not induced by face processing but reflected a vMMN induced by the rarity of the deviant emotional faces. In contrast to the results of Zhao and Li, Susac and colleagues [14] found no difference in N170 amplitudes between happy standards and neutral deviants. The MMN-like activity they recorded to change in emotional expression was found after 200 ms being maximal at 280 ms post-stimulus.

It is obvious that the relationship between the vMMN and N170 must be clarified in the future studies. This will require high-density ERP-recording and/or powerful signal analysis tools such as independent component analysis [45]. In any case, even if the present study can not provide unequivocal support for the presence of vMMN separate of face-sensitive N170, the ERPs reported in the present study provide evidence for pre-attentive change detection of positive and negative facial emotions. Against to our assumption, fearful deviants did not elicit larger negativity than happy deviants compared to standards. Our hypothesis was based on findings from the other stimulus conditions than oddball paradigm. In these studies, N170 has been larger in amplitude to fearful than to neutral or happy faces (e.g. [17]). Our data suggest that vMMN-like negativity obtained in the oddball condition does not follow this pattern.

It is suggested that MMN elicitation requires several neural computations, in this case, the registration and formation of a high-level representation (abstraction) of facial expression from several low-level features (since four different facial identities were applied), the short-term storage of these facial expression configurations, and the detection of mismatch between the expression of incoming deviants and the stored (neutral) standard expression (for the memory trace explanation of MMN, see, [46]). Therefore, for example, manipulating the time required to maintain the representation of the standard stimulus in the sensory memory (interval between the consecutive stimuli) may tell if the ERPs to facial expressions are subordinate to the sensory memory, as the MMN has found to be (for the auditory MMN, see e.g. [47]; for the visual MMN, see [8]).

In addition to posterior negative difference ERPs, we found an anterior positivity to both deviants relative to standards at 130–170 ms post-stimulus. This may correspond to the frontal positivity found concurrently with occipital vMMN in the previous studies [7, 8, 12]. In their vMMN study, Czigler and colleagues [7] interpreted the frontal difference ERPs to be related to feature-specific neural refractoriness caused by the different presentation rates of the standard (frequent) and deviant (infrequent) stimulus types. Namely, it can be assumed that the neural population responding to repeated standards may be more refractory than the population responding to rare deviants. The difference observed in ERP amplitudes may thus result from these differences in neural refractoriness (for the refractoriness explanation of MMN, see, [46]). Even if this explanation is possible in the case of the study by Czigler et al. [7], in the case of the present study, it is implausible. This is because we applied several personal identities in the facial pictures, thereby making the stimulation variable within the stimulus categories.

It is also possible that the fronto-central positivity is not specifically related to the oddball paradigm. In some previous studies investigating ERPs to facial expressions, but not applying the oddball paradigm, a frontal positivity in ERPs to emotional expressions relative to neutral ones at a latency corresponding to that observed in the present study has been found (e.g. [38, 48]). When the subjects' task was to respond to immediate stimulus repetitions, fearful faces elicited a positive shift in ERPs compared to neutral faces at fronto-central sites as early as 110 ms post-stimulus [38]. One possible functional explanation for this deflection provided by Eimer and Holmes [49] is that the fronto-central positivity is associated with early directing of attention towards emotionally significant events. In the present study, in which the subjects were concentrating on a demanding auditory task, attention was not paid to the facial pictures but to a demanding auditory task. A lack of P3a component supports the view that the participants ignored the visual stimuli. However, because attentional orienting towards a change in a facial expression may not require subjective awareness, this explanation is also possible in the case of the present experiment. Indeed, Kiss and Eimer [48] found that subliminally as well as supraliminally presented facial fear elicited differential fronto-central ERPs compared to neutral faces. This difference in ERPs emerged at 140–180 ms after stimulus onset which is comparable to the latency of our finding. Further research is needed to find out if the anterior positive difference ERPs are related to automatic directing of attention towards emotional faces or towards (infrequent) changes in visual stimuli in general (observed as a frontal positivity concurrently with the vMMN).

It is uncertain if the differential ERPs of negative polarity we found occipitally at the latency range of 280–320 ms can be associated with vMMN. Some authors have interpreted change-related differential ERPs at corresponding latencies as vMMN (210–400 ms: [35, 36]). Also vMMN to changes in emotional expression (rare neutral facial expressions among frequent happy expressions) and in facial identity was elicited around 300 ms post-stimulus [14]. In addition, vMMN to sad and happy faces elicit long-lasting negativity approximately at latency from 100 to 400 ms [13]. Another possibility is that this negativity reflects so called early posterior negativity (EPN). The EPN is a negative-going occipitotemporal potential occurring approximately 150–350 ms after stimulus onset which has been related to the perceptual encoding and early selection of visual stimuli with affective and motivational significance [50, 51]. The EPN has also been shown to be sensitive to facial expressions, but importantly, only threatening but not happy faces have been show to elicit enhanced EPN relative to neutral faces [52]. However, as in the present study the differential ERPs at the latency of 280–320 ms were not different for fearful and happy faces, we do not consider it likely that this difference would have reflected the significance of the emotional content of the deviant stimuli. Instead, we find it more likely that the differential ERPs at the 280–320 ms latency range reflected the detection of infrequent emotional expressions among a stream of neutral faces.

The analysis of the ERP latencies showed that, at the occipital channels, happy deviants were detected as different from neutral standards more than 20 ms earlier than fearful deviants were detected from neutral standards (140 ms versus 160 ms post-stimulus, respectively). In contrast, at the fronto-central recording sites, a similar change detection of fearful deviants was approximately 10 ms earlier than that of happy deviants (110 ms versus 120 ms post-stimulus, respectively). How can these seemingly contradictory results be explained? We suggest that the occipital electrodes, which were most probably recording the activity in the visual cortex, reflected activity related to structural analysis of the incoming visual input. Accordingly, we assume that the observed latency difference between processing happy and fearful deviants at the occipital channels reflects the greater visual saliency of the happy than fearful expression as compared to neutral expression. This speculation is in line with the evidence from the quantitative parameterisation of facial expressions of emotions. For example, Johnston and colleagues [53] defined facial expressions of basic emotions (happiness, surprise, fear, disgust, anger, and sadness) and emotional neutrality in a 12-dimensional space, and showed that the space representing happy faces had the least overlap with the space defining other expressions. In other words, based on the structural features of facial expressions, happy faces are the most discriminable. Previous behavioral studies (e.g. [1820, 54]) have also reported faster recognition times to happy expressions relative to negative expressions. The present data, as well as the previous electrophysiological ones, support these findings, i.e., vMMN for the happy faces were found earlier to than vMMN to sad faces [13] and positive expressions elicited earlier N170 than negative ones [16]. Interestingly, at the anterior recording sites, our results showed a reversed pattern of latency difference: fronto-centrally fearful deviants were detected earlier than happy deviants. As we earlier speculated, this anterior positivity may be related to automatic and involuntary shifts of attention towards emotional faces. Faster responses to threatening than to neutral or positive stimuli have been found in previous electrophysiological (e.g. [38]) and behavioural studies (e.g. [2123]). Our electrophysiological data from the oddball condition suggest that fearful faces elicit earlier anterior positivity as compared to positive faces, possibly indicating a frontal mechanism for threat detection.


Because there were only six electrodes in our electrode montage, the localization of the brain activation could be done only roughly. Larger number of electrodes would have allowed source localization of the activity as well as utilization of blind source separation methods such as independent component analysis. These may be needed to further examine the relationship between N170 and vMMN to facial expressions.


Differential ERPs to rare emotional faces among the frequent neutral ones presented in the oddball paradigm were found when the subjects were attending to an auditory task. These differential ERPs cannot be accounted by the differences in the low-level features in the stimuli, because the identity of the faces and, therefore, several low-level features in the faces, varied from trial to trial. Instead, the results suggest that a change detection mechanism abstracting emotion-related features among changing low-level features works pre-attentively.





Early posterior negativity


Event-related potentials


Multivariate analyses of variance


Mismatch negativity


Visual mismatch negativity.


  1. Compton RJ: The interface between emotion and attention: A review of evidence from psychology and neuroscience. Behav Cogn Neurosci Rev. 2003, 2: 115-129. 10.1177/1534582303002002003.

    Article  PubMed  Google Scholar 

  2. Näätänen R, Winkler I: The concept of auditory stimulus representation in cognitive neuroscience. Psychol Bullet. 1999, 6: 826-859. 10.1037/0033-2909.125.6.826.

    Article  Google Scholar 

  3. Courchesne E, Hillyard SA, Galambos R: Stimulus novelty, task relevance and the visual evoked potential in man. Electroencephalogr Clin Neurophysiol. 1975, 39: 131-143. 10.1016/0013-4694(75)90003-6.

    Article  CAS  PubMed  Google Scholar 

  4. Courchesne E: Changes in P3 waves with event repetition: long-term effects on scalp distribution and amplitude. Electroencephalogr Clin Neurophysiol. 1978, 45: 754-766. 10.1016/0013-4694(78)90143-8.

    Article  CAS  PubMed  Google Scholar 

  5. Näätänen R, Gaillard AWK, Mäntysalo S: Early selective attention effect on evoked potential reinterpreted. Acta Psychol. 1978, 42: 313-329. 10.1016/0001-6918(78)90006-9.

    Article  Google Scholar 

  6. Pazo-Alvarez P, Cadaveira F, Amenedo E: MMN in the visual modality: a review. Biol Psychol. 2003, 63: 199-236. 10.1016/S0301-0511(03)00049-8.

    Article  CAS  PubMed  Google Scholar 

  7. Czigler I, Balázs L, Winkler I: Memory-based detection of task-irrelevant visual changes. Psychophysiology. 2002, 39: 869-873. 10.1111/1469-8986.3960869.

    Article  PubMed  Google Scholar 

  8. Astikainen P, Lillstrang E, Ruusuvirta T: Visual mismatch negativity for changes in orientation – a sensory memory-dependent response. Eur J Neurosci. 2008, 28: 2319-2324. 10.1111/j.1460-9568.2008.06510.x.

    Article  PubMed  Google Scholar 

  9. Pazo-Alvarez P, Amenedo E, Cadaveira F: Automatic detection of motion direction changes in the human brain. Eur J Neurosci. 2004, 19: 1978-1986. 10.1111/j.1460-9568.2004.03273.x.

    Article  CAS  PubMed  Google Scholar 

  10. Heslenfeld DJ: Visual mismatch negativity. Detection of Change: Event-Related Potential and fMRI Findings. Edited by: Polich J. 2003, Boston: Kluver Academic Publishers, 41-59.

    Chapter  Google Scholar 

  11. Näätänen R, Tervaniemi M, Sussman E, Paavilainen P, Winkler I: "Primitive intelligence" in the auditory cortex. Trends Neurosci. 2001, 24: 283-288. 10.1016/S0166-2236(00)01790-2.

    Article  PubMed  Google Scholar 

  12. Winkler I, Czigles I, Sussman E, Horváth J, Balázs L: Preattentive binding of auditory and visual stimulus features. J Cogn Neurosci. 2005, 17: 320-339. 10.1162/0898929053124866.

    Article  PubMed  Google Scholar 

  13. Zhao L, Li J: Visual mismatch negativity elicited by facial expressions under non-attentional condition. Neurosci Lett. 2006, 410: 126-131. 10.1016/j.neulet.2006.09.081.

    Article  CAS  PubMed  Google Scholar 

  14. Susac A, Ilmoniemi RJ, Pihko E, Supek S: Neurodynamic studies on emotional and inverted faces in an oddball paradigm. Brain Topograph. 2004, 16: 265-268. 10.1023/B:BRAT.0000032863.39907.cb.

    Article  Google Scholar 

  15. Campanella S, Gaspard C, Debatisse D, Bruyer R, Crommelinck M, Guerit JM: Discrimination of emotional facial expressions in a visual oddball task: an ERP study. Biol Psychol. 2002, 59: 171-186. 10.1016/S0301-0511(02)00005-4.

    Article  CAS  PubMed  Google Scholar 

  16. Batty M, Taylor MJ: Early processing of the six basic facial emotional expressions. Cogn Brain Res. 2003, 17: 613-620. 10.1016/S0926-6410(03)00174-5.

    Article  Google Scholar 

  17. Leppänen JM, Kauppinen P, Peltola MJ, Hietanen JK: Differential electrocortical responses to increasing intensities of fearful and happy emotional expressions. Brain Res. 2007, 1166: 103-109. 10.1016/j.brainres.2007.06.060.

    Article  PubMed  CAS  Google Scholar 

  18. Kirita T, Endo M: Happy-face advantage in recognizing facial expressions. Acta Psychol (Amst). 1995, 89: 149-163. 10.1016/0001-6918(94)00021-8.

    Article  Google Scholar 

  19. Kolassa IT, Miltner WH: Psychophysiological correlates of face processing in social phobia. Brain Res. 2006, 1118: 130-141. 10.1016/j.brainres.2006.08.019.

    Article  CAS  PubMed  Google Scholar 

  20. Leppänen JM, Hietanen JK: Emotionally positive facial expressions are processed faster than negative facial expressions, but why?. Psychol Res. 2004, 69: 22-29. 10.1007/s00426-003-0157-2.

    Article  PubMed  Google Scholar 

  21. Fox E, Lester V, Russo R, Bowles RJ, Pichler A, Dutton K: Facial expressions of emotion: Are angry faces detected more efficiently?. Cogn Emot. 2000, 14: 61-92. 10.1080/026999300378996.

    Article  PubMed Central  PubMed  Google Scholar 

  22. Hansen CH, Hansen RH: Finding the face in the crowd: an anger superiority effect. J Pers Soc Psych. 1988, 54: 917-924. 10.1037/0022-3514.54.6.917.

    Article  CAS  Google Scholar 

  23. Öhman A, Lundqvist D, Esteves F: The face in the crowd revisited: A threat advantage with schematic stimuli. J Pers Soc Psychol. 2001, 80: 381-396. 10.1037/0022-3514.80.3.381.

    Article  PubMed  Google Scholar 

  24. Ekman P, Friesen WV: Pictures of facial affect. 1976, Palo Alto, CA: Consulting Psychologists Press

    Google Scholar 

  25. Semlitsch HV, Anderer P, Schuster P, Presslich O: A solution for reliable and valid reduction of ocular artifacts, applied to the P300 ERP. Psychophysiology. 1986, 23: 695-703. 10.1111/j.1469-8986.1986.tb00696.x.

    Article  CAS  PubMed  Google Scholar 

  26. Knight RT: Distributed cortical network for visual attention. J Cogn Neurosci. 1994, 9: 75-91. 10.1162/jocn.1997.9.1.75.

    Article  Google Scholar 

  27. Murray MM, Michel CM, Grave de Peralta R, Ortigue S, Brunet D, Gonzales Andino S, Schnider A: Rapid discrimination of visual and multisensory memories revealed by electrical neuroimaging. Neuroimage. 2004, 21: 125-135. 10.1016/j.neuroimage.2003.09.035.

    Article  PubMed  Google Scholar 

  28. Pourtois G, Schwartz S, Seghier ML, Lazeyras F, Vuilleumier P: Neural systems for orienting attention to the location of threat signals: An event-related fMRI study. NeuroImage. 2006, 31: 920-933. 10.1016/j.neuroimage.2005.12.034.

    Article  PubMed  Google Scholar 

  29. Guthrie D, Buchwald JS: Significance testing of difference potentials. Psychophysiology. 1991, 28: 240-244. 10.1111/j.1469-8986.1991.tb00417.x.

    Article  CAS  PubMed  Google Scholar 

  30. Luck SJ, Heinze HJ, Mangun GR, Hillyard SA: Visual event-related potentials index focused attention within bilateral stimulus arrays. II. Functional dissociation of P1 and N1 components. Electroencephalogr Clin Neurophysiol. 1990, 75: 528-542. 10.1016/0013-4694(90)90139-B.

    Article  CAS  PubMed  Google Scholar 

  31. Johannes S, Münte TF, Heinze HJ, Mangun GR: Luminance and spatial attention effects on early visual processing. Cogn Brain Res. 1995, 2: 189-205. 10.1016/0926-6410(95)90008-X.

    Article  CAS  Google Scholar 

  32. Bentin S, Allison T, Puce A, Perez E, McCarthy G: Electrophysiological studies of face perception in humans. J Cogn Neurosci. 1996, 8: 551-565.

    Article  PubMed Central  PubMed  Google Scholar 

  33. Czigler I, Balázs L, Pató LG: Visual change detection: event-related potentials are dependent on stimulus location in humans. Neurosci Lett. 2004, 364: 149-153. 10.1016/j.neulet.2004.04.048.

    Article  CAS  PubMed  Google Scholar 

  34. Astikainen P, Ruusuvirta T, Wikgren J, Korhonen T: The human brain processes visual changes that are not cued by attended auditory stimulation. Neurosci Lett. 2004, 368: 231-234. 10.1016/j.neulet.2004.07.025.

    Article  CAS  PubMed  Google Scholar 

  35. Tales A, Newton P, Troscianko T, Butler S: Mismatch negativity in the visual modality. NeuroReport. 1999, 10: 3363-3367. 10.1097/00001756-199911080-00020.

    Article  CAS  PubMed  Google Scholar 

  36. Stagg C, Hindley P, Tales A, Butler S: Visual mismatch negativity: The detection of stimulus change. NeuroReport. 2004, 15: 659-663. 10.1097/00001756-200403220-00017.

    Article  PubMed  Google Scholar 

  37. Ashley V, Vuilleumier P, Swick D: Time course and specificity of event-related potentials to emotional expressions. Neuroreport. 2004, 15: 211-216. 10.1097/00001756-200401190-00041.

    Article  PubMed  Google Scholar 

  38. Eimer M, Holmes A: An ERP study on the time course of emotional face processing. Neuroreport. 2002, 13: 427-431. 10.1097/00001756-200203250-00013.

    Article  PubMed  Google Scholar 

  39. Eimer M, Holmes A, McGlone FP: The role of spatial attention in the processing of facial expression: an ERP study of rapid brain responses to six basic emotions. Cogn Affect Behav Neurosci. 2003, 3: 97-110. 10.3758/CABN.3.2.97.

    Article  PubMed  Google Scholar 

  40. Holmes A, Vuilleumier P, Eimer M: The processing of emotional facial expression is gated by spatial attention: evidence from event-related brain potentials. Brain Res Cogn Brain Res. 2003, 16: 174-184. 10.1016/S0926-6410(02)00268-9.

    Article  PubMed  Google Scholar 

  41. Holmes A, Winston JS, Eimer M: The role of spatial frequency information for ERP components sensitive to faces and emotional facial expression. Brain Res Cogn Brain Res. 2005, 25: 508-520. 10.1016/j.cogbrainres.2005.08.003.

    Article  PubMed  Google Scholar 

  42. Santesso DL, Meuret AE, Hofmann SG, Mueller EM, Ratner KG, Roesch EB, Pizzagalli DA: Electrophysiological correlates of spatial orienting towards angry faces: A source localization study. Neuropsychologia. 2008, 46: 1338-1348. 10.1016/j.neuropsychologia.2007.12.013.

    Article  PubMed Central  PubMed  Google Scholar 

  43. Blau VC, Maurer U, Tottenham N, McCandliss BD: The face-specific N170 component is modulated by emotional facial expression. Behav Brain Funct. 2007, 3: 7-10.1186/1744-9081-3-7.

    Article  PubMed Central  PubMed  Google Scholar 

  44. Krombholz A, Schaefer F, Boucsein W: Modification of N170 by different emotional expression of schematic faces. Biol Psych. 2007, 76: 156-162. 10.1016/j.biopsycho.2007.07.004.

    Article  Google Scholar 

  45. Makeig S, Bell AJ, Jung TP, Sejnowski TJ: Independent component analysis of electroencephalographic data. Advances in Neural Information Processing Systems. Edited by: Touretzky D, Mozer M, Hasselmo M. 1996, Cambridge, MIT Press, 8D: 145-151.

    Google Scholar 

  46. Näätänen R: Attention and Brain Function. 1992, Hillsdale, N.J: Erlbaum

    Google Scholar 

  47. Mäntysalo S, Näätänen R: The duration of a neuronal trace of an auditory stimulus as indicated by event-related potentials. Biol Psych. 1987, 24: 183-195. 10.1016/0301-0511(87)90001-9.

    Article  Google Scholar 

  48. Kiss M, Eimer M: ERPs reveal subliminal processing of fearful faces. Psychophysiology. 2008, 45: 318-326. 10.1111/j.1469-8986.2007.00634.x.

    Article  PubMed Central  PubMed  Google Scholar 

  49. Eimer M, Holmes A: Event-related brain potential correlates of emotional face processing. Neuropsychologia. 2007, 45: 15-31. 10.1016/j.neuropsychologia.2006.04.022.

    Article  PubMed Central  PubMed  Google Scholar 

  50. Junghöfer M, Bradley MM, Elbert TR, Lang PJ: Fleeting images: a new look at early emotion discrimination. Psychophysiology. 2001, 38: 175-178. 10.1017/S0048577201000762.

    Article  PubMed  Google Scholar 

  51. Schupp HT, Stockburger J, Codispoti M, Junghöfer M, Weike AM, Hamm AO: Selective visual attention to emotion. J Neurosci. 2007, 27: 1082-1089. 10.1523/JNEUROSCI.3223-06.2007.

    Article  CAS  PubMed  Google Scholar 

  52. Schupp HT, Öhman A, Junghöfer M, Weike AI, Stockburger J, Hamm AO: The facilitated processing of threatening faces: an ERP analysis. Emotion. 2004, 4: 189-200. 10.1037/1528-3542.4.2.189.

    Article  PubMed  Google Scholar 

  53. Johnston PJ, Katsikitis M, Carr VJ: A generalised deficit can account for problems in facial emotion recognition in schizophrenia. Biol Psych. 2001, 58: 203-227. 10.1016/S0301-0511(01)00114-4.

    Article  CAS  Google Scholar 

  54. Leppänen JM, Tenhunen M, Hietanen JK: Faster choice-reaction times to positive than to negative facial expressions. J Psychophysiol. 2003, 17: 113-123. 10.1027//0269-8803.17.3.113.

    Article  Google Scholar 

Download references


The authors thank Dr. Petri Juujärvi for his comments to the previous version of this manuscript, Riina Meriläinen-Juurikka for preparing of the stimulus materials and Laura Karsio for her help in data gathering. The study was supported by the Academy of Finland (project no. 7127595 to PA and project no. 1111850 to JKH) and the Emil Aaltonen Foundation (PA).

Author information

Authors and Affiliations


Corresponding author

Correspondence to Piia Astikainen.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors' contributions

PA and JKH designed the experiment. PA collected and analyzed the data and drafted the initial manuscript version. PA and JKH wrote and accepted the final version of the manuscript.

Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

Authors’ original file for figure 1

Authors’ original file for figure 2

Authors’ original file for figure 3

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Astikainen, P., Hietanen, J.K. Event-related potentials to task-irrelevant changes in facial expressions. Behav Brain Funct 5, 30 (2009).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: