Abstract
To investigate neural mechanisms of human psychology with electroencephalography (EEG), we typically instruct participants to perform certain tasks with simultaneous recording of their brain activities. The identification of task‐related EEG responses requires data analysis techniques that are normally different from methods for analyzing resting‐state EEG. This review aims to demystify commonly used signal processing methods for identifying task‐related EEG activities for psychologists. To achieve this goal, we first highlight the different preprocessing pipelines between task‐related EEG and resting‐state EEG. We then discuss the methods to extract and visualize event‐related potentials in the time domain and event‐related oscillatory responses in the time‐frequency domain. Potential applications of advanced techniques such as source analysis and single‐trial analysis are briefly discussed. We conclude this review with a short summary of task‐related EEG data analysis, recommendations for further study, and caveats we should take heed of.
1 Introduction
Psychologists are particularly interested in revealing neural mechanisms of mental functions. Electroencephalography (EEG) is an increasingly popular, non‐invasive method to investigate how such functions are implemented in the human brain, due to its advantages in millisecond temporal resolution and financially acceptable costs. To achieve a better understanding of human mental functions, perplexing signal processing techniques are required to analyze EEG data and extract relevant EEG measures. To facilitate the applications of EEG techniques for psychologists, we introduce useful EEG signal processing techniques with MATLAB scripts (provided in the supplementary material) in two reviews.
We have extensively discussed the methods to analyze resting‐state EEG data in the previous review included in this special issue [1]. In this review, we focus on the techniques to analyze task‐related EEG data, where participants perform certain tasks while their brain activities are simultaneously recorded as EEG signals. To do so, we first highlight the importance of collecting high‐quality EEG data and describe the differences of preprocessing pipelines for task‐related EEG and resting‐state EEG. We then take a closer look at the techniques to extract event‐related potentials in the time domain and event‐related oscillatory responses in the time‐frequency domain (Fig. 1). Meanwhile, we briefly introduce the potential applications of some advanced techniques such as source analysis and single‐trial analysis. Finally, we conclude with a short summary of task‐related EEG techniques, recommendations for further study, and caveats that we should take heed of. Readers are strongly encouraged to first read relevant content in our review on methods for resting‐state EEG [1], where data preprocessing and the Fourier transform have been introduced with more details.

Overview of task‐related EEG data analysis. Raw EEG data are preprocessed to obtain the denoised data, on which time domain and time‐frequency analysis are performed. For time domain analysis, the amplitude and latency can be extracted to quantify the ERP waveform. Methods to measure the former include the simple peak amplitude, local peak amplitude, and mean amplitude; the latter can be measured using the peak latency method and 50% area latency method. For time‐frequency analysis, important measures include power‐based measures (magnitude, power, and power spectral density) and phase‐based measures (phase‐locking value, phase lag index, and weighted phase lag index).
2 EEG preprocessing
Besides neuronal activity, raw EEG data also contain physiological artifacts and non‐physiological noise. Therefore, the preprocessing of raw EEG data is crucial to denoise, remove artifacts, and ultimately improve the signal-to‐noise ratio (SNR) prior to further analysis. The preprocessing procedure for task‐related EEG is similar to that for resting‐state EEG (Table 1) [2], and their differences were highlighted in
Summary of preprocessing procedures for task‐related EEG data.
2.1 Good data first
Before discussing preprocessing differences, we are obliged to emphasize the importance of data quality, which determines the reliability and repeatability of a study. To collect good data, we should design our study rigorously to ensure that the EEG measures being selected are necessary to test our research hypothesis.
As psychologists, we are particularly interested in how cognitive functions are supported by the electrical activity generated from neuronal populations, rather than the physiological properties of the brain. We, therefore, consider EEG as a tool for us rather than a goal in and of itself. Moreover, the relationship between EEG activity and its associated cognitive processes can only be guaranteed by rigorous experiments. A poor experimental design will invalidate any data collected, so much so that they are unable to either support or falsify our research hypothesis. To make sure that EEG activity is relevant to our research goal, we need to ask three questions [3]: (1) Does the experimental design necessarily and sufficiently elicit the psychological process we are interested in? (2) Are all relevant confounding factors satisfactorily controlled? (3) Does the design meet the standards of EEG data recording and publication?
The first question requires us to ponder over experimental stimuli. For example, the stimulus intensity should be strong enough to elicit painful sensation if we want to investigate the neural mechanisms of pain processing with EEG. The second question requires that, ideally, different conditions should be identical except for the independent variables we intentionally manipulate. However, this goal is almost impossible in real life [4]. What we can do is to identify as many important confounders as possible and then try our best to control them. Generally, properties of sensory stimuli should be kept constant across conditions, unless they are the targeted independent variables. The final question invites us to contemplate issues such as eye and body movements, number of trials, time interval between trials, and stimulus probability. Of these issues, eye and body movements are of particular interest. While preprocessing can definitely help denoise EEG data, it is impossible to convert “bad data” into good data. The best practice is to acquire clean data with as little noise as possible [5]. A useful trick to do so is to minimize unwanted eye and body movements during EEG data collection.
2.2 Task‐related EEG preprocessing
When collecting task‐related EEG data, we employ a specific task wherein participants are presented with certain stimuli and/or asked to make responses. After being collected, the raw data are subject to preprocessing. As mentioned before, the preprocessing procedure for task‐related EEG is virtually the same as that for resting‐state EEG. Rare exceptions include epoch extraction, baseline correction, and bad epoch identification.
Event markers are well‐defined and even mandatory for task‐related EEG data collection since the stimuli presented to participants or the responses from participants are typically time-locked. As a result, we can naturally segment the continuous data into fixed‐length epochs containing both prestimulus and poststimulus periods (e.g., −200 ms to 800 ms). Here, the time point 0 (corresponding to an event marker) is determined by our experimental design, standing for the onset of a stimulus or the occurrence of a response. Epoching thus transforms the two‐dimensional raw data (electrodes, time) into the three‐dimensional epoched data (electrodes, time, epochs).
Since a baseline is also well‐defined (i.e., prestimulus period), we perform a baseline correction procedure immediately after extracting epochs from the continuous EEG data. Baseline correction is achieved by subtracting the average prestimulus voltage from the entire epoch in order to minimize voltage offsets and drifts due to skin hydration, skin potentials, and static charges in the electrodes [5].
Bad epochs in resting‐state EEG data are mostly those with grave artifacts. In task‐related EEG, however, the epochs incompatible with behavioral measure requirements can be regarded as bad epochs, too. For example, the epochs where a participant responds erroneously might be deleted from further analysis.
3 Task‐related EEG processing
In this section, we mainly discuss how to process task‐related EEG data in the time domain and time‐frequency domain. More advanced techniques, such as source analysis and single‐trial analysis, are mentioned briefly. Note that the advanced methods introduced in our review on resting‐state EEG data analysis can also be applied to task‐related EEG data [1].
3.1 Time domain analysis
Task‐related EEG processing in the time domain aims to extract event‐related potentials (ERPs), which show how EEG measures change as time proceeds. Spontaneous EEG activity is disturbed when a physical or mental event happens [6]. For example, when being delivered a painful stimulus in a single trial, a participant feels the pain and EEG signals, simultaneously recorded from the participant, would show some changes apart from the background EEG activity. To isolate the event‐related changes of EEG signals relative to the naturally occurring EEG activities, we simply repeat the same trial multiple times, align them to the starting time point of the targeted event, and average all the single trials after baseline correction. Since the spontaneous responses tend to fluctuate randomly, the mean response across trials would be close to zero if the number of trials is sufficiently large or, ideally, infinite, thus leaving only the eventrelated responses [Fig. 2(A)]. After obtaining the within‐participant averaged responses, we then average the responses across all participants to obtain the group‐level ERP waveform for each condition or <fig/>

Time domain analysis. (A) Comparison between single-trial event‐related potential (ERP) waveforms (colored dotted thin lines) and the averaged ERP waveform (bold black line). Single-trial ERP waveform contains lots of noise, but simple averaging makes random noise cancel each other out and the ERP responses stands out accordingly. (B) ERP waveform and topography. Ps is the point standing for the simple peak amplitude within the dashed box, whereas Pl represents the local peak amplitude. The mean amplitude of P2 is, by definition, (
The aforementioned logic of across‐trial averaging is simple and straightforward, yet there are many intricate issues after the averaging operation. To quantify the parameters in ERPs, we typically measure the amplitude and/or latency of an ERP component for each participant and each condition or group. To measure them appropriately, however, is a big challenge for researchers.
3.1.1 Measuring amplitudes
There are at least two ways of measuring the amplitude of an ERP component. The first one is to measure the voltage at the maximum point of a waveform within a specific time window, which can be called the “peak amplitude method” [5, 7]. The other is to calculate the mean voltage of all points in a defined time window, which can be termed the “mean amplitude method” [5].
The peak amplitude method is the oldest quantification of amplitudes and is used frequently [8 –10]. However, the amplitude measured by this method is sometimes counter‐intuitive, especially when the process is implemented automatically. Fig. 2(B) illustrates clearly why the most negative point within a time window, Ps, may not be properly used to define the negative peak. Normally, we would call Pl as the negative peak instead of Ps. As Luck pointed out, we can define Pl as the “local peak”, that is, the largest point in a measurement window with upward‐going and downward‐going deflections at its left and right side [5]. In contrast, absolute maximum or minimum points can be defined as the “simple peak”.
Generally, we want to measure the local peak amplitude, but not the simple peak amplitude. It is, therefore, important that we do not confuse one with the other. Mathematically, if the maximum (or minimum) point is not at the edges of a given time window, the simple peak amplitude is identical to the local peak amplitude. In practice, we can plot the ERP waveform to identify whether the maximum point occurs at the edges. If not, there is no need to distinguish them at all, and the traditional maximum algorithm produces the local peak amplitude.
The local peak method, however, is not without its own problems [5]. First and foremost, the local peak amplitude is positively biased by the noise level. That is, the measure is generally larger than the true peak amplitude when there are noises; the larger the noise, the larger the discrepancy. Unfortunately, EEG data are noisy, even all the preprocessing procedures have been performed. The noise level partly relies on the number of trials and participants, as well as the health conditions of participants. As a result, comparing the local peak amplitudes between markedly different conditions or groups can be problematic. Second, the local peak amplitude is attenuated by the latency jitter. If the latency varies across trials or participants, which is often the case, the average local peak amplitude would be smaller than the single‐trial amplitude [as illustrated in Fig. 2(A)]. This problem renders amplitude comparisons invalid when there are significant differences in latency variability and the number of trials between conditions or groups. Third, the local peak amplitude in the average ERP waveform is not a faithful representation of the average of local peak amplitude at the single‐participant level, unless different peak latencies are aligned to the same time point [11]. Without peak latency alignment, the average waveform may not reach its peak at the same time as the single-participant waveforms do; thus, the peak amplitude obtained from the average waveform is smaller than the average of single‐participant amplitudes in most cases.
By contrast, the mean amplitude is not (at least not easily) subject to any of the three problems. However, the mean amplitude method also has its own issues. The biggest drawback is that the time window, if too wide, may contain responses of other peaks or components, and thus the resulting mean amplitude is inaccurate. Fig. 2(B) illustrates a possible scenario where the mean P2 peak amplitude (area
3.1.2 Measuring latencies
Latency is the time interval between the starting time point of the targeted event (e.g., stimulus onset) and a specific ERP component. There is also more than one way of measuring latency. We can simply measure the time point where the waveform reaches its peak, or estimate the midpoint of a component by locating the time point which divides the area within the corresponding time window into two halves. The former latency is called the “peak latency”, while the latter is called the “50% area latency” [5].
These two measures of latency are based on the corresponding measures of amplitude, and thus inherit the strengths and weaknesses of the amplitude measures (see above). One thing to notice is that the peak latency is more sensitive to noises than the 50% area latency. The consequence is that the peak latency shows a larger variance, leading to lower statistical power. However, the 50% area latency is sensitive to the definition of the time window for extracting the ERP component, which would result in the inaccuracy of the measurement of ERP latency.
3.1.3 Practical issues
(1) How many trials do we need?
To derive a reliable ERP waveform, we need a sufficient number of trials [3, 5]. But how can we define “sufficient”? Unfortunately, we do not have any magic number for this question. The effect size, SNR, type of analysis, as well as many other factors determine the minimal number of trials needed to obtain a reliable and robust ERP waveform [5, 12 –14]. For a large ERP component, tens of trials would be enough; for a small ERP component, we may need hundreds of trials. One thing to remember is that the SNR improves in proportion to the square root of the trial number [5]. That is, to double the SNR, we should quadruple the number of trials. Therefore, if we already have hundreds of trials, it would be unnecessary to increase the number to a thousand. Instead, we should attempt to lower the noise level by other means.
(2) Do the baseline data reflect “pure” spontaneous activity?
No. To minimize voltage offsets and drifts due to skin hydration, skin potentials, and static charges in the electrodes, we correct the poststimulus data by subtracting the baseline mean from them, immediately after epoching the data [5]. However, the prestimulus baseline may be contaminated by the ERPs from the preceding trial or participant’s anticipations of the onset of the stimulus. Since the baseline‐corrected data are the difference between the uncorrected data and the baseline mean, the poststimulus responses are influenced by the baseline activity. We, therefore, should look at the baseline activity to ensure that the effect of interest is not an artifact caused by unwanted group or condition differences in the baseline. In practice, any differences that begin in the baseline or within 100 ms of stimulus onset may be spurious [5].
(3) What reference electrode should we choose?
The reference electrode has a huge impact on the ERP waveform. But there is no universally agreed‐upon reference electrode that fits well with all research questions. Cz, Fz, linkedears, linked‐mastoids, the ipsilateral‐ear, the contralateral‐ear, the tip of the nose, the average of all electrodes, and a point at infinite have been reported in the literature as the reference [15 –17]. In general, we should avoid the reference sites that are heavily biased toward one hemisphere, near the electrode of interest, or extremely noisy [5]. A practically‐accepted option is to choose the conventional one used in the relevant research field, thus allowing for comparisons of the results with those in previous studies.
(4) How can we visualize the results?
In the time domain, we have information about electric potentials across time and electrodes. We thus can depict graphs showing how electrical potentials vary across time points (i.e., ERP waveform) and how electrical potentials vary across electrodes (i.e., ERP topography). Providing these two kinds of graphs are mandatory for ERP studies [7].
ERP waveforms display the temporal characteristics of voltage without spatial characteristics [see the waveform in Fig. 2(B)]. Oftentimes we only choose one electrode of interest or compute the average waveform of several surrounding electrodes. We can, of course, draw the ERP waveform for every electrode and put them in the same graph (e.g., butterfly plots), but that might not be helpful as the reader may not be able to see the details like the difference of ERP waveforms between different conditions or groups. An oft‐used graph derived from ERP waveforms is the differential waveform, which could be created by subtracting one ERP waveform from another, showing how the difference changes over time.
ERP topographies, on the other hand, depict how the potentials are distributed across the scalp electrodes [see the topography in Fig. 2(B)]. We need to control another variable in this sort of graph, that is, time. The time point chosen usually corresponds to the latency of the component we are interested in. Alternatively, we can calculate the mean potentials within a time window and plot the corresponding scalp topography. One rule to remember is that the time variable (e.g., the interval for computing the average) must be identical for all the electrodes [7]; otherwise, the generated scalp topography would be confusing and difficult to interpret. However, there is no restriction on how many topographies we can draw so long as every one of them follows the rule. If necessary, we could provide a series of topographies for several time points (or several time intervals) for the ERP waveform.
3.2 Time‐frequency analysis
ERP waveforms are the neurophysiological response that is both time‐locked and phase-locked to the stimulus onset. However, there are some important event‐related brain responses that are time‐locked, but not phase‐locked to the stimulus onset. To derive these brain responses, we need to perform the time‐frequency analysis, which characterizes the spectral perturbations of the data over time. Please note that Fourier transform, which is often used to extract the spectral features of resting‐state EEG [1], would not be applicable in the task‐EEG data processing, since Fourier transform ignores the temporal information and hinges on the unrealistic assumption of stationary, that is, the frequency features of the signal are assumed to be invariant over time [Fig. 3(A)].

Time‐frequency analysis. (A) Shortcomings of the Fourier transform. Signals
To overcome this disadvantage, we need to perform the time‐frequency analysis. Commonly used time‐frequency methods include the short‐time Fourier transform (STFT), wavelet transform, and Hilbert transform. Practically, they produce highly similar results and can be considered fundamentally equivalent [18, 19]. Therefore, we only focus on the short‐time Fourier transform in the next section, a method more intuitive and intelligible.
3.2.1 Basic ideas of the STFT
The intuition of the STFT is straightforward. It cuts a single trial into numerous brief overlaping segments (i.e., time windows) where the assumption of stationarity supposedly holds. The STFT extracts the spectrum in every time window using the Fourier transform and assign the spectrum to the midpoint in the entire window, thereby providing temporal information. The STFT thus successfully addresses — or assumes away — the problems of unrealistic stationarity assumption and the lack of temporal information of the Fourier transform.
Mathematically, the STFT for a discrete signal
where
“1. Select a window function of finite length. 2. Place the window on top of the signal at 3. Segment the signal using this window. 4. Compute the spectrum of the windowed data segment. 5. Incrementally slide the window along time. 6. Go to step 3, until the window reaches the end of the signal.”
Note that what we actually select is a window function rather than a simple window to diminish edge artifacts. Typical window functions include Hann (sometimes called Hanning), Hamming, and Gaussian windows. The Hann window is preferred since it, by definition, tapers the data to zero at its both sides [13]. The process of windowing and sliding the window along time indicates a vital issue in time‐frequency analysis: The spectra of contiguous time points would be strikingly similar, meaning the time precision of time‐frequency analysis is inherently by far poorer than the time domain analysis. Surely we can improve time resolution by narrowing the time window, but by doing so, we would have to compromise the frequency resolution, a phenomenon known as the uncertainty principle in signal processing [20].
3.2.2 Power‐based measures
The resulting
Magnitude is conceptualized as the absolute value of
where
PSD is a derived measure from power and calculated as
where
Since all three measures are, by definition, non‐negative, we solve the non‐phase locked problem in time domain analysis, which results from the cancellation of positive and negative potentials. This problem is successfully solved only if we do the STFT at the single‐trial level.
With these measures at hand, we can define two important concepts describing task‐related power changes relative to the baseline activity: event‐related synchronization (ERS) and eventrelated desynchronization (ERD). The former means the poststimulus oscillatory power is greater than the baseline power; the latter, on the other hand, means the poststimulus oscillatory power is smaller than the baseline power. The exact formulas computing them rely on the baseline correction method, a topic we discuss later in
3.2.3 Phase‐based measures
One particularly important measure we can extract from time‐frequency results is phase, which is the angle between the real part and the imaginary part of the complex series
where
PLV can also be used to quantify connectivity between electrodes. The formula calculating PLV is
where
PLV suffers from the problem of a common source, meaning that spurious connectivity occurs if activities from electrodes
where
where
In practice, these phase‐based measures (e.g., PLV, PLI, and wPLI) were computed on the band‐passed filtered data. Note that these measures range from 0 to 1 and do not follow the normal distribution. When we compare phase‐based connectivity measures between conditions or groups, the results are bounded between −1 and 1 and are not normal-distributed, either. To facilitate meaningful statistical tests, we can Fisher‐
where ln() is the natural logarithm function, and
3.2.4 Practical issues
(1) How long should the time window for analysis be?
The length of the time window determines the trade‐off between time and frequency resolution. Shorter windows improve time resolution, but worsen frequency resolution; and vice versa. The window length also restricts the lowest frequency we can analyze. In theory, we can never extract the frequency information whose period is longer than the time window [13]. For example, the lowest frequency is 5 Hz if the window is 200 ms. To achieve a better SNR, we would want the window to contain two or three periods of certain frequency information, which further limits the lowest frequency we extract. In practice, we generally try out several different window lengths and choose the one that displays the best visually and statistically reasonable results [21].
(2) One STFT, one window function?
No. We can apply multiple window functions to the same data segment and average the spectrums, a method called “multitaper” [25]. In practice, to do multitaper analysis, we employ Slepian tapers, which are orthogonal to each other, rather than any arbitrarily chosen window functions [26]. The advantage of this method is that it improves the SNR, making it useful for analyzing high‐frequency activity and single‐trial data [13]. However, a higher SNR is achieved at the expense of worse time precision. The multitaper method thus is not particularly helpful for analysis that focuses on lower frequencies (e.g., below 30 Hz) or demands precise timing.
(3) How can we do baseline correction?
As in time domain analysis, baseline correction is necessary for time‐frequency analysis. One crucial reason is that raw time‐frequency power values follow the so‐called “1/
There are four major approaches to perform the baseline correction: subtraction, percentage, decibel, and
Subtraction :
Percentage :
Decibel :
where Mean(Powerbaseline) is the average baseline power and SD(Powerbaseline ) is the standard deviation of baseline power. It is noteworthy that the baseline need not be the entire prestimulus period. In fact, we had better choose a narrower baseline. The algorithm of the STFT implies that the power at the stimulus onset reflects the EEG activity of a time window whose midpoint is the stimulus onset. As a result, to avoid any contamination from the poststimulus activity, we should choose a baseline such that the distance between its ending and the stimulus onset is not larger than one‐half of the time window [27]. For example, if the window length is 200 ms, the baseline needs to end before −100 ms.
As for the approach per se, methods correcting the data by taking relative rather than absolute changes, that is, the percentage, decibel, and
(4) Should we compute connectivity measures over time or trial?
As we mentioned before, phase‐based measures, such as PLV, PLI, and wPLI, can be computed in two ways: over time or over trial. We are then forced to choose one of them. These two ways both have their own advantages and limitations [13]. Connectivity over time is less sensitive to trial‐to‐trial jitter of event timing and phase angle differences between trials. However, the epoch should be long enough for this kind of connectivity. Consequently, connectivity over time is more appropriate if we focus on high‐frequency connectivity. On the other hand, connectivity over trial puts no constraint on epoch length and achieves a higher level of evidence. Moreover, since connectivity is computed for every time point, the temporal precision is by far higher for connectivity over trial. Indeed, the STFT is not well suited for computing this connectivity; other methods like wavelet transform and Hilbert transform are preferred. The limitation is that connectivity over trial is highly sensitive to the event timing. Therefore, any uncertainty of timing may invalidate this measure.
(5) How can we visualize the results?
Time‐frequency analysis performed on a single trial produces three‐dimensional results: time, frequency, and certain measure (e.g., magnitude). A simple two‐dimensional plot thus is not adequate for visualization. To overcome this challenge, we could color the plot to add an extra dimension representing the measure we have computed [Fig. 3(B)]. Coloring the time‐frequency plot, however, is not easy. Please note that power follows the 1/
In addition, we need to average the data across trials or participants. Although time-frequency results are available for every electrode, we may not want to show the results from all electrodes. Instead, we could choose one representative electrode or average several surrounding electrodes to plot time‐frequency results. To provide spatial information, we can draw a scalp topography by averaging power within the time‐frequency region of interest for each electrode [Fig. 3(C)].
3.3 Advanced techniques
Powerful as they are, time domain and time-frequency analysis are some routine analysis for task‐related EEG processing. To extract more information from the raw data, we need more advanced techniques like source analysis and single‐trial analysis.
3.3.1 Source analysis
One key feature of EEG data is that their spatial resolution is inherently poor. EEG data provide only the topographic distribution of electric potentials, not the spatial distribution of neural sources. Source analysis could be used to infer the latter from the former, which is a problem known as the inverse problem [29]. The inverse problem is notoriously difficult to solve. In actuality, this problem has no unique solution, meaning that more than one configuration of neural activity can produce the same topographic distribution and that we are theoretically unable to identify which configuration is absolutely correct for an EEG dataset. Localization of brain generators for specific electrical potentials is thus a challenging, if not impossible, task. Nevertheless, if conducted and interpreted carefully, source analysis provides valuable information about the brain regions generating the EEG activities recorded at the scalp, which is very helpful in a lot of basic and clinical applications. In practice, there are many methods to estimate the source, which are implemented in several freely- available softwares, for example, FieldTrip, sLORETA, and RECOR [29].
3.3.2 Single‐trial analysis
Traditionally, we average all single trials to boost the SNR and extract more accurate and consist brain features. However, the operation of averaging depends on an implicit and unrealistic assumption: The single‐trial response is largely invariant across trials [30]. We have mentioned in
4 Concluding remarks and caveats
EEG is an old but powerful technique for revealing the neural implementation of human psychology. In this paper, we have introduced some of the basic and advanced EEG signal processing methods for task‐related EEG. Important temporal (ERP amplitude and latency)
and spectral measures (PSD, PLV, PLI, and wPLI) can be extracted to answer specific research questions (Fig. 1). For those who want to learn EEG signal processing more comprehensively and deeply, we highly recommend three book‐length materials: (1)
Powerful as it is, we should not abuse EEG technique. Importantly, the proper use of EEG requires us to pay attention to the following critical issues.
4.1 Always good data first
The importance of rigorous experimental design and good data can never be overstated. Fancy data analysis will not save a poorly designed study. This is why we reiterate this point here again, even though we have discussed it in
4.2 Always checking assumptions
All mathematical models have their own assumptions and are only effective when the assumptions are satisfied. It is, therefore, always mandatory to check them. For example, the Fourier transform relies on the assumption of stationarity. When the signal is nonstationary, spectral analysis with the Fourier transform can lead to misleading results. We can check assumptions mathematically or visually. The former is surely stricter and more precise; the latter, however, is more intuitive. If the model is robust against violation of some assumptions, a visual inspection can be extremely useful. We thus encourage more frequent use of visualization in assumption checks.
4.3 Taking advantage of the common knowledge
There is currently no universally accepted standard for EEG analysis. Recommendations from researchers in different fields may vary tremendously, and it is hard to determine which one is the correct one. “Correctness” of the recommendation may even be ill‐defined since the criteria for perfect analysis are themselves controversial and, in some cases, inconsistent. The solution is to respect the common knowledge in the field. For example, if most researchers reference the data to the nose tip, we can safely do the same thing. At least, our results are comparable to those already published, and any divergence can be explained more easily without involving methodological inconsistency.
4.4 Basic analysis before advanced ones
In the present article, we have mainly introduced the basic analysis and briefly mentioned the more advanced ones. This practice is, of course, partly due to the abstruseness of advanced analysis, but it should also remind us that the basics are more important. Findings from advanced analysis are suspicious, though not definitely false, if not supported by basic results, for example, ERP or time‐frequency results. For non‐methodological studies, we believe that the focus should be on the basic analysis and the resulting measures, unless there are a priori reasons to do otherwise.
4.5 Describing analysis in the methods section
As a general requirement, we should describe our analysis in detail in the “Methods” section. For example, we should pro″vide: the type of electrode, interelectrode impedance, the locations of the recording electrodes, the reference electrode, the sampling rate, filtering parameters, artifact identification and rejection procedures, the length of baseline, the baseline correction method, ERP amplitude and latency determination methods, the type and length of the window in time‐frequency analysis, and the like [7, 13].
With these caveats in mind, we can exploit the power of EEG technique appropriately to gain valuable insights into how the brain supports and implements human psychology.
Supplemental Material
Supplemental Material, 10.26599_BSA.2020.9050018 - Demystifying signal processing techniques to extract task-related EEG responses for psychologists
Supplemental Material, 10.26599_BSA.2020.9050018 for Demystifying signal processing techniques to extract task-related EEG responses for psychologists by Libo Zhang, Zhenjiang Li, Fengrui Zhang, Ruolei Gu, Weiwei Peng and Li Hu in Brain Science Advances
Supplemental Material
Supplemental Material, BSA20200018_suplemental_1Matlab_scripts - Demystifying signal processing techniques to extract task-related EEG responses for psychologists
Supplemental Material, BSA20200018_suplemental_1Matlab_scripts for Demystifying signal processing techniques to extract task-related EEG responses for psychologists by Libo Zhang, Zhenjiang Li, Fengrui Zhang, Ruolei Gu, Weiwei Peng and Li Hu in Brain Science Advances
Supplemental Material
Supplemental Material, BSA20200018_Suplemental_2sub_stft - Demystifying signal processing techniques to extract task-related EEG responses for psychologists
Supplemental Material, BSA20200018_Suplemental_2sub_stft for Demystifying signal processing techniques to extract task-related EEG responses for psychologists by Libo Zhang, Zhenjiang Li, Fengrui Zhang, Ruolei Gu, Weiwei Peng and Li Hu in Brain Science Advances
Footnotes
Conflict of interests
The authors have declared that no competing interests exist.
Acknowledgments
This work was supported by the National Natural Science Foundation of China (No. 31822025, No. 31671141). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
