Participants
Twenty-eight volunteers (9 men, 19 women) with a mean age of 23.13 years (SD = 3.64), with normal or corrected to normal vision and with no psychological or neurological diseases were included in the experiment. All of them were right-handed according to the Edinburgh Handedness Inventory 43. The experiment was approved by the Ethical Committee of the Leibniz Research Centre for Working Environment and Human Factors and was conducted in accordance with the Declaration of Helsinki. All participants gave their written informed consent prior to the beginning of the experimental procedure. Please find additional information in the methods section in the supplementary material.
Procedure
In each trial, participants performed both a number classification and a working memory task. These tasks were framed by a relevance cue and a feedback screen, both supporting participants to adapt their mental effort in a task-specific way throughout the trial (see Fig. 1).
Each trial started with a relevance cue indicating on which of the two tasks participants should focus their attention in the following trial. The cue was a triangle with either the number five in it (indicating that the number classification task should be focused) or a triangle with vertical stripes indicating that the working memory task should be focused (see Fig. 1). The relevance cue was presented for 500 ms and was followed by a fixation dot for 2500 ms. After that, the working memory task started with two Gabor patches (memory items) appearing on the left and right side of the screen for 400 ms, followed by a fixation dot for 2200 ms. Participants were instructed to remember the orientation of both patches. During the retention interval of the working memory task, participants performed a cued number-classification task, in which they classified a digit (number classification task target) as either smaller or bigger than five or as odd or even by pressing one of two buttons. A letter (classification cue, A, B, X or Y) appearing on screen for 200 ms indicated which of two classifications needed to be done. Bigger/smaller five was assigned to A or B and odd/even to X or Y for half of the participants, for the other half it was the other way round. The following classification digit (1, 2, 3, 4, 6, 7, 8, or 9) was presented for 200 ms. Its presentation was framed by a fixation dot period of 1000 ms before its appearance and 2600 ms after its disappearance. Responses to the number classification task were registered in the time interval starting 200 ms and ending 1800 ms after the onset of the digit. Responses outside this time interval were regarded as missing answers. Then, the working memory task continued with a retro-cue indicating which of the two Gabor Patches would be asked for in the following working memory retrieval phase. The retro-cue was an arrow, presented for 200 ms, pointing to the side (left or right) on which the to be retrieved Gabor patch had been presented before. It was followed by a fixation dot for 1000 ms. After that, the working memory target, a Gabor Patch with a random orientation, appeared on the screen. Participants were instructed to rotate the orientation of this Gabor patch in the same orientation as the retro-cued memory item using a computer mouse. If no response was made within the time interval of 200 ms to 4500 ms after the appearance of the target stimulus, the trial continued and the working memory answer was registered as missing. After a fixation dot interval of 200 ms, the trial ended with a feedback score appearing on the screen for 300 ms. The feedback score was calculated based on performance in both tasks, however the performance in the cued, i.e., in the more important, task was weighted three times as much as the performance in the un-cued task (for details see below). The ITI was 1200 ms.
After arrival, subjects were informed in detail about the experimental procedure. They were instructed to always perform both tasks but to put more effort in the respectively cued task than in the un-cued task. To motivate them to follow the task instructions, participants were additionally told that in case they collected more feedback points than 50% of all participants, they had the opportunity to take part in a lottery. In fact, the lottery was done including all participants to not disadvantage less skillful subjects. Three participants won money in the lottery (€50, €45 and €30, respectively) which they got in addition to their normal study incentive. In addition, it was explained in detail how feedback scores were calculated.
Feedback points (FP) were assigned to responses in the following way: The maximal amount of feedback points (max) was 75 for the cued task and 25 for the un-cued task. In each trial, feedback points received for both tasks were added, and the sum (number between zero and 100) was depicted on screen accompanied by a percentage sign. Thus, performance in the cued task weighted three time as much as the performance in the un-cued task. In the number classification task, missing and incorrect answers always resulted in zero points. A response time of 200–400 ms with respect to the number classification digit resulted in the highest possible score (max). In the working memory task, missing answers or degree deviations between target and cued memory item bigger than 45° resulted in zero points. Apart from that, feedback points were assigned with the following formulas to reaction times (RT) or degree deviations:
$${FP}_{Number classification task}=max-\left[\frac{max}{(1800 ms-400 ms)}*(RT-400 ms)\right]$$
(1)
$${FP}_{Working memory task}=max-\frac{max}{45^\circ }*degree deviation$$
(2)
Thus, in the number classification task, subjects got a high amount of feedback points by answering as quickly and as accurately as possible. In the working memory task, they should answer as accurately as possible in order to be successful. In general, they had a higher likelihood to reach high scores if they varied their effort throughout the trial.
The experiment consisted of 10 blocks with 60 trials each, providing the possibility for breaks in between blocks. Trial order within each block was randomized. The levels of the experimental factors prioritized task (working memory task more important vs. number classification task more important), retro-cue orientation (left vs. right) and number classification type (odd / even vs. smaller / bigger five) appeared equally often within each block and co-occurrences were balanced. Participants needed roughly 2.5 hours for the 10 blocks. During this time, EEG was continuously recorded. Please find additional information regarding the experimental procedure and stimulus material in the supplementary material.
Eeg Data Acquisition And Preprocessing
EEG was recorded using BrainVision Brainamp DC amplifier, BrainVision Recording software and 64 Ag/AgCl actiCAP slim active electrodes arranged according to the international 10–10 system (BrainProducts, Gilching, Germany). The ground electrode was placed at AFz and FCz served as online reference. Data was recorded with a sampling rate of 1000 Hz and impedances were kept below 10 kΩ during recording.
EEG preprocessing was based on EEGLAB 44 in combination with custom MATLAB code (R2020a, The MathWorks Inc., Natick, Massachusetts). Data were down-sampled to 200 Hz, high-pass filtered (FIR filter with Hamming window, order: 1321, transition band-width: 0.5 Hz, -6dB cutoff: 0.25 Hz, passband-edge: 0.5 Hz), low-pass filtered (FIR filter with Hamming window, order: 67, transition band-width: 10 Hz, -6dB cutoff: 45 Hz, passband-edge: 40 Hz) and re-referenced to CPz in order to be able to include the former reference position FPz in the analysis. Next, channels with poor data quality were detected and excluded from further analysis based on kurtosis and probability criteria (M = 0.54 channels, SD = 0.75 channels). After re-referencing to common average reference, data were segmented in epochs starting 3700 ms prior to and ending 13700 ms after the relevance cue. A baseline subtraction was done using the mean of each trial. Then, trials containing artifacts were detected and deleted automatically. On average, 10.71% of trials were removed (SD = 7.92%). An independent component analysis (ICA) was performed on a random sample of the data (200 epochs). Prior to ICA, a dimensionality reduction was done, using a standard PCA and removing the 3 dimensions that explained least variance, to deal with potential linear dependencies amongst channels. After the ICA, on average, 31% independent components (SD = 10.47%) were removed per participant using the EEGLab plugin ICLabel (Pion-Tonachini et al., 2019). Finally, previously removed channels were interpolated. Please find more details on EEG preprocessing in the method section of supplementary material.
Time Frequency Analysis
Data were decomposed into frequency bands using complex Morlet wavelet convolution as described in Cohen 46. The frequencies of the wavelets ranged from 2–20 Hz (19 wavelets, linearly spaced). They had a full width of half maximum (fwhmf, Cohen, 2019) between 0.75 Hz for the lowest frequency and 4.25 Hz for the highest frequency, which corresponds to an fwhmf of 1000 to 200 ms in the time domain. Power values were extracted from the complex convolution result. To diminsih edge artifacts, the first and the last 700 ms of each epoch was deleted after the convolution. Raw power values were decibel-normalized using condition-unspecific baseline values based on the mean of the trials belonging to both conditions in the time interval from 700 to 200 ms prior to the relevance cue.
Lateralized power was computed based on raw power by pairing each electrode on the left hemisphere with its corresponding channel on the right hemisphere. For each electrode pair, we calculated one power average over all trials for contralateral electrodes (i.e., right hemisphere electrodes on trials in which the working memory task cue pointed to the left and left hemisphere electrodes on trials in which the working memory task cue pointed to the right) and one average for ipsilateral electrodes. Trials used for analysis were randomly drawn in a way that each combination of retro-cue direction (left vs. right) and response side in number classification (left vs. right) occurred equally often. To get a normalized asymmetry measure, we then computed a lateralization index 48–51 as follows:
(3)\(\text{L}\text{a}\text{t}\text{e}\text{r}\text{a}\text{l}\text{i}\text{z}\text{a}\text{t}\text{i}\text{o}\text{n} \text{i}\text{n}\text{d}\text{e}\text{x}=\frac{\text{i}\text{p}\text{s}\text{i}\text{l}\text{a}\text{t}\text{e}\text{r}\text{a}\text{l} \text{p}\text{o}\text{w}\text{e}\text{r} - \text{c}\text{o}\text{n}\text{t}\text{r}\text{a}\text{l}\text{a}\text{t}\text{e}\text{r}\text{a}\text{l} \text{p}\text{o}\text{w}\text{e}\text{r}}{\text{i}\text{p}\text{s}\text{i}\text{l}\text{a}\text{t}\text{e}\text{r}\text{a}\text{l} \text{p}\text{o}\text{w}\text{e}\text{r} + \text{c}\text{o}\text{n}\text{t}\text{r}\text{a}\text{l}\text{a}\text{t}\text{e}\text{r}\text{a}\text{l} \text{p}\text{o}\text{w}\text{e}\text{r}}\)
The index was calculated for every electrode pair, frequency, time point and subject, separately. Based on previous literature 29,30,52, we were interested in posterior alpha asymmetry which has been observed to be a correlate of attentional shifts within working memory. We therefore averaged the index values across a posterior electrode cluster (T7/T8, C5/C6, C3(C4, C1/C2, TP9/TP10, TP7/TP8, CP5/CP6, CP3/CP4, CP1/CP2, P7/P8, P5/P6, P3/4, P1/P2, PO7/PO8, PO3/PO4, PO9/PO10, O1/O2) and across frequencies ranging from 8 to 15 Hz (see Fig. 3).
Decoding Analysis
We did separate decoding analyses for both experimental conditions, i.e. for trials in which the number classification task was more important and trials in which the working memory task was more important. Within these two conditions, we decoded from the scalp topography the trial’s number classification type (smaller/bigger five vs. odd /even).
Decoding was done both based on broadband ERP data as well as based on alpha power. The decoding procedure was the same in both analyses except for the features used for training the models: For ERP decoding, the preprocessed EEG data of all channels were used as features, whereas for alpha power decoding time-frequency decomposed data at eight, nine, 10, 11 and 12 Hz of all channels were used. Employing Matlab scripts provided by Bae & Luck 53, we calculated classification accuracies based on a cross-validation with linear support vector machines, applied to trial averages rather than to single trials to boost signal-to-noise ratio. See supplementary material for more details.
Statistical analysis
For details regarding statistical analysis of behavioral data see supplementary material. To statistically analyze EEG data, cluster-based permutation tests 54 were applied as they provide an elegant way to correct for multiple comparisons in high-dimensional EEG data.
In the non-lateralized time-frequency data analysis, the two experimental conditions (number classification task important and working memory task important) were compared at each pixel in the electrode x time point x frequency space to exploratively identify time points, frequencies and electrodes at which effort manipulation affects EEG activity. To this end, a cluster-based permutation test was conducted on decibel-normalized power values using the Matlab toolbox Fieldtrip 55. The test was computed across all frequencies obtained from time-frequency decomposition, i.e. across 2–20 Hz, across all 64 electrodes and across the whole epoch starting 3000 ms prior to and ending 13000 ms after the relevance cue (see also supplementary material).
Statistical analysis of decoding accuracies and power asymmetry values was done in two different ways: exploratively, using cluster-based permutation tests and hypothesis-driven, with Bayesian analyses applied to time intervals of interest. The cluster-based permutation tests were run over all time points of the trial between the relevance cue and 1000 ms after the onset of the working memory task target display. They were computed with the two Matlab functions cluster_test() and cluster_test_helper() which are part of the online code provided by Wolff et al. 56. See supplementary material for more details regarding these tests.
To test for general effects, cluster-based permutation tests were used to compare lateralization index against the case of zero asymmetry and decoding accuracies against chance level. To test for effects of the experimental manipulation, decoding accuracies and lateralization indices of working memory task important trials were compared against the respective measure obtained from number classification task important trials.
To facilitate result interpretation, a second approach was applied for decoding and lateralization index analysis: a time interval of interest analysis including Bayesian statistics. In frequentist statistics, higher p-values do not indicate more evidence for the null hypothesis. In contrast, Bayesian statistics quantifies the evidence in favor for one of both hypotheses (alternative hypothesis H1 and null hypothesis H0). The Bayes factor B01 is the likelihood ratio expressing this evidence. More exactly, B01 is the likelihood for the measured data given that H0 is true divided by the likelihood for the measured data given that H1 is true. Thus, B01 = 2 means that the data are twice as likely H0 than under H1. Keysers & Wagenmakers 57 suggest that whereas 1/3 < B01 < 3 provides not enough evidence for either hypothesis, B01 > 3 indicates substantial evidence for H0 and likewise, B10 < 1/3 indicates substantial evidence for H1 (see also Jeffreys, 58 ). A B01 > 10 is considered as strong evidence for H0. The time interval of interest analysis comprised two steps: We first calculated t-tests using Matlab. We then computed Bayes factors based on the outputs derived from the frequentist t-test as suggested by Morey & Wagenmaker 59, using a Matlab toolbox from Bart Krekelberg 60. Default effect size priors from the toolbox were used (Cauchy distribution scale = 0.707). Using this approach, we compared on the one hand lateralization indices versus zero / decoding accuracies versus chance levels and on the other hand values in number classification task important trials versus values in working memory task important trials. Two-sided one-sample t tests were computed in the first case and two-sided paired t tests in the other case.
Effect sizes were calculated for significant effects. η refers to adjusted partial eta squared 61. This is a bias-free effect size based on variance which is suited to show were pixels of a cluster are most pronounced. D refers to Cohen’s d. In some cases, both effect sizes are reported to provide a better comparability with other research.
Data / Code Availability Statement
Data and analysis code related to this manuscript will be made available online as an OSF repository upon publication.