Participants
Fifteen healthy participants were recruited for this experiment (mean age 29.7 ± 8.4 years, 6 females) with normal or corrected-to-normal vision and no history of neurological or psychiatric diseases. All participants were right-handed, as confirmed by a handedness questionnaire adapted from (Oldfield, 1971): mean laterality index of 85.0 ± 9.0. All gave informed written consent before participating in accordance with the declaration of Helsinki, and the study followed the safety guidelines for magnetic resonance imaging research on humans. The work was approved by the Ethics Committee of the Faculty of Medicine of the University of Coimbra.
fMRI data acquisition
Scanning was performed on a 3T Siemens Magnetom Prisma fit, using a 64-channel head/neck coil, at the Institute of Nuclear Sciences Applied to Health (ICNAS), University of Coimbra, Portugal. The scanning session started with the acquisition of one 3D anatomical magnetization-prepared rapid acquisition gradient echo (MPRAGE) pulse sequence (TR = 2530 ms, echo time (TE) = 3.42 ms, flip angle (FA) = 7°, 176 slices, voxel size 1.0 × 1.0 × 1.0 mm3, field of view (FOV) = 256 × 256 mm2).
The functional runs were acquired using a 2D multi-band (MB) gradient-echo (GE) echo-planar imaging (EPI) sequence from the Center for Magnetic Resonance Research, University of Minnesota (Release R016a). We tested four different temporal resolutions: TR = 0.5 s (MB factor = 6, FA = 53º, 42 slices), TR = 0.75 s (MB factor = 4, FA = 63º, 40 slices), TR = 1 s (MB factor = 3, FA = 68º, 42 slices), TR = 2.5 s (MB factor = 1, FA = 85º, 42 slices). The remaining parameters were matched: TE = 30.2 ms, interleaved slices with 0.5 mm gap, voxel size 2.5 × 2.5 × 2.5 mm3, FoV 192 × 192 mm2 (Table 1).
Table 1
Main parameters of the four fMRI sequences.
Sequence | 1 | 2 | 3 | 4 |
---|
TR (s) | 0.5 | 0.75 | 1 | 2.5 |
TE (ms) | 30.2 |
MB factor | 6 | 4 | 3 | 1 |
Voxel size (mm) | 2.5 |
Slice gap (mm) | 0.5 |
Number of slices | 42 | 40 | 42 | 42 |
Flip Angle (º) | 53 | 63 | 68 | 85 |
Bandwidth (Hz/px) | 2742 | 2742 | 2632 | 1994 |
Echo Spacing (ms) | 0.51 | 0.49 | 0.49 | 0.59 |
Excite pulse duration (us) | 2560 | 2560 | 2560 | 4840 |
EPI factor | 76 |
GRAPPA | Off | Off | Off | 2 |
LeakBlock | On | On | On | Off |
For mapping and correction of image distortions related to magnetic field inhomogeneities, we acquired a pair of spin-echo images with anterior-posterior (AP) and posterior-anterior (PA) phase encoding polarity with matching geometry and echo-spacing to each of the functional scans. These were acquired before the functional runs.
The participants’ physiological signals (respiration and pulse) were recorded during the functional runs using the scanner’s Physiological Measurement Unit (PMU). The respiratory signal was recorded at 50 Hz using a respiratory cushion, and the cardiac cycle was recorded at 200 Hz using a pulse sensor.
Functional tasks
We implemented three functional tasks based on a moving plaid stimulus. These stimuli are created by superimposing two gratings leading to a bistable percept: stimuli can be perceived moving coherently as a single surface (integration) or incoherently as two separate surfaces sliding over each other (segregation) (see supplementary materials and Fig. 1).
The first is a localizer for the region of interest - hMT+. For this purpose, we created three conditions: i. ‘Fixation’ - a fixation cross; ii. ‘Static plaid’ - a stationary plaid; iii. ‘moving plaid’ - a moving plaid, which is inherently a bistable stimulus (coherent versus incoherent, e.g. integration versus segregation). The run lasted for 2.9 min and was composed of nine trials with the sequential presentation of each condition for 6 seconds.
We named the following two tasks as ‘ambiguous’ and ‘unambiguous’ runs. These runs are composed of trials considering three conditions: ‘static’ (static plaid), ‘motion’ (ambiguous or unambiguous moving plaid), and ‘MAE’ (a period during which motion aftereffect is expected). During the ambiguous runs, the ‘motion’ condition showed the participants the moving plaid without any overlaid dots. As such, the stimulus is entirely ambiguous (the percept alternates between coherent and incoherent). Here, the participants were instructed to report the perceived type of motion (coherent or incoherent) using two buttons of a response box. During the unambiguous runs, the plaid is shown with overlaid dots moving either coherently down or incoherently inwards, which disambiguates the perception of the plaid (unambiguously coherent or incoherent). Based on the responses given by the participant in the previous ambiguous run, we manipulated the switches between coherent and incoherent motion in the unambiguous runs to match the previous responses precisely. This matched the time of each percept across both types of runs. The participants received the same instruction - to report the perceived type of motion at all times.
In this work, we considered the localizer and the four unambiguous runs, one for each temporal resolution.
fMRI data preprocessing
The data were organized according to the Brain Imaging Data Structure (BIDS) (Gorgolewski et al., 2016), using BIDSkit (Tyszka, 2023) and dcm2niix (Li et al., 2016). Results included in this manuscript were obtained after data preprocessing performed using fMRIPrep 23.0.2 (Esteban et al., 2023), which is based on Nipype 1.8.6 (Gorgolewski et al., 2011; Esteban et al., 2022). For a detailed description of the fmriPrep pipeline please see supplementary materials. To perform quality checks on the anatomical and functional images and extract metrics for further analysis and sequence comparison we ran mriqc 23.0.1 (Esteban et al., 2017). All subsequent analyses were performed in Python using Nilearn (Abraham et al., 2014).
fMRI data analysis
Regarding the localizer data, hMT + was functionally localized for each subject using a standard GLM analysis. The design matrix included predictors for all experimental conditions (‘fixation’, ‘static plaid', ‘moving plaid’) and confound regressors based on the mean and first derivative of the voxels in the CSF and the six head motion parameters and their derivatives. Temporal high-pass filtering (cut-off = 0.03 Hz) and spatial smoothing (FWHM = 6 mm) were applied. hMT + was manually selected on both hemispheres based on the activation map for the contrast ‘moving plaid > static plaid’ corrected with Bonferroni p = 0.05, cluster threshold k = 50, and confirmed using the hMT + mask of Neurosynth (Yarkoni et al., 2011)). Then, a spherical mask was designed around the center coordinates of each defined ROI (radius = 6 mm).
The first step to analyzing the unambiguous runs was the definition of the coherent and incoherent predictors. To this end, using the perceptual reports acquired via button presses with a temporal resolution of 60 Hz, we calculated the percentage of coherent reports over the total number of reports for each volume. We used this value to label each volume as coherent or incoherent. This information defines which of the two percepts the participant is experiencing at any given time, allowing us to design predictors for coherent and incoherent events (for the GLM analyses) and to have the onsets and offsets of each percept to use for the neurofeedback algorithm.
We estimated the functional activation maps of the unambiguous runs for each TR using a GLM. The design matrix included predictors for all experimental conditions (‘static’, ‘unambiguous motion’, ‘MAE’) and confound regressors based on the mean and first derivative of the voxels in the CSF and the six head motion parameters and their derivatives. Temporal high-pass filtering (cut-off = 0.003 Hz) was applied before extracting ROI activation measures (beta statistic and t-value). Our contrast of interest here was ‘unambiguous motion’ vs. ‘static’.
We then used a different design matrix, replacing the ‘unambiguous motion’ predictor with predictors for coherent and incoherent percepts. As in the previous model, we extract the ROI beta statistic and t-value for the contrasts between coherent/incoherent and static plaid.
For the following analyses, we extract the time course of the ROI and normalize it as a percent signal change to the mean value across time. We estimate the feedback based on this time course as we would in an actual neurofeedback experiment, by displaying the signal variation during the ‘unambiguous motion’ upregulation condition vs. the ‘static’ baseline/downregulation condition.
Next, we use the time courses of the left and right hMT + to estimate Pearson’s correlation over time, a measure of inter-hemispheric functional connectivity. We use a sliding window of 6 seconds for the three lower TRs and of 7.5 seconds for the higher TR (a minimum of 3 data points was considered for obtaining a measurable correlation value). Using the perceptual report information, we center all correlation windows at the transitions between coherent and incoherent percepts and average them across trials and participants. Given the size of the windows, we only considered coherent and incoherent events that lasted at least 7.5 seconds.
To use the functional connectivity information as feedback, we study its characteristics by defining a feedback rule. We defined that successful automatic detection of a perceptual transition meant a 10% minimum increase or decrease in correlation, for coherent and incoherent trials, respectively. We chose this threshold based on (Sousa et al., 2019), which reported an average difference in correlation between percepts of 0.13 ± 0.04. Based on this rule, we extract the transitions’ detection ratio and the time to detection from the first correlation window that includes the volume of transition.
Data and code availability
All the code for the above-mentioned analyses and data (including the activation maps and hMT+ timecourses) can be found at https://github.com/alexsayal/vpmb-tr.