Clinical Trials Logo

Clinical Trial Summary

How does one know what to look at in a scene? Imagine a "Where's Waldo" game - it's challenging to find Waldo because there are many 'salient' locations in the picture, each vying for one's attention. One can only attend to a small location on the picture at a given moment, so to find Waldo, one needs to direct their attention to different locations. One prominent theory about how one accomplishes this claims that important locations are identified based on distinct feature types (for example, motion or color), with locations most unique compared to the background most likely to be attended. An important component of this theory is that individual feature dimensions (again, color or motion) are computed within their own 'feature maps', which are thought to be implemented in specific brain regions. However, whether and how specific brain regions contribute to these feature maps remains unknown. The goal of this study is to determine how brain regions that respond strongly to different feature types (color and motion) and which encode spatial locations of visual stimuli transform 'feature dimension maps' based on stimulus properties as a function of task instructions. The investigators hypothesize that feature-selective brain regions act as neural feature dimension maps, and thus encode representations of relevant location(s) based on their preferred feature dimension, such that the stimulus representation in the most relevant feature map is up-regulated to support adaptive behavior. The investigators will scan healthy human participants using functional MRI (fMRI) in a repeated-measures design while they view visual stimuli made relevant based on a cued feature dimension (e.g., color or motion). The investigators will employ state-of-the-art multivariate analysis techniques that allow them to reconstruct an 'image' of the stimulus representation encoded by each brain region to dissect how neural tissue identifies salient locations. Each participant will perform a challenging discrimination task based on the cued feature (report motion direction or color of stimulus dots) of a stimulus presented in the periphery, which are identical across trial types. Across trials the investigators will manipulate the attended feature value (color, motion, or fixation point). This manipulation will help the investigators fully understand these critical relevance computations in the healthy human visual system.


Clinical Trial Description

In this experiment, participants will engage in all task conditions in a repeated-measures design. Participants are not randomly assigned to groups, as all participants will experience the same set of experimental manipulations. In this experiment, participants will engage in a series of challenging visual attention tasks while their eye position is tracked during fMRI scanning. In all tasks, participants will perform challenging discrimination judgments based on a stimulus presented at the fixation point (discriminate the aspect ratio of a + target - wide or tall?), or a stimulus presented in the periphery (were the dots moving clockwise/counterclockwise? were the dots orange/cyan?). Behavioral responses will be recorded with a button press, which participants will make using a fMRI-compatible button box held in their right hand. In this Experiment, the investigators will manipulate aspects of the behavioral task while keeping the stimulus display constant. These manipulations will allow the investigators to test the role of feature-selective retinotopic regions of interest (ROIs) in transforming spatial representations of salient locations as a function of task relevance to guide visual attention. In this Experiment (Experiment 2.1), the investigators will present a single stimulus at a peripheral location on a blank background containing equiluminant colored moving dots (orange and cyan dots, moving both clockwise and counterclockwise around the stimulus center). Participants will be cued at the beginning of each trial to report either the most prominent color of the dots (orange or cyan), the most prominent motion direction (clockwise or counterclockwise), or to perform a fixation task. If attending to a feature dimension modulates the activation profiles within the corresponding dimension map, the investigators expect to see a selective enhancement of the stimulus representation in the dimension map of a region preferring the attended feature. Participants will also be scanned for an anatomical & retinotopic mapping session, which will allow the investigators to identify brain regions for further analysis using well-established and standardized procedures. STATISTICAL DESIGN & POWER The fMRI studies described in this study record employ an inverted encoding model (IEM) for spatial position to quantify stimulus representations in reconstructed spatial maps of the visual field based on activation patterns measured in retinotopic feature-selective ROIs. The investigators rigorously identify ROIs using independent retinotopic mapping and localizer techniques, and use a 'mapping' task to estimate a 'fixed' encoding model for use across all conditions in each Experiment reported. These design decisions ensure that the investigators can maximize their ability to detect effects of their manipulations of interest within individual participants and brain regions and maximize the statistical power. The investigators use a compromise between deep imaging of several experimental and stimulus conditions within individual participants and aggregation of data across a moderate sample of these deeply-imaged participants (n = 10; see below). This allows the investigators to attain high-quality, reproducible estimates of model-based stimulus representations across task and stimulus manipulations within individual participants and conduct statistical inference on these measurements across the study sample. fMRI analyses will be conducted within each participant's individual brain, and voxels are assigned 'region' labels according to independent criteria (functional retinotopic mapping). Accordingly, there are no comparisons that require precise alignment of brain tissue between participants, and no generation of group-averaged 'maps' of brain activation. As such, concerns about reproducibility of brain maps and associated statistical power concerns are irrelevant to this study design. The statistical design of the study is a repeated-measures design, whereby each participant is exposed to all manipulations in the study. The order of manipulations each participant experiences is randomized across participants. The investigators will employ nonparametric randomization tests for all statistical comparisons whereby they will conduct hypothesis testing (e.g., repeated-measures analysis of variance) using 'shuffled' data (misaligned condition labels relative to measured map activation on each trial) to generate a null distribution of test statistics under the null hypothesis of no effect of their independent variable(s). Once this procedure is repeated extensively (1,000 times) per test, the p-value can be estimated by comparing the test statistic computed using intact labels to this null distribution, and corrected for multiple comparisons as appropriate (e.g., via false discovery rate). Using permutation procedures to generate a null distribution minimizes reliance on parametric assumptions. Additionally, the experiments within the study are designed such that sufficient data will be acquired that data from each individual participant can be used to test the effects of interest. Accordingly, each participant can be considered independent 'replication' of each other participant. Previous studies adopting a similar methodology whereby IEM-based reconstructions of visual stimuli are compared between conditions have employed relatively small sample sizes (n = 7-8). Other studies using population receptive field models or location-specific functional localizer, which are in principle very similar to the approach employed here, have used smaller sample sizes (e.g., n = 6). Sample size & statistical power: In this study, the investigators will acquire an intermediate sample size with extensive data per task condition (n = 10; 2 experimental fMRI sessions, each 1.5-2 hrs, for each participant; along with a 2-hr anatomical imaging and retinotopic mapping fMRI session). Of particular interest, one study used n = 6 participants to establish with a large effect size dz = 3.52 that V1 voxels tuned to a stimulus location where a salient stimulus was defined by feature contrast respond more strongly than when feature contrast is absent. In another study, similar effect sizes were reported by this group in a color-selective ROI known as hV4 (n = 6; dz = 1.06 and 1.80 for orientation- and motion-based contrast, respectively). Accordingly, assuming a conservative effect size of 0.90 (based on those reported previously), the investigators expect a sample size of n = 10 will allow the study to be well-powered (80%, α = 0.05) to detect a similar change in Experiment 1.1, which is most analogous to this study (one-tailed paired T-test). Additionally, the investigators used their pilot data (n = 3) to measure the effect size for the critical comparison between salience-related modulations between feature-selective regions to be dz = 3.10 for the salience-defining feature. These values are commensurate with those cited above, and further support the selection of sample size. If analyses of data acquired during further pilot testing & experiment refinement suggest smaller effect sizes, the investigators will refine the power analyses and adjust the projected enrollment accordingly to ensure robust and reproducible results. Note that this power analysis relies on parametric assumptions which will not be required for the proposed analyses, which invoke randomization methods to compute empirical null distributions. ;


Study Design


NCT number NCT06281457
Study type Interventional
Source University of California, Santa Barbara
Contact
Status Enrolling by invitation
Phase N/A
Start date April 1, 2024
Completion date December 31, 2024