Literature DB >> 25730671

Attention alters orientation processing in the human lateral geniculate nucleus.

Sam Ling1, Michael S Pratte2, Frank Tong2.   

Abstract

Orientation selectivity is a cornerstone property of vision, commonly believed to emerge in the primary visual cortex. We found that reliable orientation information could be detected even earlier, in the human lateral geniculate nucleus, and that attentional feedback selectively altered these orientation responses. This attentional modulation may allow the visual system to modify incoming feature-specific signals at the earliest possible processing site.

Entities:  

Mesh:

Year:  2015        PMID: 25730671      PMCID: PMC4556110          DOI: 10.1038/nn.3967

Source DB:  PubMed          Journal:  Nat Neurosci        ISSN: 1097-6256            Impact factor:   24.884


What role does the human lateral geniculate nucleus (LGN) play in visual perception? Most models of vision treat the LGN as a passive relay station to V1[1,2], characterized by simple circular receptive fields. However, a growing body of neurophysiological evidence suggests that some neurons in the LGN possess receptive fields with elongated aspect ratios[3-6] — a property that could support subcortical orientation-selective processing. Moreover, there is physiological evidence to suggest a plausible mechanism for orientation-specific feedback from V1 to the LGN[7,8]. In particular, the distribution of orientation preferences in cat LGN has been shown to depend on whether V1 is lesioned or spared, with lesions leading to a diminished preference for oblique orientations but a preserved preference for cardinal orientations[5]. This result implies that feedback projections from V1 serve to strengthen the representation of oblique orientations in the LGN. In the present study, we present converging lines of evidence indicating that orientation signals can be detected in the human LGN. Furthermore, we demonstrate that attention selectively enhances representations of oblique orientations, revealing a potential consequence of attentional feedback on these orientation-selective responses. Many fMRI studies have measured orientation-selective signals throughout the visual cortex, using multivariate pattern analysis to classify viewed orientation from voxel activity patterns[9,10]. However, the presence of orientation-selective responses has not, to our knowledge, been demonstrated in the human LGN. fMRI BOLD activity was measured while participants were shown sinusoidal gratings, which were oriented either cardinally (0° or 90°) or obliquely (45° or 135°). The spatial phase of the gratings was randomized across presentations (1 Hz) within each 16-s stimulus block to ensure that pattern classification relied on orientation-selective information, rather than on retinotopic luminance differences[9]. Attention was manipulated by presenting a cue at the start of each block to indicate whether the participant should covertly attend to the oriented grating (attended condition) or to a sequence of letters that appeared in rapid succession at fixation (unattended condition). The grating task involved detecting and discriminating intermittent, near threshold changes in orientation, with magnitudes titrated to match for difficulty across conditions. The letter discrimination task involved reporting whenever a ‘J’ or ‘K’ appeared in the letter sequence. This central letter task is cognitively demanding, and effectively withdraws attention from the visual periphery. Our first goal was to determine whether viewed orientation could be successfully classified from voxel activity patterns in the LGN. We localized the LGN region of interest in individual participants using a combination of functional[11,12] and anatomical localization methods (Fig. 1a & Supplementary Fig. 1a). Our pattern analysis results indicated that reliable orientation information was present not only in V1, but in the human LGN, as well (Fig. 1b; all p’s<0.01). Next, we asked whether the top-down effects of attentional feedback might influence orientation-selective responses in the LGN. To do so, we measured the extent to which attention improved orientation classification performance, separately for oblique and cardinal orientations, based on the hypothesis that cortical feedback might selectively bolster the representation of oblique orientations in the LGN[5]. We observed a significant interaction between the effects of attentional modulation, orientation, and visual area on classification performance, F(4,40) = 5.51, p = 0.001. In V1, attention increased the strength of orientation responses for both oblique and cardinal orientations (Fig. 1b, all p’s < .01) to a comparable extent, t(5) = 0.39, p = 0.71. In the LGN, however, the attentional effect was significantly greater for oblique orientations than for cardinal orientations (Fig. 1b; t(5) = 4.76, p = 0.002). Specifically, the withdrawal of attentional feedback led to substantially weaker orientation-selective activity patterns for oblique orientations in the LGN, but had no reliable effect on cardinal orientations (Fig. 1d). The qualitatively different pattern of results observed in the LGN and V1 suggests that a distinct type of feature-selective modulation is occurring in the LGN, leading to the selective enhancement of oblique orientations (Supplementary Figs. 1b & 2).
Figure 1

Attention selectively augments orientation representation in the human LGN. a) LGN region of interest in a representative subject, identified by the intersection (green) of functional localizers (orange) and proton density-weighted structural imaging (blue). b) Accuracy of orientation decoding (d’ units) for multivariate activity patterns in LGN and V1, for attended and unattended gratings. Attention led to enhanced orientation-selective responses in V1 for all orientation conditions, whereas in the LGN only oblique orientations were modulated by attention. Oblique and cardinal orientation pairs were tested on different scan sessions. Individual points correspond to individual subjects. A non-parametric permutation test confirmed that V1 decoding performance fell outside of the 95% CI bounds of the null distribution in all conditions for every individual participant (24/24 cases), and that LGN decoding performance fell outside of the bounds of the null in 12/12 cases in the attended condition, and 9/12 cases for the unattended condition. c) Attention had comparable effects on mean BOLD activity across orientations, in both the LGN and V1. BOLD response was normalized by the mean intensity across the time series for each run. d) Attentional modulation indices (AMI) for decoding performance of oblique (green) and cardinal (blue) orientations, in LGN and V1. Higher positive values indicate larger effects of attention. Whereas attentional modulation was comparable for cardinal and oblique orientations in area V1, attention selectively modulated responses to oblique orientations in the LGN. Error bars denote ±1 s.e.m.

Do these attentional effects reflect changes in the pattern-specific component of these orientation responses, or might they reflect gross changes in mean BOLD activity? We compared the overall magnitude of the BOLD response for cardinal and oblique orientations. Attention led to enhanced responses in both the LGN and V1 (Fig. 1c, F(4,40) = 20.19, p < .001), but the magnitude of modulation was comparable for cardinal and oblique orientations in the LGN and V1, with no evidence of an interaction effect, F(4,40) = 0.38, p = .56. Taken together, our results indicate that attention enhances the overall BOLD response of the LGN regardless of stimulus orientation, but that the impact of attentional feedback on feature selectivity is more nuanced: the degree to which attention affects LGN processing of a visual stimulus depends on stimulus orientation. The amount of orientation information we observed in LGN activity patterns was modest, when compared to the high classification performance of area V1. This was to be expected due to several factors, including broader orientation selectivity of LGN neurons, poorer quality of fMRI signals from subcortical than cortical regions[10], and the relatively smaller size of the LGN structure. Nevertheless, we observe reliable orientation-selective responses in the human LGN, and these may arise from several sources. For example, animal studies have suggested that ganglion cell receptive fields are not uniformly circular[13], but instead exhibit modest orientation preferences organized at both fine[14] and coarse spatial scales[14]. To determine whether one such coarse-scale orientation bias, known as the radial bias, exists in the human LGN, we conducted an additional experiment to examine whether LGN responses depend on the correspondence between stimulus orientation and retinotopic preference[15]. LGN voxels were localized based on their retinotopic preference for either of the two diagonal radial axes (Fig. 2a), and mean BOLD responses were significantly greater for full-field gratings that were collinear rather than orthogonal to a voxel’s preferred radial axis (Fig. 2b, Supplementary Fig. 3, LGN: t(3) = 6.649, p = .003; V1: t(3) = 7.536; p = .002). These results indicate that the human LGN exhibits a coarse-scale preference for radial orientations, similar to what has been previously found in the human V1. In another experiment, we found that that the orientation of logarithmic spiral gratings could also be decoded from LGN activity patterns, indicating the presence of other sources of orientation preference in human LGN, distinct from radial bias[16] (Supplementary Fig. 4).
Figure 2

Retinotopic preference for radial orientation and effects of orientation-specific masking in the human LGN and V1. a) Schematic of configurations used to test for radial bias. Functional localizers were used to determine voxels with retinotopic preference that fell along either the 45° or 135° axis (depicted by green shaded areas with diagonally opposing quadrants). We then presented a series of full-field gratings, which were oriented 45° or 135°. The relationship between localizer configuration and stimulus orientation determined the Collinear and Orthogonal conditions. b) Radial bias indices for mean BOLD responses in both LGN and V1. Higher positive values indicate larger radial biases. The results reveal that, in both V1 and LGN, the strength of the response depended critically on the match between the orientation of a stimulus, and the retinotopic preference of a region of interest. c) Illustration of stimuli used to test for orientation-tuned masking in LGN and V1. Stimuli were composed of linear sinusoidal gratings summed with orientation bandpass filtered noise. The noise orientation and grating were configured either collinear or orthogonal to each other. d) Orientation masking indices for mean BOLD responses in both LGN and V1 revealed that collinear stimuli had stronger suppression (lower BOLD response) than orthogonal stimuli. Error bars denote ±1 s.e.m.

In a final experiment, we used univariate fMRI analyses to test whether the LGN is sensitive to orientation-tuned masking. By combining target and mask images that have collinear or orthogonal orientations, one can test for nonlinear orientation-selective responses to the linearly summed image pairs[17,18]. We found that BOLD responses in the LGN and V1 exhibit a predicted dependence on orientation: responses were weaker when target and mask were collinear rather than orthogonal (Fig. 2c & d, LGN: t(3) = 7.44, p = .002; V1: t(3) = 4.11; p = .013). These differential responses could not result from LGN detectors that lacked orientation tuning, providing further evidence that orientation responses are present in the human LGN (Supplementary Fig. 6). Recent studies demonstrate that top-down attention can impact the neural responsivity of the LGN[11,19], but we are in the nascent stages of understanding of how it might affect feature- selective processing[20]. Our study provides converging evidence, using both multivariate and univariate fMRI approaches, for orientation-selective responses in the human LGN. These orientation-selective responses exhibit some degree of coarse organization coincident with retinotopic preference, and they appear to exhibit orientation-dependent interactions. Taken together, these results suggest that the LGN may play an underappreciated role in processing orientation information, a process that has traditionally been attributed to the visual cortex. Moreover, we found that attention altered these orientation-selective responses suggesting a remarkably early locus by which top-down processes act upon feature-selective responses. Even modest levels of attentional modulation at this stage could potentially lead to larger effects downstream, suggesting that this early modulation may have an important role in visual perception.

Methods

Observers

Nine healthy adult volunteers (aged 22–41 years, 3 female), with normal or corrected-to-normal vision, participated in the study. All experiments were matched for sex (aside from the spirals study, which had 2 females and 1 male). All subjects gave informed written consent. The study was approved by the Vanderbilt University Institutional Review Board. In each study, participants completed 13–22 fMRI runs (~5 min duration each) per scanning session. A power analysis indicated that 6 subjects would be sufficient to detect the predicted decoding and attention effects. Indeed, this sample size is consistent with previous fMRI decoding studies[21-23].

Apparatus

The stimuli were generated using Matlab and the Psychophysics Toolbox[24,25], and were displayed on a rear-projection screen using a gamma-corrected Eiki LC-X60 LCD projector with a Navitar zoom lens, which participants viewed through a front surface mirror. To minimize head motion, participants used a customized bite bar and padding.

MRI acquisition

Most functional and anatomical data were acquired on a Philips 3T Intera Achieva MRI scanner, equipped with an 8-channel head coil. Each MRI session lasted 2.5 hours, during which we acquired: i) a T1-weighted 3D anatomical scan of the entire brain (1 mm isotropic), ii) a T2-weighted structural in-plane for EPI alignment, iii) 2–4 functional runs to identify retinotopic regions in the LGN and visual cortex that correspond to the stimulus location, and iv) 13–22 fMRI runs to measure BOLD activity during the experimental task. BOLD activity was measured using gradient-echo T2*-weighted echoplanar imaging; twenty slices were acquired axially, with through-plane coverage of the thalamus and the occipital pole, and a voxel size of 2 × 2 × 2 mm (TR 2 s, TE 35 ms; flip angle 79°; FOV 192 × 192 mm). Per subject, the study consisted of 2 scan sessions for the attention study, 1 session for the spirals study, 1 session for the radial bias study, and 1 session for the tuned masking study. Proton density-weighted (PD) images, which were used to anatomically delineate the LGN, were acquired in a separate scanning session on a Philips 7T Achieva MRI scanner, equipped with a 32-channel head coil. Each PD scan lasted 1 hour, during which T1-weighted structural in-plane images was acquired for alignment in addition to 18–20 PD volume acquisitions. PD images were acquired with whole-brain coverage, and a voxel size of 1 × 1 × 1 mm (TR 4.3 s, TE 1.9 ms).

Stimuli and design

Attention experiment

Participants were instructed to maintain fixation on a small circle (diameter: 0.1°) at the center of the display throughout the course of the experiment. Retinotopic regions of interest (ROIs) were identified in the visual cortex and the LGN based on 2 functional localizer scans, in which subjects viewed a flickering checkerboard of the same size and location as the stimuli used in the main experiment (presented in a circular aperture centered on fixation, with an outer diameter of 16°, and an inner diameter of 0.7°). Each localizer run started with 12 s of fixation, followed by 12 s of a flickering checkerboard stimulus (100% contrast; counter-phase reversing at 6 Hz). This off/on stimulus cycle was repeated 8 times in each run, and ended with a final 12 s of fixation. A given run took, in total, 204 seconds. For the main experiment, observers viewed gratings presented at 0°, 45°, 90° or 135°. The cardinal (0°=vertical, 90°=horizontal) and oblique (45°, 135°) orientations were presented in different scan sessions, and the ordering of scans was counterbalanced across subjects (experimenter was not blind to condition), such that 3 subjects were scanned with the oblique orientations first, and the other three were scanned with the cardinal orientations first. To ensure that the classifier relied on orientation-selective information, rather than retinotopic luminance differences, the spatial phase of pairs of gratings was randomized (500 ms on, 500 ms off) within each 16-s stimulus block[9]. To further rule out the possibility that fine-scale retinotopic differences in luminance might contribute to orientation decoding, we attempted to decode stimulus orientation based on the pixel-wise average of counter-phased stimuli, across each block of trials. Decoding performance of these images was at chance, confirming that above-chance decoding of LGN activity patterns must arise from orientation-selective responses[9]. Each run started with 16 s of fixation, followed by 16 s of an oriented grating. This was repeated 16 times in each run, and ended with a final 16 s of fixation. Each run took 288 seconds to complete. Each fMRI run included an equal number of stimulus blocks for each orientation, presented in a randomly generated order. To manipulate attention towards or away from the oriented gratings, a series of letters was presented at central fixation throughout the experiment. A cue appeared at the beginning of each block to indicate whether the participant should attend to the oriented grating (attended condition) or to the sequence of letters at fixation (unattended condition). The grating task involved discriminating near-threshold changes in orientation, and difficulty was titrated to be comparable (~90% accuracy) across conditions and participants. For the letter discrimination task, observers were instructed to report whenever a ‘J’ or ‘K’ appeared in the letter sequence (5 letters/s). The attentional task alternated between blocks within each run.

Spiral grating experiment

Previous fMRI studies of the human visual cortex have shown that orientation-selective signals can be found at multiple spatial scales, ranging from the scale of cortical columns, to a coarse scale of >1cm, such as a retinotopically organized radial bias in which individual voxels exhibit a general preference for orientations that radiate away from the fovea[15,26,27]. Spiral stimuli, however, can mitigate this radial bias[16], although other course-scale biases may be present[28]. Three participants were shown annular logarithmic spiral gratings in a block design (Supplementary Figure 2; 1–9° eccentricity, phase-randomized: 500 ms on, 500 ms off, 10 cycles/rotation), at either ±45° pitch. Each run started with 16 s of fixation, followed by 16 s of an oriented grating. This was repeated 16 times in each run, and ended with a final 16 s of fixation. Each run took 288 seconds to complete. Each fMRI run included an equal number of stimulus blocks for each spiral pitch, presented in a randomly generated order. Classification performance for the pitch of the spirals was evaluated to test for orientation information in LGN and V1 that does not correspond strictly to the radial bias.

Radial bias experiment

To assess the relationship between orientation preference and topography within the LGN and V1, we conducted a separate experiment to explicitly test for radial orientation biases in the LGN. Using a paradigm adopted from Sasaki et al (2006), we examined whether BOLD responses for voxels in the LGN and V1 depended on whether the orientation of a stimulus fell along the axis corresponding to the voxel’s retinotopic polar angle preference, or was orthogonal to the radial axis (Figure 2a). On functional localizer runs, flickering checkerboards were presented in diagonally opposing quadrants (Figure 2a) to identify voxels with retinotopic preference that corresponded better with either the 45° or 135° radial axis. Each localizer run started with 12 s of fixation, followed by 12 s of a wedge-pair stimulus (100% contrast; counter-phase reversing at 6 Hz). This off/on stimulus cycle was repeated 10 times in each run, alternating between the 45° and 135° wedge-pairs across successive cycles, and ended with a final 12 s of fixation. Each localizer run lasted 252 seconds. In the remaining runs, full-field gratings oriented at 0°, 45°, 90° or 135° (100% contrast) were shown to assess whether BOLD responses depended on an interaction between retinotopic preference and orientation. The spatial phase of the grating was phase-reversed (500 ms on, 500 ms off) within each 12-s stimulus block[9]. Each run started with 12 s of fixation. This was followed by repeated cycles consisting of 12 s of an oriented grating, and 12 s of subsequent blank fixation. This off/on stimulus cycle was repeated 10 times in each run, and ended with a final 12 s of fixation. Each run lasted 312 seconds. Each fMRI run included an equal number of stimulus blocks for each orientation, presented in a randomly generated order. To maintain fixation, a series of letters was presented at central fixation throughout the experiment, and observers were instructed to report whenever a ‘J’ or ‘K’ appeared in the letter sequence (5 letters/s). The inclusion of 0° and 90° orientations, while not relevant in this design for the investigation of the radial bias, allowed us to also evaluate the difference in BOLD response between oblique (45° and 135°) and cardinal orientations (0° and 90°) within the same scanning session (Supplementary Figure 4).

Orientation tuned masking experiment

To further assess the evidence for orientation responses in the human LGN, we combined a critical-band noise-masking technique[18,29-31] and univariate fMRI analyses to test for sensitivity to orientation-specific masking. LGN and V1 ROI’s were acquired using a localizer paradigm identical to the attention experiment. Noise masking involves embedding an oriented signal within orientation bandpass-filtered noise. Under these conditions, a population of orientation-selective units will respond more weakly as the orientation content of the signal and noise components becomes more similar, as the stimuli will activate a common set of orientation units rather than two distinct sub-populations. Simulations of neural responses to such stimuli, using both a V1 simple cell Gabor model with half-wave rectification[32,33] and population coding models[34] confirmed that a population of orientation detectors will respond more weakly as the signal and noise orientations become more similar. In contrast, a population of non-oriented center-surround units, based on a difference of Gaussians model[35], yielded a flat pattern of responses as a function of the orientation difference between signal and noise components. Here, we apply this technique to mean BOLD responses in both V1 and LGN, to assess whether the LGN reveals effects of orientation-dependent masking. Participants viewed a sinusoidal grating oriented at either 45° or 135° (40% contrast, 2 cycles/°, counter-phasing at 4 Hz), embedded in orientation bandpass-filtered noise (40% contrast, orientation bandwidth of 5°). The center frequency of this noise band was either identical to the grating orientation (collinear) or rotated 90° from the grating orientation (orthogonal). Stimuli were presented using a block design to measure mean BOLD responses to each of the four possible combinations of grating orientation and noise orientation (45°/45°, 45°/135°, 135°/45°, and 135°/135°). Each run started with 12 s of fixation. This was followed by repeated cycles consisting of 12 s of an oriented grating, and 12 s of subsequent blank fixation. This off/on stimulus cycle was repeated 16 times in each run, and ended with a final 12 s of fixation. Each run lasted 408 s. The order of conditions was pseudo-randomized in a given scan. To encourage fixation, a series of letters was presented at central fixation throughout the experiment, and observers were instructed to report whenever a ‘J’ or ‘K’ appeared in the letter sequence (5 letters/s). Each fMRI run included an equal number of stimulus blocks for each orientation, presented in a random order.

fMRI analyses

The functional data were preprocessed using standard motion-correction procedures[36]. For each run, the raw MRI time series for every voxel was converted to units of percent signal change by dividing each voxel by its mean intensity across that run. To functionally localize ROIs, the localizer time series were fit with a general linear model, which assumed a temporally shifted double-gamma function as a model of the hemodynamic response function. This localizer was combined with standard retinotopic mapping procedures on data obtained in a separate scan session to identify the 150 most responsive voxels within area V1, in the native space for each participant. Contiguous voxels in the thalamus that responded to the localizer stimulus at a threshold of p<0.01 defined the LGN functional region of interest (ROI)[11,37-39]. To anatomically localize the LGN, the PD images were motion-corrected and averaged across scans, and the LGN was identified as the contiguous voxels within the thalamus that appeared as a higher intensity, teardrop-shaped structure medial and superior to the hippocampus, and lateral to the pulvinar[40], often with a gap corresponding to the different density of the white matter sitting between the LGN and the pulvinar. Analyses were conducted using 25 ± 7 (mean ± s.d.) voxels across both left and right LGN. The selection of voxels from this anatomical localizer was conservatively lateralized, to avoid including the adjacent lateral inferior pulvinar. Pattern classification in the LGN was performed on voxels that corresponded to the intersection of the functionally defined LGN ROI and the structurally defined ROI. On average, 61% of proton density-defined LGN overlapped with the functionally defined LGN. Note also that the pattern of results when using only the functional localizer look quite similar to the restricted ROI. Thus, it is unlikely that our conservative voxel selection biased the analysis. (Supplementary Figure 1). In the radial bias experiment, a GLM was fit to the localizer data to generate two ROI’s, distinguishing between voxels that responded to the 45° or 135° wedge-pairs. For the attention experiment, patterns of activity for each stimulus block were obtained by temporally averaging the BOLD activity across each block for every voxel in an ROI, after time shifting by 4 seconds to account for hemodynamic lag. Multivoxel pattern analysis was performed using linear support vector machines (SVM)[9]. Classifiers were trained using a ‘leave-one-run-out’ cross-validation scheme, with performance assessed by testing each left-out run and averaging the resulting classification accuracies over runs. The penalty parameter, C, was tuned within every training set using a cross-validation procedure, which ensured an unbiased selection of C[41]. Classification results presented for the attention experiment represent performance within the cardinal and oblique orientation pairs. For all analyses, classification accuracy was converted to d’[42]. The interaction was tested with repeated-measures ANOVA. A power analysis revealed that 6 subjects would be sufficient to detect the predicted decoding and attention effects. Indeed, this sample size is consistent with previous fMRI decoding studies[21-23]. In addition to the reported t-tests showing that decoding was reliably above chance at the group level, we performed a subject-wise permutation analysis in which chance-level decoding performance was calculated by shuffling the orientation labels 1000 times, per subject, experimental condition, and ROI. In V1, decoding performance fell outside of the bounds of the null distribution’s 95% confidence interval in all instances (24/24). In the LGN, decoding performance fell outside of the bounds of the null in 12/12 instances (i.e., every subject and orientation condition) in the attended condition, and 9/12 instances for unattended. In the attention experiment, the block design used for classification did not have any blank fixation periods between blocks (to maximize the number of samples for classification). Because our measure of % BOLD signal change is normalized to the entire time series, the resulting response estimates for attended and unattended conditions are positive and negative respectively, rather than all being above zero. In the Radial Bias and Orientation-Tuned Masking experiments, BOLD responses were estimated per condition by fitting the time series with a GLM. However, in these experiments there were epochs of blank fixation (12 s) between each stimulus presentation, thereby lead to positive values for % signal change BOLD responses across all conditions.

Eye-position monitoring

Participants were instructed to maintain fixation on a central fixation point throughout each fMRI run. For the experiment involving spiral stimuli, the participant’s eye position was monitored using an MRI-compatible ASL EYE-TRAC eye-tracking system. We applied pattern classifiers to these data to evaluate whether any reliable information about the viewed orientation could be decoded from eye-position signals[22]. Analyses of eye movement data during a subset of scans revealed that our ability to decode orientation from eye position was at chance level, t(3)=0.34, p=0.75. We also conducted additional high-resolution measurements of eye position (500 Hz, SR-Research Eyelink II) outside of the MRI scanner for 5 of the subjects that ran in the study, using a visual stimulation paradigm identical to that used in the main attention experiment. We found that when we tried to decode the orientation of a viewed grating based solely on the high-resolution eye tracking data, performance was at chance level, t(4)=1.50, p>0.2. These findings replicate other published results from our lab, in which we attempted to decode stimulus orientation from eye position[21,22,43]. Taken together, these results indicate that eye movements are unlikely to have contributed to our ability to classify orientation based on fMRI activity patterns.
  42 in total

1.  Attention modulates responses in the human lateral geniculate nucleus.

Authors:  Daniel H O'Connor; Miki M Fukui; Mark A Pinsk; Sabine Kastner
Journal:  Nat Neurosci       Date:  2002-11       Impact factor: 24.884

2.  Are primate lateral geniculate nucleus (LGN) cells really sensitive to orientation or direction?

Authors:  Xiangmin Xu; Jennifer Ichida; Yuri Shostak; A B Bonds; Vivien A Casagrande
Journal:  Vis Neurosci       Date:  2002 Jan-Feb       Impact factor: 3.241

3.  Robust object recognition with cortex-like mechanisms.

Authors:  Thomas Serre; Lior Wolf; Stanley Bileschi; Maximilian Riesenhuber; Tomaso Poggio
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2007-03       Impact factor: 6.226

4.  Auditory filter shapes derived with noise stimuli.

Authors:  R D Patterson
Journal:  J Acoust Soc Am       Date:  1976-03       Impact factor: 1.840

5.  Comparison of pattern recognition methods in classifying high-resolution BOLD signals obtained at high magnetic field in monkeys.

Authors:  Shih-pi Ku; Arthur Gretton; Jakob Macke; Nikos K Logothetis
Journal:  Magn Reson Imaging       Date:  2008-09       Impact factor: 2.546

6.  The Psychophysics Toolbox.

Authors:  D H Brainard
Journal:  Spat Vis       Date:  1997

7.  How attention extracts objects from noise.

Authors:  Michael S Pratte; Sam Ling; Jascha D Swisher; Frank Tong
Journal:  J Neurophysiol       Date:  2013-06-26       Impact factor: 2.714

8.  Spatial-frequency masking in vision: critical bands and spread of masking.

Authors:  C F Stromeyer; B Julesz
Journal:  J Opt Soc Am       Date:  1972-10

9.  Guarding the gateway to cortex with attention in visual thalamus.

Authors:  Kerry McAlonan; James Cavanaugh; Robert H Wurtz
Journal:  Nature       Date:  2008-10-05       Impact factor: 49.962

10.  fMRI orientation decoding in V1 does not require global maps or globally coherent orientation stimuli.

Authors:  Arjen Alink; Alexandra Krugliak; Alexander Walther; Nikolaus Kriegeskorte
Journal:  Front Psychol       Date:  2013-08-12
View more
  26 in total

1.  The influence of low-level stimulus features on the representation of contexts, items, and their mnemonic associations.

Authors:  Derek J Huffman; Craig E L Stark
Journal:  Neuroimage       Date:  2017-04-08       Impact factor: 6.556

2.  Primary Tactile Thalamus Spiking Reflects Cognitive Signals.

Authors:  Christian Waiblinger; Clarissa J Whitmire; Audrey Sederberg; Garrett B Stanley; Cornelius Schwarz
Journal:  J Neurosci       Date:  2018-04-27       Impact factor: 6.167

3.  Figure-Ground Modulation in the Human Lateral Geniculate Nucleus Is Distinguishable from Top-Down Attention.

Authors:  Sonia Poltoratski; Alexander Maier; Allen T Newton; Frank Tong
Journal:  Curr Biol       Date:  2019-06-06       Impact factor: 10.834

4.  Oculomotor inhibition covaries with conscious detection.

Authors:  Alex L White; Martin Rolfs
Journal:  J Neurophysiol       Date:  2016-07-06       Impact factor: 2.714

5.  Characterizing the effects of feature salience and top-down attention in the early visual system.

Authors:  Sonia Poltoratski; Sam Ling; Devin McCormack; Frank Tong
Journal:  J Neurophysiol       Date:  2017-04-05       Impact factor: 2.714

6.  Decoding the different states of visual attention using functional and effective connectivity features in fMRI data.

Authors:  Behdad Parhizi; Mohammad Reza Daliri; Mehdi Behroozi
Journal:  Cogn Neurodyn       Date:  2017-11-25       Impact factor: 5.082

7.  Robust functional mapping of layer-selective responses in human lateral geniculate nucleus with high-resolution 7T fMRI.

Authors:  Yazhu Qian; Jinyou Zou; Zihao Zhang; Jing An; Zhentao Zuo; Yan Zhuo; Danny J J Wang; Peng Zhang
Journal:  Proc Biol Sci       Date:  2020-04-15       Impact factor: 5.349

8.  A cross-species comparison of corticogeniculate structure and function.

Authors:  J Michael Hasse; Farran Briggs
Journal:  Vis Neurosci       Date:  2017-11-16       Impact factor: 3.241

9.  Radial bias is not necessary for orientation decoding.

Authors:  Michael S Pratte; Jocelyn L Sy; Jascha D Swisher; Frank Tong
Journal:  Neuroimage       Date:  2015-12-05       Impact factor: 6.556

10.  Modulation of the Primary Auditory Thalamus When Recognizing Speech with Background Noise.

Authors:  Paul Glad Mihai; Nadja Tschentscher; Katharina von Kriegstein
Journal:  J Neurosci       Date:  2021-07-09       Impact factor: 6.167

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.