Academic Appointments
-
Associate Professor, Psychology
-
Member, Bio-X
-
Faculty Affiliate, Institute for Human-Centered Artificial Intelligence (HAI)
-
Member, Wu Tsai Neurosciences Institute
Administrative Appointments
-
co-Director, Neurosciences Interdisciplinary Graduate Program (2021 - Present)
Program Affiliations
-
Symbolic Systems Program
Professional Education
-
PhD, University of California, Berkeley and UCSF, Bioengineering (2002)
-
BS, Yale University, Computer Science (1993)
Current Research and Scholarly Interests
How does neural activity in the human cortex create our sense of visual perception? We use a combination of functional magnetic resonance imaging, computational modeling and analysis, and psychophysical measurements to link human perception to cortical brain activity.
2024-25 Courses
- Brain decoding
PSYCH 164 (Win) - Cognitive Neuroscience: Vision
PSYCH 263 (Win) - Introduction to Cognitive Neuroscience
PSYCH 50 (Aut) - Neurosciences Cognitive Core
NEPR 207 (Spr) - Practicum in Teaching: Intro to Cognitive Neuroscience
PSYCH 50A (Aut) -
Independent Studies (8)
- Directed Reading in Neurosciences
NEPR 299 (Aut, Win, Spr, Sum) - Graduate Research
NEPR 399 (Aut, Win, Spr, Sum) - Graduate Research
PSYCH 275 (Aut, Win, Spr) - Independent Study
SYMSYS 296 (Aut, Win, Spr) - Practicum in Teaching
PSYCH 281 (Aut, Win, Spr) - Reading and Special Work
PSYCH 194 (Aut, Win, Spr) - Senior Honors Tutorial
SYMSYS 190 (Aut, Win, Spr) - Special Laboratory Projects
PSYCH 195 (Aut, Win, Spr)
- Directed Reading in Neurosciences
-
Prior Year Courses
2023-24 Courses
- Brain decoding
PSYCH 164 (Spr) - Cognitive Neuroscience: Vision
PSYCH 263 (Spr) - Introduction to Cognitive Neuroscience
PSYCH 50 (Win) - Neuroscience research
PSYCH 196A (Spr) - Neurosciences Cognitive Core
NEPR 207 (Spr) - Practicum in Teaching: Intro to Cognitive Neuroscience
PSYCH 50A (Win)
2022-23 Courses
- Cognitive Neuroscience: Vision
PSYCH 263 (Aut) - Neuroscience research
PSYCH 196A (Spr) - Neurosciences Cognitive Core
NEPR 207 (Spr)
2021-22 Courses
- Brain decoding
PSYCH 164 (Win) - Foundational Topics in Neuroscience
PSYCH 196B (Sum) - Introduction to Cognitive Neuroscience
PSYCH 50 (Win) - Neuroscience research
PSYCH 196A (Spr) - Neurosciences Cognitive Core
NEPR 207 (Spr) - Practicum in Teaching: Intro to Cognitive Neuroscience
PSYCH 50A (Win)
- Brain decoding
Stanford Advisees
-
Doctoral Dissertation Reader (AC)
Minseung Choi, Insub Kim -
Postdoctoral Faculty Sponsor
Jiwon Yeon -
Doctoral Dissertation Advisor (AC)
Austin Kuo, Joshua Ryu -
Doctoral Dissertation Co-Advisor (AC)
Alex Durango -
Doctoral (Program)
Hyunwoo Gu, Josh Wilson -
Postdoctoral Research Mentor
Jiwon Yeon
All Publications
-
Gain, not concomitant changes in spatial receptive field properties, improves task performance in a neural network attention model.
eLife
2023; 12
Abstract
Attention allows us to focus sensory processing on behaviorally relevant aspects of the visual world. One potential mechanism of attention is a change in the gain of sensory responses. However, changing gain at early stages could have multiple downstream consequences for visual processing. Which, if any, of these effects can account for the benefits of attention for detection and discrimination? Using a model of primate visual cortex we document how a Gaussian-shaped gain modulation results in changes to spatial tuning properties. Forcing the model to use only these changes failed to produce any benefit in task performance. Instead, we found that gain alone was both necessary and sufficient to explain category detection and discrimination during attention. Our results show how gain can give rise to changes in receptive fields which are not necessary for enhancing task performance.
View details for DOI 10.7554/eLife.78392
View details for PubMedID 37184221
-
Texture-like representation of objects in human visual cortex.
Proceedings of the National Academy of Sciences of the United States of America
2022; 119 (17): e2115302119
Abstract
SignificanceHumans are exquisitely sensitive to the spatial arrangement of visual features in objects and scenes, but not in visual textures. Category-selective regions in the visual cortex are widely believed to underlie object perception, suggesting such regions should distinguish natural images of objects from synthesized images containing similar visual features in scrambled arrangements. Contrarily, we demonstrate that representations in category-selective cortex do not discriminate natural images from feature-matched scrambles but can discriminate images of different categories, suggesting a texture-like encoding. We find similar insensitivity to feature arrangement in Imagenet-trained deep convolutional neural networks. This suggests the need to reconceptualize the role of category-selective cortex as representing a basis set of complex texture-like features, useful for a myriad of behaviors.
View details for DOI 10.1073/pnas.2115302119
View details for PubMedID 35439063
-
Population Models, Not Analyses, of Human Neuroscience Measurements.
Annual review of vision science
2021
Abstract
Selectivity for many basic properties of visual stimuli, such as orientation, is thought to be organized at the scale of cortical columns, making it difficult or impossible to measure directly with noninvasive human neuroscience measurement. However, computational analyses of neuroimaging data have shown that selectivity for orientation can be recovered by considering the pattern of response across a region of cortex. This suggests that computational analyses can reveal representation encoded at a finer spatial scale than is implied by the spatial resolution limits of measurement techniques. This potentially opens up the possibility to study a much wider range of neural phenomena that are otherwise inaccessible through noninvasive measurement. However, as we review in this article, a large body of evidence suggests an alternative hypothesis to this superresolution account: that orientation information is available at the spatial scale of cortical maps and thus easily measurable at the spatial resolution of standard techniques. In fact, a population model shows that this orientation information need not even come from single-unit selectivity for orientation tuning, but instead can result from population selectivity for spatial frequency. Thus, a categorical error of interpretation can result whereby orientation selectivity can be confused with spatial frequency selectivity. This is similarly problematic for the interpretation of results from numerous studies of more complex representations and cognitive functions that have built upon the computational techniques used to reveal stimulus orientation. We suggest in this review that these interpretational ambiguities can be avoided by treating computational analyses as models of the neural processes that give rise to measurement. Building upon the modeling tradition in vision science using considerations of whether population models meet a set of core criteria is important for creating the foundation for a cumulative and replicable approach to making valid inferences from human neuroscience measurements. Expected final online publication date for the Annual Review of Vision Science, Volume 7 is September 2021. Please see http://www.annualreviews.org/page/journal/pubdates for revised estimates.
View details for DOI 10.1146/annurev-vision-093019-111124
View details for PubMedID 34283926
-
Optimality and heuristics in perceptual neuroscience
NATURE NEUROSCIENCE
2019; 22 (4): 514–23
View details for DOI 10.1038/s41593-019-0340-4
View details for Web of Science ID 000462154300005
-
A flexible readout mechanism of human sensory representations.
Nature communications
2019; 10 (1): 3500
Abstract
Attention can both enhance and suppress cortical sensory representations. However, changing sensory representations can also be detrimental to behavior. Behavioral consequences can be avoided by flexibly changing sensory readout, while leaving the representations unchanged. Here, we asked human observers to attend to and report about either one of two features which control the visibility of motion while making concurrent measurements of cortical activity with BOLD imaging (fMRI). We extend a well-established linking model to account for the relationship between these measurements and find that changes in sensory representation during directed attention are insufficient to explain perceptual reports. Adding a flexible downstream readout is necessary to best explain our data. Such a model implies that observers should be able to recover information about ignored features, a prediction which we confirm behaviorally. Thus, flexible readout is a critical component of the cortical implementation of human adaptive behavior.
View details for DOI 10.1038/s41467-019-11448-7
View details for PubMedID 31375665
-
A Switching Observer for Human Perceptual Estimation
Neuron
2018; 97 (2): 462-474
View details for DOI 10.1016/j.neuron.2017.12.011
-
Plaudits for logits in sensory neuroscience.
Neuron
2024; 112 (17): 2825-2827
Abstract
A workhorse tool of economic decision-making has long sought to get inside people's heads through careful examination of their choices. In this issue of Neuron, Carandini1 flips the script, showing how it can model how the brain makes sensory choices.
View details for DOI 10.1016/j.neuron.2024.08.008
View details for PubMedID 39236675
-
Efficient coding of natural images in the mouse visual cortex.
Nature communications
2024; 15 (1): 2466
Abstract
How the activity of neurons gives rise to natural vision remains a matter of intense investigation. The mid-level visual areas along the ventral stream are selective to a common class of natural images-textures-but a circuit-level understanding of this selectivity and its link to perception remains unclear. We addressed these questions in mice, first showing that they can perceptually discriminate between textures and statistically simpler spectrally matched stimuli, and between texture types. Then, at the neural level, we found that the secondary visual area (LM) exhibited a higher degree of selectivity for textures compared to the primary visual area (V1). Furthermore, textures were represented in distinct neural activity subspaces whose relative distances were found to correlate with the statistical similarity of the images and the mice's ability to discriminate between them. Notably, these dependencies were more pronounced in LM, where the texture-related subspaces were smaller than in V1, resulting in superior stimulus decoding capabilities. Together, our results demonstrate texture vision in mice, finding a linking framework between stimulus statistics, neural representations, and perceptual sensitivity-a distinct hallmark of efficient coding computations.
View details for DOI 10.1038/s41467-024-45919-3
View details for PubMedID 38503746
View details for PubMedCentralID 5629359
-
Mouse visual cortex as a limited resource system that self-learns an ecologically-general representation.
PLoS computational biology
2023; 19 (10): e1011506
Abstract
Studies of the mouse visual system have revealed a variety of visual brain areas that are thought to support a multitude of behavioral capacities, ranging from stimulus-reward associations, to goal-directed navigation, and object-centric discriminations. However, an overall understanding of the mouse's visual cortex, and how it supports a range of behaviors, remains unknown. Here, we take a computational approach to help address these questions, providing a high-fidelity quantitative model of mouse visual cortex and identifying key structural and functional principles underlying that model's success. Structurally, we find that a comparatively shallow network structure with a low-resolution input is optimal for modeling mouse visual cortex. Our main finding is functional-that models trained with task-agnostic, self-supervised objective functions based on the concept of contrastive embeddings are much better matches to mouse cortex, than models trained on supervised objectives or alternative self-supervised methods. This result is very much unlike in primates where prior work showed that the two were roughly equivalent, naturally leading us to ask the question of why these self-supervised objectives are better matches than supervised ones in mouse. To this end, we show that the self-supervised, contrastive objective builds a general-purpose visual representation that enables the system to achieve better transfer on out-of-distribution visual scene understanding and reward-based navigation tasks. Our results suggest that mouse visual cortex is a low-resolution, shallow network that makes best use of the mouse's limited resources to create a light-weight, general-purpose visual system-in contrast to the deep, high-resolution, and more categorization-dominated visual system of primates.
View details for DOI 10.1371/journal.pcbi.1011506
View details for PubMedID 37782673
-
What has vision science taught us about functional MRI?
NeuroImage
2022: 119536
Abstract
In the domain of human neuroimaging, much attention has been paid to the question of whether and how the development of functional magnetic resonance imaging (fMRI) has advanced our scientific knowledge of the human brain. However, the opposite question is also important; how has our knowledge of the visual system advanced our understanding of fMRI? Here, we discuss how and why scientific knowledge about the human and animal visual system has been used to answer fundamental questions about fMRI as a brain measurement tool and how these answers have contributed to scientific discoveries beyond vision science.
View details for DOI 10.1016/j.neuroimage.2022.119536
View details for PubMedID 35931310
-
Increasing neural network robustness improves match to macaque V1 eigenspectrum, spatial frequency preference and predictivity.
PLoS computational biology
2022; 18 (1): e1009739
Abstract
Task-optimized convolutional neural networks (CNNs) show striking similarities to the ventral visual stream. However, human-imperceptible image perturbations can cause a CNN to make incorrect predictions. Here we provide insight into this brittleness by investigating the representations of models that are either robust or not robust to image perturbations. Theory suggests that the robustness of a system to these perturbations could be related to the power law exponent of the eigenspectrum of its set of neural responses, where power law exponents closer to and larger than one would indicate a system that is less susceptible to input perturbations. We show that neural responses in mouse and macaque primary visual cortex (V1) obey the predictions of this theory, where their eigenspectra have power law exponents of at least one. We also find that the eigenspectra of model representations decay slowly relative to those observed in neurophysiology and that robust models have eigenspectra that decay slightly faster and have higher power law exponents than those of non-robust models. The slow decay of the eigenspectra suggests that substantial variance in the model responses is related to the encoding of fine stimulus features. We therefore investigated the spatial frequency tuning of artificial neurons and found that a large proportion of them preferred high spatial frequencies and that robust models had preferred spatial frequency distributions more aligned with the measured spatial frequency distribution of macaque V1 cells. Furthermore, robust models were quantitatively better models of V1 than non-robust models. Our results are consistent with other findings that there is a misalignment between human and machine perception. They also suggest that it may be useful to penalize slow-decaying eigenspectra or to bias models to extract features of lower spatial frequencies during task-optimization in order to improve robustness and V1 neural response predictivity.
View details for DOI 10.1371/journal.pcbi.1009739
View details for PubMedID 34995280
-
Voxel-Wise Linearity Analysis of Increments and Decrements in BOLD Responses in Human Visual Cortex Using a Contrast Adaptation Paradigm.
Frontiers in human neuroscience
2021; 15: 541314
Abstract
The linearity of BOLD responses is a fundamental presumption in most analysis procedures for BOLD fMRI studies. Previous studies have examined the linearity of BOLD signal increments, but less is known about the linearity of BOLD signal decrements. The present study assessed the linearity of both BOLD signal increments and decrements in the human primary visual cortex using a contrast adaptation paradigm. Results showed that both BOLD signal increments and decrements kept linearity to long stimuli (e.g., 3 s, 6 s), yet, deviated from linearity to transient stimuli (e.g., 1 s). Furthermore, a voxel-wise analysis showed that the deviation patterns were different for BOLD signal increments and decrements: while the BOLD signal increments demonstrated a consistent overestimation pattern, the patterns for BOLD signal decrements varied from overestimation to underestimation. Our results suggested that corrections to deviations from linearity of transient responses should consider the different effects of BOLD signal increments and decrements.
View details for DOI 10.3389/fnhum.2021.541314
View details for PubMedID 34531731
-
Context effects on probability estimation
PLOS BIOLOGY
2020; 18 (3)
View details for DOI 10.1371/journal.pbio.3000634.r006
View details for Web of Science ID 000522940500005
-
Humans perceive binocular rivalry and fusion in a tristable dynamic state.
The Journal of neuroscience : the official journal of the Society for Neuroscience
2019
Abstract
Human vision combines inputs from the two eyes into one percept. Small differences 'fuse' together, while larger differences are seen 'rivalrously' from one eye at a time. These outcomes are typically treated as mutually exclusive processes, with paradigms targeting one or the other and fusion being unreported in most rivalry studies. Is fusion truly a default, stable state that only breaks into rivalry for non-fusible stimuli? Or are monocular and fused percepts three sub-states of one dynamical system? To determine whether fusion and rivalry are separate processes, we measured human perception of Gabor patches with a range of inter-ocular orientation disparities. Observers (10 female, 5 male) reported rivalrous, fused and uncertain percepts over time. We found a dynamic "tristable" zone spanning from 25-35 degrees of orientation disparity where fused, left- or right-eye dominant percepts could all occur. The temporal characteristics of fusion and non-fusion periods during tristability matched other bistable processes. We tested statistical models with fusion as a higher-level bistable process alternating with rivalry against our findings. None of these fit our data, but a simple bistable model extended to have three states reproduced many of our observations. We conclude that rivalry and fusion are multistable sub-states capable of direct competition, rather than separate bistable processes.SIGNIFICANCE STATEMENTWhen inputs to the two eyes differ, they can either fuse together or engage in binocular rivalry, where each eye's view is seen exclusively in turn. Visual stimuli have often been tailored to produce either fusion or rivalry, implicitly treating them as separate mutually-exclusive perceptual processes. We have found that some similar-but-different stimuli can result in both outcomes over time. Comparing various simple models with our results suggests that rivalry and fusion are not independent processes, but compete within a single multistable system. This conceptual shift is a step toward unifying fusion and rivalry, and understanding how they both contribute to the visual system's production of a unified interpretation of the conflicting images cast on the retina by real-world scenes.
View details for DOI 10.1523/JNEUROSCI.0713-19.2019
View details for PubMedID 31519817
-
Computing Social Value Conversion in the Human Brain.
The Journal of neuroscience : the official journal of the Society for Neuroscience
2019
Abstract
Social signals play powerful roles in shaping self-oriented reward valuation and decision-making. These signals activate social and valuation/decision areas, but the core computation for their integration into the self-oriented decision machinery remains unclear. Here, we study how a fundamental social signal-social value (others' reward value)-is converted into self-oriented decision-making in the human brain. Using behavioral analysis, modeling, and neuroimaging, we show three-stage processing of social value conversion from the offer to the effective value and then to the final decision value. First, a value of others' bonus on offer, called offered value, was encoded uniquely in the right temporoparietal junction (rTPJ), and also in the left dorsolateral prefrontal cortex (ldlPFC), which is commonly activated by offered self-bonus value. The effective value, an intermediate value representing the effective influence of the offer on the decision, was represented in the right anterior insula (rAI), and the final decision value was encoded in the medial prefrontal cortex (mPFC). Second, using psychophysiological interaction (PPI) and dynamic causal modeling (DCM) analyses, we demonstrated three-stage feedforward processing; from the rTPJ and ldPFC to the rAI and then from rAI to the mPFC. Further, we showed that these characteristics of social conversion underlie distinct sociobehavioral phenotypes. We demonstrate that the variability in the conversion underlies difference between prosocial and selfish subjects, as seen from the differential strength of the rAI and ldlPFC coupling to the mPFC responses, respectively. Together, these findings identified fundamental neural computation processes for social value conversion underlying complex social decision-making behaviors.SIGNIFICANCE STATEMENTIn daily life, we make decisions based on self-interest but also in consideration for others' status. These social influences modulate valuation and decision signals in the brain, suggesting a fundamental process called value conversion that translates social information into self-referenced decisions. However, little is known about the conversion process and its underlying brain mechanisms. We investigated value conversion using human fMRI with computational modeling and found three essential stages in a progressive brain circuit from social to empathic and decision areas. Interestingly, the brain mechanism of conversion differed between prosocial and individualistic subjects. These findings reveal how the brain processes and merges social information into the elemental flow of self-interested decision-making.
View details for PubMedID 31000587
-
Inverted Encoding Models Reconstruct an Arbitrary Model Response, Not the Stimulus.
eNeuro
2019; 6 (2)
Abstract
Probing how large populations of neurons represent stimuli is key to understanding sensory representations as many stimulus characteristics can only be discerned from population activity and not from individual single-units. Recently, inverted encoding models have been used to produce channel response functions from large spatial-scale measurements of human brain activity that are reminiscent of single-unit tuning functions and have been proposed to assay "population-level stimulus representations" (Sprague et al., 2018a). However, these channel response functions do not assay population tuning. We show by derivation that the channel response function is only determined up to an invertible linear transform. Thus, these channel response functions are arbitrary, one of an infinite family and therefore not a unique description of population representation. Indeed, simulations demonstrate that bimodal, even random, channel basis functions can account perfectly well for population responses without any underlying neural response units that are so tuned. However, the approach can be salvaged by extending it to reconstruct the stimulus, not the assumed model. We show that when this is done, even using bimodal and random channel basis functions, a unimodal function peaking at the appropriate value of the stimulus is recovered which can be interpreted as a measure of population selectivity. More precisely, the recovered function signifies how likely any value of the stimulus is, given the observed population response. Whether an analysis is recovering the hypothetical responses of an arbitrary model rather than assessing the selectivity of population representations is not an issue unique to the inverted encoding model and human neuroscience, but a general problem that must be confronted as more complex analyses intervene between measurement of population activity and presentation of data.
View details for PubMedID 30923743
-
A quantitative framework for motion visibility in human cortex
JOURNAL OF NEUROPHYSIOLOGY
2018; 120 (4): 1824-1839
View details for DOI 10.1152/jn.00433.2018
View details for Web of Science ID 000451350100033
-
Task-dependent enhancement of facial expression and identity representations in human cortex
NEUROIMAGE
2018; 172: 689–702
Abstract
What cortical mechanisms allow humans to easily discern the expression or identity of a face? Subjects detected changes in expression or identity of a stream of dynamic faces while we measured BOLD responses from topographically and functionally defined areas throughout the visual hierarchy. Responses in dorsal areas increased during the expression task, whereas responses in ventral areas increased during the identity task, consistent with previous studies. Similar to ventral areas, early visual areas showed increased activity during the identity task. If visual responses are weighted by perceptual mechanisms according to their magnitude, these increased responses would lead to improved attentional selection of the task-appropriate facial aspect. Alternatively, increased responses could be a signature of a sensitivity enhancement mechanism that improves representations of the attended facial aspect. Consistent with the latter sensitivity enhancement mechanism, attending to expression led to enhanced decoding of exemplars of expression both in early visual and dorsal areas relative to attending identity. Similarly, decoding identity exemplars when attending to identity was improved in dorsal and ventral areas. We conclude that attending to expression or identity of dynamic faces is associated with increased selectivity in representations consistent with sensitivity enhancement.
View details for PubMedID 29432802
-
Inverted Encoding Models of Human Population Response Conflate Noise and Neural Tuning Width.
The Journal of neuroscience : the official journal of the Society for Neuroscience
2018; 38 (2): 398–408
Abstract
Channel-encoding models offer the ability to bridge different scales of neuronal measurement by interpreting population responses, typically measured with BOLD imaging in humans, as linear sums of groups of neurons (channels) tuned for visual stimulus properties. Inverting these models to form predicted channel responses from population measurements in humans seemingly offers the potential to infer neuronal tuning properties. Here, we test the ability to make inferences about neural tuning width from inverted encoding models. We examined contrast invariance of orientation selectivity in human V1 (both sexes) and found that inverting the encoding model resulted in channel response functions that became broader with lower contrast, thus apparently violating contrast invariance. Simulations showed that this broadening could be explained by contrast-invariant single-unit tuning with the measured decrease in response amplitude at lower contrast. The decrease in response lowers the signal-to-noise ratio of population responses that results in poorer population representation of orientation. Simulations further showed that increasing signal to noise makes channel response functions less sensitive to underlying neural tuning width, and in the limit of zero noise will reconstruct the channel function assumed by the model regardless of the bandwidth of single units. We conclude that our data are consistent with contrast-invariant orientation tuning in human V1. More generally, our results demonstrate that population selectivity measures obtained by encoding models can deviate substantially from the behavior of single units because they conflate neural tuning width and noise and are therefore better used to estimate the uncertainty of decoded stimulus properties.SIGNIFICANCE STATEMENT It is widely recognized that perceptual experience arises from large populations of neurons, rather than a few single units. Yet, much theory and experiment have examined links between single units and perception. Encoding models offer a way to bridge this gap by explicitly interpreting population activity as the aggregate response of many single neurons with known tuning properties. Here we use this approach to examine contrast-invariant orientation tuning of human V1. We show with experiment and modeling that due to lower signal to noise, contrast-invariant orientation tuning of single units manifests in population response functions that broaden at lower contrast, rather than remain contrast-invariant. These results highlight the need for explicit quantitative modeling when making a reverse inference from population response profiles to single-unit responses.
View details for PubMedID 29167406
-
Adaptable history biases in human perceptual decisions
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA
2016; 113 (25): E3548-E3557
Abstract
When making choices under conditions of perceptual uncertainty, past experience can play a vital role. However, it can also lead to biases that worsen decisions. Consistent with previous observations, we found that human choices are influenced by the success or failure of past choices even in a standard two-alternative detection task, where choice history is irrelevant. The typical bias was one that made the subject switch choices after a failure. These choice history biases led to poorer performance and were similar for observers in different countries. They were well captured by a simple logistic regression model that had been previously applied to describe psychophysical performance in mice. Such irrational biases seem at odds with the principles of reinforcement learning, which would predict exquisite adaptability to choice history. We therefore asked whether subjects could adapt their irrational biases following changes in trial order statistics. Adaptability was strong in the direction that confirmed a subject's default biases, but weaker in the opposite direction, so that existing biases could not be eradicated. We conclude that humans can adapt choice history biases, but cannot easily overcome existing biases even if irrational in the current context: adaptation is more sensitive to confirmatory than contradictory statistics.
View details for DOI 10.1073/pnas.1518786113
View details for Web of Science ID 000378272400014
View details for PubMedID 27330086
View details for PubMedCentralID PMC4922170
-
Parietal and prefrontal: categorical differences?
Nature neuroscience
2015; 19 (1): 5-7
View details for DOI 10.1038/nn.4204
View details for PubMedID 26713741
-
A CASE FOR HUMAN SYSTEMS NEUROSCIENCE
NEUROSCIENCE
2015; 296: 130-137
Abstract
Can the human brain itself serve as a model for a systems neuroscience approach to understanding the human brain? After all, how the brain is able to create the richness and complexity of human behavior is still largely mysterious. What better choice to study that complexity than to study it in humans? However, measurements of brain activity typically need to be made non-invasively which puts severe constraints on what can be learned about the internal workings of the brain. Our approach has been to use a combination of psychophysics in which we can use human behavioral flexibility to make quantitative measurements of behavior and link those through computational models to measurements of cortical activity through magnetic resonance imaging. In particular, we have tested various computational hypotheses about what neural mechanisms could account for behavioral enhancement with spatial attention (Pestilli et al., 2011). Resting both on quantitative measurements and considerations of what is known through animal models, we concluded that weighting of sensory signals by the magnitude of their response is a neural mechanism for efficient selection of sensory signals and consequent improvements in behavioral performance with attention. While animal models have many technical advantages over studying the brain in humans, we believe that human systems neuroscience should endeavor to validate, replicate and extend basic knowledge learned from animal model systems and thus form a bridge to understanding how the brain creates the complex and rich cognitive capacities of humans.
View details for DOI 10.1016/j.neuroscience.2014.06.052
View details for Web of Science ID 000353828300015
View details for PubMedID 24997268
-
Encoding of graded changes in spatial specificity of prior cues in human visual cortex
JOURNAL OF NEUROPHYSIOLOGY
2014; 112 (11): 2834-2849
Abstract
Prior information about the relevance of spatial locations can vary in specificity; a single location, a subset of locations, or all locations may be of potential importance. Using a contrast-discrimination task with four possible targets, we asked whether performance benefits are graded with the spatial specificity of a prior cue and whether we could quantitatively account for behavioral performance with cortical activity changes measured by blood oxygenation level-dependent (BOLD) imaging. Thus we changed the prior probability that each location contained the target from 100 to 50 to 25% by cueing in advance 1, 2, or 4 of the possible locations. We found that behavioral performance (discrimination thresholds) improved in a graded fashion with spatial specificity. However, concurrently measured cortical responses from retinotopically defined visual areas were not strictly graded; response magnitude decreased when all 4 locations were cued (25% prior probability) relative to the 100 and 50% prior probability conditions, but no significant difference in response magnitude was found between the 100 and 50% prior probability conditions for either cued or uncued locations. Also, although cueing locations increased responses relative to noncueing, this cue sensitivity was not graded with prior probability. Furthermore, contrast sensitivity of cortical responses, which could improve contrast discrimination performance, was not graded. Instead, an efficient-selection model showed that even if sensory responses do not strictly scale with prior probability, selection of sensory responses by weighting larger responses more can result in graded behavioral performance benefits with increasing spatial specificity of prior information.
View details for DOI 10.1152/jn.00729.2013
View details for Web of Science ID 000346023000015
View details for PubMedID 25185808
-
Differing effects of attention in single-units and populations are well predicted by heterogeneous tuning and the normalization model of attention
FRONTIERS IN COMPUTATIONAL NEUROSCIENCE
2014; 8
Abstract
Single-unit measurements have reported many different effects of attention on contrast-response (e.g., contrast-gain, response-gain, additive-offset dependent on visibility), while functional imaging measurements have more uniformly reported increases in response across all contrasts (additive-offset). The normalization model of attention elegantly predicts the diversity of effects of attention reported in single-units well-tuned to the stimulus, but what predictions does it make for more realistic populations of neurons with heterogeneous tuning? Are predictions in accordance with population-scale measurements? We used functional imaging data from humans to determine a realistic ratio of attention-field to stimulus-drive size (a key parameter for the model) and predicted effects of attention in a population of model neurons with heterogeneous tuning. We found that within the population, neurons well-tuned to the stimulus showed a response-gain effect, while less-well-tuned neurons showed a contrast-gain effect. Averaged across the population, these disparate effects of attention gave rise to additive-offsets in contrast-response, similar to reports in human functional imaging as well as population averages of single-units. Differences in predictions for single-units and populations were observed across a wide range of model parameters (ratios of attention-field to stimulus-drive size and the amount of baseline response modifiable by attention), offering an explanation for disparity in physiological reports. Thus, by accounting for heterogeneity in tuning of realistic neuronal populations, the normalization model of attention can not only predict responses of well-tuned neurons, but also the activity of large populations of neurons. More generally, computational models can unify physiological findings across different scales of measurement, and make links to behavior, but only if factors such as heterogeneous tuning within a population are properly accounted for.
View details for DOI 10.3389/fncom.2014.00012
View details for Web of Science ID 000332510400001
View details for PubMedID 24600380
View details for PubMedCentralID PMC3928538
-
Cortical Correlates of Human Motion Perception Biases
JOURNAL OF NEUROSCIENCE
2014; 34 (7): 2592-2604
Abstract
Human sensory perception is not a faithful reproduction of the sensory environment. For example, at low contrast, objects appear to move slower and flicker faster than veridical. Although these biases have been observed robustly, their neural underpinning is unknown, thus suggesting a possible disconnect of the well established link between motion perception and cortical responses. We used functional imaging to examine the encoding of speed in the human cortex at the scale of neuronal populations and asked where and how these biases are encoded. Decoding, voxel population, and forward-encoding analyses revealed biases toward slow speeds and high temporal frequencies at low contrast in the earliest visual cortical regions, matching perception. These findings thus offer a resolution to the disconnect between cortical responses and motion perception in humans. Moreover, biases in speed perception are considered a leading example of Bayesian inference because they can be interpreted as a prior for slow speeds. Therefore, our data suggest that perceptual priors of this sort can be encoded by neural populations in the same early cortical areas that provide sensory evidence.
View details for DOI 10.1523/JNEUROSCI.2809-13.2014
View details for Web of Science ID 000331614700021
View details for PubMedID 24523549
-
Functional Signalers of Changes in Visual Stimuli: Cortical Responses to Increments and Decrements in Motion Coherence
CEREBRAL CORTEX
2014; 24 (1): 110-118
Abstract
How does our brain detect changes in a natural scene? While changes by increments of specific visual attributes, such as contrast or motion coherence, can be signaled by an increase in neuronal activity in early visual areas, like the primary visual cortex (V1) or the human middle temporal complex (hMT+), respectively, the mechanisms for signaling changes resulting from decrements in a stimulus attribute are largely unknown. We have discovered opposing patterns of cortical responses to changes in motion coherence: unlike areas hMT+, V3A and parieto-occipital complex (V6+) that respond to changes in the level of motion coherence monotonically, human areas V4 (hV4), V3B, and ventral occipital always respond positively to both transient increments and decrements. This pattern of responding always positively to stimulus changes can emerge in the presence of either coherence-selective neuron populations, or neurons that are not tuned to particular coherences but adapt to a particular coherence level in a stimulus-selective manner. Our findings provide evidence that these areas possess physiological properties suited for signaling increments and decrements in a stimulus and may form a part of cortical vigilance system for detecting salient changes in the environment.
View details for DOI 10.1093/cercor/bhs294
View details for Web of Science ID 000328373300007
View details for PubMedID 23010749
-
Demonstration of Tuning to Stimulus Orientation in the Human Visual Cortex: A High-Resolution fMRI Study with a Novel Continuous and Periodic Stimulation Paradigm
CEREBRAL CORTEX
2013; 23 (7): 1618-1629
Abstract
Cells in the animal early visual cortex are sensitive to contour orientations and form repeated structures known as orientation columns. At the behavioral level, there exist 2 well-known global biases in orientation perception (oblique effect and radial bias) in both animals and humans. However, their neural bases are still under debate. To unveil how these behavioral biases are achieved in the early visual cortex, we conducted high-resolution functional magnetic resonance imaging experiments with a novel continuous and periodic stimulation paradigm. By inserting resting recovery periods between successive stimulation periods and introducing a pair of orthogonal stimulation conditions that differed by 90° continuously, we focused on analyzing a blood oxygenation level-dependent response modulated by the change in stimulus orientation and reliably extracted orientation preferences of single voxels. We found that there are more voxels preferring horizontal and vertical orientations, a physiological substrate underlying the oblique effect, and that these over-representations of horizontal and vertical orientations are prevalent in the cortical regions near the horizontal- and vertical-meridian representations, a phenomenon related to the radial bias. Behaviorally, we also confirmed that there exists perceptual superiority for horizontal and vertical orientations around horizontal and vertical meridians, respectively. Our results, thus, refined the neural mechanisms of these 2 global biases in orientation perception.
View details for DOI 10.1093/cercor/bhs149
View details for Web of Science ID 000321163700012
View details for PubMedID 22661413
-
Modulation of Visual Responses by Gaze Direction in Human Visual Cortex
JOURNAL OF NEUROSCIENCE
2013; 33 (24): 9879-9889
Abstract
To locate visual objects, the brain combines information about retinal location and direction of gaze. Studies in monkeys have demonstrated that eye position modulates the gain of visual signals with "gain fields," so that single neurons represent both retinotopic location and eye position. We wished to know whether eye position and retinotopic stimulus location are both represented in human visual cortex. Using functional magnetic resonance imaging, we measured separately for each of several different gaze positions cortical responses to stimuli that varied periodically in retinal locus. Visually evoked responses were periodic following the periodic retinotopic stimulation. Only the response amplitudes depended on eye position; response phases were indistinguishable across eye positions. We used multivoxel pattern analysis to decode eye position from the spatial pattern of response amplitudes. The decoder reliably discriminated eye position in five of the early visual cortical areas by taking advantage of a spatially heterogeneous eye position-dependent modulation of cortical activity. We conclude that responses in retinotopically organized visual cortical areas are modulated by gain fields qualitatively similar to those previously observed neurophysiologically.
View details for DOI 10.1523/JNEUROSCI.0500-12.2013
View details for Web of Science ID 000320235300003
View details for PubMedID 23761883
-
Learning to Simulate Others' Decisions
NEURON
2012; 74 (6): 1125-1137
Abstract
A fundamental challenge in social cognition is how humans learn another person's values to predict their decision-making behavior. This form of learning is often assumed to require simulation of the other by direct recruitment of one's own valuation process to model the other's process. However, the cognitive and neural mechanism of simulation learning is not known. Using behavior, modeling, and fMRI, we show that simulation involves two learning signals in a hierarchical arrangement. A simulated-other's reward prediction error processed in ventromedial prefrontal cortex mediated simulation by direct recruitment, being identical for valuation of the self and simulated-other. However, direct recruitment was insufficient for learning, and also required observation of the other's choices to generate a simulated-other's action prediction error encoded in dorsomedial/dorsolateral prefrontal cortex. These findings show that simulation uses a core prefrontal circuit for modeling the other's valuation to generate prediction and an adjunct circuit for tracking behavioral variation to refine prediction.
View details for DOI 10.1016/j.neuron.2012.04.030
View details for Web of Science ID 000305659700017
View details for PubMedID 22726841
-
Attentional Enhancement via Selection and Pooling of Early Sensory Responses in Human Visual Cortex
NEURON
2011; 72 (5): 832-846
Abstract
The computational processes by which attention improves behavioral performance were characterized by measuring visual cortical activity with functional magnetic resonance imaging as humans performed a contrast-discrimination task with focal and distributed attention. Focal attention yielded robust improvements in behavioral performance accompanied by increases in cortical responses. Quantitative analysis revealed that if performance were limited only by the sensitivity of the measured sensory signals, the improvements in behavioral performance would have corresponded to an unrealistically large reduction in response variability. Instead, behavioral performance was well characterized by a pooling and selection process for which the largest sensory responses, those most strongly modulated by attention, dominated the perceptual decision. This characterization predicts that high-contrast distracters that evoke large responses should negatively impact behavioral performance. We tested and confirmed this prediction. We conclude that attention enhanced behavioral performance predominantly by enabling efficient selection of the behaviorally relevant sensory signals.
View details for DOI 10.1016/j.neuron.2011.09.025
View details for Web of Science ID 000297971100016
View details for PubMedID 22153378
View details for PubMedCentralID PMC3264681
-
Feature-Specific Attentional Priority Signals in Human Cortex
JOURNAL OF NEUROSCIENCE
2011; 31 (12): 4484-4495
Abstract
Human can flexibly attend to a variety of stimulus dimensions, including spatial location and various features such as color and direction of motion. Although the locus of spatial attention has been hypothesized to be represented by priority maps encoded in several dorsal frontal and parietal areas, it is unknown how the brain represents attended features. Here we examined the distribution and organization of neural signals related to deployment of feature-based attention. Subjects viewed a compound stimulus containing two superimposed motion directions (or colors) and were instructed to perform an attention-demanding task on one of the directions (or colors). We found elevated and sustained functional magnetic resonance imaging response for the attention task compared with a neutral condition, without reliable differences in overall response amplitude between attending to different features. However, using multivoxel pattern analysis, we were able to decode the attended feature in both early visual areas (primary visual cortex to human motion complex hMT+) and frontal and parietal areas (e.g., intraparietal sulcus areas IPS1-IPS4 and frontal eye fields) that are commonly associated with spatial attention. Furthermore, analysis of the classifier weight maps showed that attending to motion and color evoked different patterns of activity, suggesting that different neuronal subpopulations in these regions are recruited for attending to different feature dimensions. Thus, our finding suggests that, rather than a purely spatial representation of priority, frontal and parietal cortical areas also contain multiplexed signals related to the priority of different nonspatial features.
View details for DOI 10.1523/JNEUROSCI.5745-10.2011
View details for Web of Science ID 000288750700015
View details for PubMedID 21430149
-
Is cortical vasculature functionally organized?
NEUROIMAGE
2010; 49 (3): 1953-1956
Abstract
The cortical vasculature is a well-structured and organized system, but the extent to which it is organized with respect to the neuronal functional architecture is unknown. In particular, does vasculature follow the same functional organization as cortical columns? In principle, cortical columns that share tuning for stimulus features like orientation may often be active together and thus require oxygen and metabolic nutrients together. If the cortical vasculature is built to serve these needs, it may also tend to aggregate and amplify orientation specific signals and explain why they are available in fMRI data at very low resolution.
View details for DOI 10.1016/j.neuroimage.2009.07.004
View details for Web of Science ID 000273626400003
View details for PubMedID 19596071
-
Differential roles for frontal eye fields (FEFs) and intraparietal sulcus (IPS) in visual working memory and visual attention
JOURNAL OF VISION
2010; 10 (11)
Abstract
Cortical activity was measured with functional magnetic resonance imaging to probe the involvement of the superior precentral sulcus (including putative human frontal eye fields, FEFs) and the intraparietal sulcus (IPS) in visual short-term memory and visual attention. In two experimental tasks, human subjects viewed two visual stimuli separated by a variable delay period. The tasks placed differential demands on short-term memory and attention, but the stimuli were visually identical until after the delay period. An earlier study (S. Offen, D. Schluppeck, & D. J. Heeger, 2009) had found a dissociation in early visual cortex that suggested different computational mechanisms underlying the two processes. In contrast, the results reported here show that the patterns of activation in prefrontal and parietal cortex were different from one another but were similar for the two tasks. In particular, the FEF showed evidence for sustained delay period activity for both the working memory and the attention task, while the IPS did not show evidence for sustained delay period activity for either task. The results imply differential roles for the FEF and IPS in these tasks; the results also suggest that feedback of sustained activity from frontal cortex to visual cortex might be gated by task demands.
View details for DOI 10.1167/10.11.28
View details for Web of Science ID 000283783500028
View details for PubMedID 20884523
-
Executed and Observed Movements Have Different Distributed Representations in Human aIPS
JOURNAL OF NEUROSCIENCE
2008; 28 (44): 11231-11239
Abstract
How similar are the representations of executed and observed hand movements in the human brain? We used functional magnetic resonance imaging (fMRI) and multivariate pattern classification analysis to compare spatial distributions of cortical activity in response to several observed and executed movements. Subjects played the rock-paper-scissors game against a videotaped opponent, freely choosing their movement on each trial and observing the opponent's hand movement after a short delay. The identities of executed movements were correctly classified from fMRI responses in several areas of motor cortex, observed movements were classified from responses in visual cortex, and both observed and executed movements were classified from responses in either left or right anterior intraparietal sulcus (aIPS). We interpret above chance classification as evidence for reproducible, distributed patterns of cortical activity that were unique for execution and/or observation of each movement. Responses in aIPS enabled accurate classification of movement identity within each modality (visual or motor), but did not enable accurate classification across modalities (i.e., decoding observed movements from a classifier trained on executed movements and vice versa). These results support theories regarding the central role of aIPS in the perception and execution of movements. However, the spatial pattern of activity for a particular observed movement was distinctly different from that for the same movement when executed, suggesting that observed and executed movements are mostly represented by distinctly different subpopulations of neurons in aIPS.
View details for DOI 10.1523/JNEUROSCI.3585-08.2008
View details for Web of Science ID 000260502400014
View details for PubMedID 18971465
-
Maps of visual space in human occipital cortex are retinotopic, not spatiotopic
JOURNAL OF NEUROSCIENCE
2008; 28 (15): 3988-3999
Abstract
We experience the visual world as phenomenally invariant to eye position, but almost all cortical maps of visual space in monkeys use a retinotopic reference frame, that is, the cortical representation of a point in the visual world is different across eye positions. It was recently reported that human cortical area MT (unlike monkey MT) represents stimuli in a reference frame linked to the position of stimuli in space, a "spatiotopic" reference frame. We used visuotopic mapping with blood oxygen level-dependent functional magnetic resonance imaging signals to define 12 human visual cortical areas, and then determined whether the reference frame in each area was spatiotopic or retinotopic. We found that all 12 areas, including MT, represented stimuli in a retinotopic reference frame. Although there were patches of cortex in and around these visual areas that were ostensibly spatiotopic, none of these patches exhibited reliable stimulus-evoked responses. We conclude that the early, visuotopically organized visual cortical areas in the human brain (like their counterparts in the monkey brain) represent stimuli in a retinotopic reference frame.
View details for DOI 10.1523/JNEUROSCI.5476-07.2008
View details for Web of Science ID 000255012400018
View details for PubMedID 18400898
View details for PubMedCentralID PMC2515359
-
A temporal frequency-dependent functional architecture in human V1 revealed by high-resolution fMRI
NATURE NEUROSCIENCE
2007; 10 (11): 1404-1406
Abstract
Although cortical neurons with similar functional properties often cluster together in a columnar organization, only ocular dominance columns, the columnar structure representing segregated anatomical input (from one of the two eyes), have been found in human primary visual cortex (V1). It has yet to be shown whether other columnar organizations that arise only from differential responses to stimulus properties also exist in human V1. Using high-resolution functional magnetic resonance imaging, we have found such a functional architecture containing domains that respond preferentially to either low or high temporal frequency.
View details for DOI 10.1038/nn1983
View details for Web of Science ID 000250508400017
View details for PubMedID 17934459
-
Contrast adaptation and representation in human early visual cortex
NEURON
2005; 47 (4): 607-620
Abstract
The human visual system can distinguish variations in image contrast over a much larger range than measurements of the static relationship between contrast and response in visual cortex would suggest. This discrepancy may be explained if adaptation serves to re-center contrast response functions around the ambient contrast, yet experiments on humans have yet to report such an effect. By using event-related fMRI and a data-driven analysis approach, we found that contrast response functions in V1, V2, and V3 shift to approximately center on the adapting contrast. Furthermore, we discovered that, unlike earlier areas, human V4 (hV4) responds positively to contrast changes, whether increments or decrements, suggesting that hV4 does not faithfully represent contrast, but instead responds to salient changes. These findings suggest that the visual system discounts slow uninformative changes in contrast with adaptation, yet remains exquisitely sensitive to changes that may signal important events in the environment.
View details for DOI 10.1016/j.neuron.2005.07.016
View details for Web of Science ID 000231411100016
View details for PubMedID 16102542
View details for PubMedCentralID PMC1475737
-
A population decoding framework for motion aftereffects on smooth pursuit eye movements
JOURNAL OF NEUROSCIENCE
2004; 24 (41): 9035-9048
Abstract
Both perceptual and motor systems must decode visual information from the distributed activity of large populations of cortical neurons. We have sought a common framework for understanding decoding strategies for visually guided movement and perception by asking whether the strong motion aftereffects seen in the perceptual domain lead to similar expressions in motor output. We found that motion adaptation indeed has strong sequelae in the direction and speed of smooth pursuit eye movements. After adaptation with a stimulus that moves in a given direction for 7 sec, the direction of pursuit is repelled from the direction of pursuit targets that move within 90 degrees of the adapting direction. The speed of pursuit decreases for targets that move at the direction and speed of the adapting stimulus and is repelled from the adapting speed in the sense that the decrease either becomes greater or smaller (eventually turning to an increase) when tracking targets move slower or faster than the adapting speed. The effects of adaptation are spatially specific and fixed to the retinal location of the adapting stimulus. The magnitude of adaptation of pursuit speed and direction is uncorrelated, suggesting that the two parameters are decoded independently. Computer simulation of motion adaptation in the middle temporal visual area (MT) shows that vector-averaging decoding of the population response in MT can account for the effects of adaptation on the direction of pursuit. Our results suggest a unified framework for thinking, in terms of population decoding, about motion adaptation for both perception and action.
View details for DOI 10.1523/JNEUROSCI.0337-04.2004
View details for Web of Science ID 000224461800015
View details for PubMedID 15483122
-
Directional anisotropies reveal a functional segregation of visual motion processing for perception and action
NEURON
2003; 37 (6): 1001-1011
Abstract
Human exhibits an anisotropy in direction perception: discrimination is superior when motion is around horizontal or vertical rather than diagonal axes. In contrast to the consistent directional anisotropy in perception, we found only small idiosyncratic anisotropies in smooth pursuit eye movements, a motor action requiring accurate discrimination of visual motion direction. Both pursuit and perceptual direction discrimination rely on signals from the middle temporal visual area (MT), yet analysis of multiple measures of MT neuronal responses in the macaque failed to provide evidence of a directional anisotropy. We conclude that MT represents different motion directions uniformly, and subsequent processing creates a directional anisotropy in pathways unique to perception. Our data support the hypothesis that, at least for visual motion, perception and action are guided by inputs from separate sensory streams. The directional anisotropy of perception appears to originate after the two streams have segregated and downstream from area MT.
View details for Web of Science ID 000181899600013
View details for PubMedID 12670428
-
Serial linkage of target selection for orienting and tracking eye movements
NATURE NEUROSCIENCE
2002; 5 (9): 892-899
Abstract
Many natural actions require the coordination of two different kinds of movements. How are targets chosen under these circumstances: do central commands instruct different movement systems in parallel, or does the execution of one movement activate a serial chain that automatically chooses targets for the other movement? We examined a natural eye tracking action that consists of orienting saccades and tracking smooth pursuit eye movements, and found strong physiological evidence for a serial strategy. Monkeys chose freely between two identical spots that appeared at different sites in the visual field and moved in orthogonal directions. If a saccade was evoked to one of the moving targets by microstimulation in either the frontal eye field (FEF) or the superior colliculus (SC), then the same target was automatically chosen for pursuit. Our results imply that the neural signals responsible for saccade execution can also act as an internal command of target choice for other movement systems.
View details for DOI 10.1038/nn897
View details for Web of Science ID 000177656300020
View details for PubMedID 12145637
View details for PubMedCentralID PMC2548313
-
Linked target selection for saccadic and smooth pursuit eye movements
JOURNAL OF NEUROSCIENCE
2001; 21 (6): 2075-2084
Abstract
In natural situations, motor activity must often choose a single target when multiple distractors are present. The present paper asks how primate smooth pursuit eye movements choose targets, by analysis of a natural target-selection task. Monkeys tracked two targets that started 1.5 degrees eccentric and moved in different directions (up, right, down, and left) toward the position of fixation. As expected from previous results, the smooth pursuit before the first saccade reflected a vector average of the responses to the two target motions individually. However, post-saccadic smooth eye velocity showed enhancement that was spatially selective for the motion at the endpoint of the saccade. If the saccade endpoint was close to one of the two targets, creating a targeting saccade, then pursuit was selectively enhanced for the visual motion of that target and suppressed for the other target. If the endpoint landed between the two targets, creating an averaging saccade, then post-saccadic smooth eye velocity also reflected a vector average of the two target motions. Saccades with latencies >200 msec were almost always targeting saccades. However, pursuit did not transition from vector-averaging to target-selecting until the occurrence of a saccade, even when saccade latencies were >300 msec. Thus, our data demonstrate that post-saccadic enhancement of pursuit is spatially selective and that noncued target selection for pursuit is time-locked to the occurrence of a saccade. This raises the possibility that the motor commands for saccades play a causal role, not only in enhancing visuomotor transmission for pursuit but also in choosing a target for pursuit.
View details for Web of Science ID 000167422200029
View details for PubMedID 11245691
-
Linear and nonlinear contributions to orientation tuning of simple cells in the cat's striate cortex
VISUAL NEUROSCIENCE
1999; 16 (6): 1115-1121
Abstract
Orientation selectivity is one of the most conspicuous receptive-field (RF) properties that distinguishes neurons in the striate cortex from those in the lateral geniculate nucleus (LGN). It has been suggested that orientation selectivity arises from an elongated array of feedforward LGN inputs (Hubel & Wiesel, 1962). Others have argued that cortical mechanisms underlie orientation selectivity (e.g. Sillito, 1975; Somers et al., 1995). However, isolation of each mechanism is experimentally difficult and no single study has analyzed both processes simultaneously to address their relative roles. An alternative approach, which we have employed in this study, is to examine the relative contributions of linear and nonlinear mechanisms in sharpening orientation tuning. Since the input stage of simple cells is remarkably linear, the nonlinear contribution can be attributed solely to cortical factors. Therefore, if the nonlinear component is substantial compared to the linear contribution, it can be concluded that cortical factors play a prominent role in sharpening orientation tuning. To obtain the linear contribution, we first measure RF profiles of simple cells in the cat's striate cortex using a binary m-sequence noise stimulus. Then, based on linear spatial summation of the RF profile, we obtain a predicted orientation-tuning curve, which represents the linear contribution. The nonlinear contribution is estimated as the difference between the predicted tuning curve and that measured with drifting sinusoidal gratings. We find that measured tuning curves are generally more sharply tuned for orientation than predicted curves, which indicates that the linear mechanism is not enough to account for the sharpness of orientation-tuning. Therefore, cortical factors must play an important role in sharpening orientation tuning of simple cells. We also examine the relationship of RF shape (subregion aspect ratio) and size (subregion length and width) to orientation-tuning halfwidth. As expected, predicted tuning halfwidths are found to depend strongly on both subregion length and subregion aspect ratio. However, we find that measured tuning halfwidths show only a weak correlation with subregion aspect ratio, and no significant correlation with RF length and width. These results suggest that cortical mechanisms not only serve to sharpen orientation tuning, but also serve to make orientation tuning less dependent on the size and shape of the RF. This ensures that orientation is represented equally well regardless of RF size and shape.
View details for Web of Science ID 000084409800011
View details for PubMedID 10614591