Neuro Tuesday Spring 2002




NeuroTuesday Spring 2002




5-6:30 PM Tuesdays in Warren Weaver Hall Room 1314



February 5: Nava Rubin (NYU) The dynamics of perceptual bi-stability in ambiguous motion displays

February 19: Kukjin Kang (NYU) Population codes and distance measures in probability

March 5: Tai Sing Lee (Carnegie Mellon) Dynamics and plasticity in early visual processing

March 12: No Seminar Spring Break

April 16: John Foxe's seminar has been POSTPONED

April 23: Roger Tootell (MGH) Forward into the Past: fMRI of visual cortex in awake behaving macaques

April 30: John Foxe (Nathan Kline Institute & Albert Einstein) The case for early feedforward integration of multisensory inputs




The Computational Neuroscience Forum & CIMS NeuroTuesday Joint WebPage





Abstracts


Tai Sing Lee (Carnegie Mellon) Dynamics and plasticity in early visual processing


The primary visual cortex (V1) has traditionally been considered to be a piece of neural machinery that is relatively hardwired to perform and edge detection. In this talk, I will present some recent evidence from primate electrophysiological experiments that demonstrate that V1 neurons are in fact quite dynamic and plastic in what they are encoding and in the way they are computing. For example, they are sensitive to the statistics of the stimulus environment, changing their tuning properties on the fly, and that their processing of perceptual information is a function of higher order stimulus properties, behavioral experience and task demands. Taken together, we conclude that V1 is not simply the first stage of visual processing, but an integral part of the reciprocally connected system that participates in many levels of visual perceptual inference. I will also sketch a computational framework on how V1 can serve as a high-resolution buffer for integrating information and coordinating computation across multiple visual modules.


John Foxe (NKI & Albert Einstein) The case for early feedforward integration of multisensory inputs


A widely held view in the field of multisensory science is that sensory information from a single object (e.g. a barking dog) is first processed extensively through the respective unisensory processing streams before the information is combined in higher-order multisensory regions of cortex. Under this view, multisensory modulations which have recently been found in early sensory cortices during hemodynamic imaging studies have been interpreted as reflecting feedback modulations that occur subsequent to multisensory processing in the higher-order areas. Of course, hemodynamic imaging cannot address the relative timing of such activations. In this talk, I will focus on recent evidence from both human and monkey electrophysiological investigations as well as hemodynamic imaging studies that challenge an exclusively feedback model. Specifically, these studies show that multisensory integration effects are seen at very early latencies relative to the well-established timecourse of sensory activation within the respective unisensory processing streams. In fact, intracranial recordings show that multisensory inputs found in early sensory cortices display a characteristic feedforward pattern of inputs. For many years, it was believed that there was no direct anatomic connections between early sensory cortices of the different sensory modalities but recent anatomic tracer studies have begun to uncover direct connections between early visual and auditory sensory regions, providing the necessary anatomic substrate for such early interactions. In light of these recent findings, it is becoming evident that models of multisensory cortical processing will need to incorporate early feedforward integration mechanisms as well as feedback. These data highlight the tremendous potential of "multimodality" imaging, whereby data from the macrospcopic noninvasive level (fMRI and high-density scalp recordings in humans) are integrated with equivalent data from the more microscopic/cellular level which we record intracranially in non-human primates.


Roger Tootell (MGH) Forward into the Past: fMRI in awake behaving macaques


For many years, our understanding of the brain relied largely on data from macaque monkeys and other animals, because it is usually not possible to study the human brain invasively. During the past decade, this information from animals has been increasingly supplemented by noninvasive studies of human brain, using functional magnetic resonance imaging (fMRI). However it still remained impossible to compare these two data sets directly , since the data from monkeys was generated using quite different techniques (e.g., single unit recording, neural tracing) compared to the techniques used to study human brain (e.g. fMRI, psychophysics, etc.). Thus one often could not say whether a given discrepancy arose from differences between the species tested (e.g. humans versus macaques) or in the techniques used to reveal those differences (e.g. single unit recording versus fMRI).

To resolve some of these issues, and to open up a wide range of potential future experiments, we began doing fMRI in visual cortex of macaque monkeys. Such studies were done at 3 T, in a horizontal bore scanner, using the BOLD signal, in awake behaving primates --- exactly as in parallel fMRI experiments on human subjects. To minimize the experimental differences even more, we showed both human and macaque subjects the same visual stimuli, and had them perform equivalent tasks in the scanner. Data analysis (in flattened cortical surface format) was also equivalent for both primate families. To compensate for the much-smaller size of the macaque brain, we scanned macaques at higher spatial resolution (to 1 mm3).

Initial studies included fMRI tests for retinotopy, color, motion, stereopsis and face selectivity. As one might expect, we found broad similarities in the organization of visual cortex in macaques and humans --- but distinctive differences were found as well. The distinctions include: 1) a difference in the location of a supplementary motion-selective area --- V3 in macaques, instead V3A in humans; 3) retinotopic differences (e.g. V3, V4d), 3) a somewhat different organization for the processing of stereopsis --- although V3A and caudal intraparietal areas are clearly specialized for stereopsis in both humans and macaques. Comforting likenesses were also found. For instance, the fMRI revealed a face-selective area in inferotemporal cortex of macaque monkey --- quite similar to previous descriptions of a face-selective area in human inferotemporal cortex ('FFA').

Such differences and similarities should be expected from members of the same order, which have been evolutionarily separated by ~ 30 million years. It is crucial to become informed about such differences and similarities, if the macaque is to serve optimally as a ``model system'' for the human brain.




Nava Rubin (NYU) The dynamics of perceptual bi-stability in ambiguous motion displays


Perceptual bi-stability occurs when the same stimulus leads to two (or more) different interpretations. Upon prolonged observation, the perception of the stimulus alternates haphazardly over time. Plaids are ambiguous stimuli that can be seen as a single pattern translating rigidly ("coherency") or split into two gratings sliding over each other ("transparency"). Although they cause bi-stability, nearly all the work using plaids to study motion integration and segmentation was with brief presentations, neglecting the dynamical aspect. We studied the dynamics of perceptual alternations in plaids by measuring the cummulative time spent perceiving each interpretation, as well as the mean duration of individual epochs. We show that the dynamics approach provides more sensitive measures of the relative strength of the two interpretations, and therefore can shed new light on motion integration/segmentation. We also show that a fundamental result about the dynamics of alternations in binocular rivalry (Levelt 1967) can be generalized to the case of plaids, suggesting there may be general rules governing the dynamics of bi-stable perception. Specifically, it suggests a suppressive/inhibitory coupling between rivaling stimulus interpretations as a general principle in the brain.

Joint work with Jean-Michel Hupe


A Recap of NeuroTuesday Spring 2001

Louis Tao, My CIMS web page, Applied Mathematics Laboratory & CNS, New York, NY 10012. This page last updated Jan 2001.