Contextual and top-down influences in vision

Time/Room: Friday, May 10, 1:00 – 3:00 pm, Royal 4-5
Organizer: Uri Polat, Tel-Aviv University
Presenters: Charles Gilbert, Uri Polat, Rudiger von der Heydt, Pieter Roelfsema, Dennis Levi, Dov Sagi

< Back to 2013 Symposia

Symposium Description

According to classical models of spatial vision, the output of neurons in the early visual cortex is determined by the local features of the stimuli and integrated at later stages of processing (feedforward). However, experimental results obtained during the last two decades show contextual modulation: local perceptual effects are modulated by global image properties. The receptive field properties of cortical neurons are subject to learning and to top-down influences of attention, expectation and perceptual task. Even at early cortical stages of visual processing neurons are subject to contextual influences that play a role in intermediate level vision, contour integration and surface segmentation, which enables them to integrate information over large parts of the visual field. These influences are not fixed but are subject to experience, enabling neurons to encode learned information. The dynamic properties of context modulations are mediated by an interaction between reentrant signals to the cortex and intrinsic cortical connections, changing effective connectivity within the cortical network. The evolving view of the nature of the receptive field includes contextual influences which change in the long term as a result of perceptual learning and in the short term as a result of a changing behavioral context. In the symposia we will present anatomical, physiological and psychophysical data showing contextual effects in lateral interactions, grouping, border ownership, crowding and perceptual learning.

Presentations

Contextual modulation in the visual cortex

Speaker: Charles Gilbert, The Rockefeller University, New York

Vision is an active process. The receptive field properties of cortical neurons are subject to learning and to top-down influences of attention, expectation and perceptual task. Even at early cortical stages of visual processing neurons are subject to contextual influences that play a role in intermediate level vision, contour integration and surface segmentation, which enables them to integrate information over large parts of the visual field. These influences are not fixed but are subject to experience, enabling neurons to encode learned information. Even in the adult the visual cortex there is considerable plasticity, where cortical circuits undergo exuberant changes in axonal arbors following manipulation of sensory experience. The integrative properties of cortical neurons, the contextual influences that confer selectivity to complex stimuli, are mediated in part by a plexus of long range horizontal connections that enable neurons to integrate information over an area of visual cortex representing large parts of the visual field. These connections are the substrate for an association field, a set of interactions playing a role in contour integration and saliency. The association field is not fixed. Rather, neurons can select components of this field to express difference functional properties. As a consequence neurons can be thought of as adaptive processors, changing their function according to behavioral context, and their responses reflect the demands of the perceptual task being performed. The top-down signal facilitates our ability to segment the visual scene despite its complex arrangement of objects and backgrounds. It plays a role in encoding and recall of learned information. The resulting feedforward signals carried by neurons convey different meanings according to the behavioral context. We propose that these dynamic properties are mediated by an interaction between reentrant signals to the cortex and intrinsic cortical connections, changing effective connectivity within the cortical network. The evolving view of the nature of the receptive field includes contextual influences which change in the long term as a result of perceptual learning and in the short term as a result of a changing behavioral context.

Spatial and temporal rules for contextual modulations

Speaker: Uri Polat, Tel-Aviv University, Tel-Aviv, Israel

Most contextual modulations, such as center-surround and crowding exhibit a suppressive effect. In contrast, collinear configuration is a unique case of contextual modulation in which the effect can be either facilitative or suppressive, depending on the context. Physiological and psychophysical studies revealed several spatial and temporal rules that determine the modulation effect: 1) spatial configuration: collinear configuration can be either facilitative or suppressive, whereas non-collinear configurations may be suppressive; 2) separation between the elements: suppression for close separation that coincides with the size of the receptive field and facilitation outside the receptive field; 3) activity dependent: facilitation for low contrast (near the threshold) and suppression for high contrast; 4) temporal properties: suppression is fast and transient, whereas facilitation is delayed and sustained; 5) attention may enhance the facilitation; 6) slow modulation: perceptual learning can increase the facilitatory effect over a time scale of several days; 7) fovea and periphery: similar rules can be applied when spatial scaling to the size of receptive field is done. It is believed that the role of collinear facilitation is to enhance contour integration and object segmentation, whereas center-surround is important for pop-out. Our recent studies suggest that these rules can serve as a unified model for spatial and temporal masking as well as for crowding.

Border ownership and context

Speaker: Rudiger von der Heydt, The Johns Hopkins University, Baltimore, Maryland, USA

A long history of studies of perception has shown that the visual system organizes the incoming information early on, interpreting the 2D image in terms of a 3D world, and producing a structure that enables object-based attention and tracking of object identity. Recordings from monkey visual cortex show that many neurons, especially in area V2, are selective for border ownership. These neurons are edge selective and have ordinary classical receptive fields, but in addition, their responses are modulated (enhanced or suppressed) depending on the location of a ‘figure’ relative to the edge in their receptive field. Each neuron has a fixed preference for location on one side or the other. This selectivity is derived from the image context far beyond the classical receptive field. This talk will review evidence indicating that border ownership selectivity reflects mechanisms of object definition. The evidence includes experiments showing (1) reversal of border ownership signals with change of perceived object structure, (2) border ownership specific enhancement of responses in object-based selective attention, )3) persistence of border ownership signals in accordance with continuity of object perception, and (4) remapping of border ownership signals across saccades and object movements. Some of these findings can be explained by assuming that grouping circuits detect ‘objectness’ according to simple rules, and, via recurrent projections, enhance the low-level feature signals representing the object. This might be the mechanism of object-based attention. Additional circuits may provide persistence and remapping.

Visual cortical mechanisms for perceptual grouping

Speaker: Pieter Roelfsema, Netherlands Institute for Neuroscience, Amsterdam, the Netherlands

A fundamental task of vision is to group the image elements that belong to one object and to segregate them from other objects and the background. I will discuss a new conceptual framework that explains how the binding problem is solved by the visual cortex. According to this framework, two mechanisms are responsible for binding: base-grouping and incremental grouping. Base-groupings are coded by single neurons tuned to multiple features, like the combination of a color and an orientation. They are computed rapidly because they reflect the selectivity of feedforward connections that propagate information from lower to higher areas of the visual cortex. However, not all conceivable feature combinations are coded by dedicated neurons. Therefore, a second, flexible incremental grouping mechanism is required. Incremental grouping relies on horizontal connections between neurons in the same area and feedback connections that propagate information from higher to lower areas. These connections spread an enhanced response (not synchrony) to all the neurons that code image elements that belong to the same perceptual object. This response enhancement acts as a label that tags those neurons that respond to image elements to be bound in perception. The enhancement of neuronal activity during incremental grouping has a correlate in psychology because object-based attention is directed to the features labeled with the enhanced neuronal response. Our recent results demonstrate that feedforward and feedback processing rely on different receptors for glutamate and on processing in different cortical layers.

Crowding in context

Speaker: Dennis Levi, UC Berkeley, Berkeley, CA, USA

In peripheral vision, objects that can be readily recognized when viewed in isolation, become unrecognizable in clutter. This is the interesting phenomenon known as visual crowding. Crowding represents an essential bottleneck, setting limits on object perception, eye movements, visual search, reading and perhaps other functions in peripheral, amblyopic and developing vision (Whitney & Levi, 2011). It is generally defined as the deleterious influence of nearby contours on visual discrimination, but the effects of crowding go well beyond impaired discrimination. Crowding impairs the ability to recognize and respond appropriately to objects in clutter. Thus, studying crowding may lead to a better understanding of the processes involved in object recognition. Crowding also has important clinical implications for patients with macular degeneration, amblyopia and dyslexia. Crowding is strongly dependent on context. The focus of this talk will be on trying to put crowding into context with other visual phenomena.

Perceptual learning in context

Speaker: Dov Sagi, The Weizmann Institute of Science, Rehovot, Israel

Studies of perceptual learning show a large diversity of effects, with learning rate and specificity varying across stimuli and experimental conditions. Most notably, there is an initial fast phase of within session (online) learning followed by a slower phase, taking place over days, which is highly specific to basic image features. Our results show that the latter phase is highly sensitive to contextual modulation. While thresholds for contrast discrimination of a single Gabor patch are relatively stable and unaffected by training, the addition of close flankers induces dramatic improvements of thresholds, indicating increased gain of the contrast response function (“context enabled learning”). Cross-orientation masking effects can be practically eliminated by practice. In texture discrimination, learning was found to interact with slowly evolving adaptive effects reducing the effects of learning. These deteriorative effects can be eliminated by cross-orientation interactions found to counteract sensory adaptation. The experimental results are explained by plasticity within local networks of early vision assuming excitatory-inhibitory interactions, where context modulates the balance between excitation and inhibition. We suggest that reduced inhibitory effects increases learning efficiency, making learning faster and generalizable. Specificity of learning seems to be the result of experience dependent local contextual interactions.

< Back to 2013 Symposia

The structure of visual working memory

Time/Room: Friday, May 10, 1:00 – 3:00 pm, Royal 1-3
Organizer: Wei Ji Ma, Baylor College of Medicine
Presenters: Steven J. Luck, Wei Ji Ma, Paul M. Bays, George Alvarez, Robert Jacobs

< Back to 2013 Symposia

Symposium Description

TWO THEORETICAL ISSUES Working memory is an essential component of perception, cognition, and action. The past eight years have seen a surge of activity aimed at understanding the structure of visual working memory. This symposium brings together some of the leading thinkers in this field to discuss two central theoretical issues: slots versus resources, and the role of context. SLOTS VERSUS RESOURCES Working memory is widely believed to be subject to an item limit: no more than a fixed number of items can be stored and any additional items are forgotten. In 2004, Wilken and Ma challenged this notion and advocated for an alternative framework in which a continuous memory resource is divided over all items and errors are explained in terms of the quality of encoding rather than the quantity of remembered items. Since then, arguments have been made on both sides, notably by speakers in this symposium (Luck, Bays, Alvarez, Ma). New concepts that have been introduced in this debate include variable precision, non-target reports, Bayesian inference, and the neural substrate of memory resource. Intriguingly, all speakers have recently used the same visual working memory paradigm – delayed estimation – to draw sometimes conflicting conclusions. Therefore, we expect a vivid exchange of thoughts. THE ROLE OF CONTEXT In the slots-versus-resources debate, items are routinely assumed to be encoded independently in working memory. This assumption is likely to be wrong, but how wrong? Recent work has pointed out the large effects of the context in which an item is presented. Items seem to be remembered in groups or ensembles organized by space or feature, and this introduces predictable biases. Hierarchical Bayesian models have been proposed by the groups of Alvarez and Jacobs to quantify context effects. They will both be speaking about these data and models. TARGET AUDIENCE The symposium aims to present current debates and open questions in the study of visual working memory to a broad audience. We believe this symposium will be of interest to students, postdocs, and faculty. The contents should be useful to a very large VSS audience: anyone studying multiple-object working memory or attention using psychophysics, electrophysiology, modeling, neuroimaging, or EEG/MEG. The symposium could benefit them by suggesting new theoretical frameworks to think about data, as well as new experimental paradigms.

Presentations

Continuous versus discrete models of visual working memory capacity

Speaker: Steven J. Luck, University of California, Davis
Authors: Weiwei Zhang, University of California, Davis

Working memory plays a key role in visual cognition, allowing the visual system to span the gaps created by blinks and saccades and providing a major source of control over attention and eye movements. Moreover, measurements of visual working memory capacity for simple visual features are strongly correlated with individual differences in higher cognitive abilities and are related to psychiatric and neurological disorders. It is therefore critically important that we understand the nature of capacity limits in visual working memory. Two major classes of theories have been proposed, discrete theories in which a limited number of items can be concurrently stored with high resolution, and continuous theories in which a potentially limitless number of items can be stored by reducing the precision of the representations. In this talk, we will review 15 years of research on the nature of visual working memory representations and present new evidence that favors discrete representations.

Continuous resources and variable precision in working memory

Speaker: Wei Ji Ma, Baylor College of Medicine
Authors: Ronald van den Berg, Baylor College of Medicine; Hongsup Shin, Baylor College of Medicine

In comparisons between item-limit and continuous-resource models of working memory, the continuous-resource model tested is usually a stereotyped one in which memory resource is divided equally among items. This model cannot account for human behavior. We recently introduced the notion that resource (mnemonic precision) is variable across items and trials. This model provides excellent fits to data and outperforms item-limit models in explaining delayed-estimation data. When studying change detection, a model of memory is not enough, since the task contains a decision stage. Augmenting the variable-precision model of memory with a Bayesian decision model provides the best available account of change detection performance across set sizes and change magnitudes. Finally, we argue that variable, continuous precision has a plausible neural basis in the gain of a neural population. Our results and those of other groups overhaul long-held beliefs about the limitations of working memory.

Working memory capacity and allocation reflect noise in neural storage

Speaker: Paul M. Bays, University College London

A key claim differentiating “resource” from “slot” models of WM is that resources can be allocated flexibly, enhancing the mnemonic precision of some visual elements at a cost to others. While salient visual events are found to have a short-lived influence on WM that is rapidly suppressed, informative cues lead to a long-lasting reallocation of resources. We argue that resource limits in working memory are a direct consequence of stochasticity (noise) in neural representations. A model based on population coding reproduces the empirical relationship between error distributions and memory load and demonstrates that observers allocate limited neural resources in a near-optimal fashion.

Beyond Slots vs. Resources

Speaker: George Alvarez, Harvard University
Authors: Timothy Brady, Harvard University; Daryl Fougnie, Harvard University; Jordan Suchow, Harvard University

Slot and resource models have been influential in the study of visual working memory capacity. However, several recent empirical findings are not explicitly predicted by either model. These findings include: (1) a shared limit on the fidelity of working memory and long-term memory, (2) stochastic variability in working memory that is not explained by uneven allocation of a commodity such as slots or resources, and (3) the existence of structured representations. Together, these findings demand either significant modification of existing slot and resource models, or the introduction of a new framework for understanding visual working memory capacity.

A Probabilistic Clustering Theory of the Organization of Visual Short-Term Memory

Speaker: Robert Jacobs, University of Rochester
Authors: A. Emin Orhan, University of Rochester

Some models of visual short-term memory (VSTM) assume that memories for individual items are independent. Recent experimental evidence indicates that this assumption is false. People’s memories for individual items are influenced by the other items in a scene. We develop a Probabilistic Clustering Theory (PCT) for modeling the organization of VSTM. PCT states that VSTM represents a set of items in terms of a probability distribution over all possible clusterings or partitions of those items. Because PCT considers multiple possible partitions, it can represent an item at multiple granularities or scales simultaneously. Moreover, using standard probabilistic inference, it automatically determines the appropriate partitions for the particular set of items at hand, and the probabilities or weights that should be allocated to each partition. A consequence of these properties is that PCT accounts for experimental data that have previously motivated hierarchical models of VSTM, thereby providing an appealing alternative to hierarchical models with pre-specified, fixed structures.

< Back to 2013 Symposia

2013 Symposia

The structure of visual working memory

Organizer: Wei Ji Ma, Baylor College of Medicine
Time/Room: Friday, May 10, 1:00 – 3:00 pm, Royal 1-3

Working memory is an essential component of perception, cognition, and action. The past eight years have seen a surge of activity aimed at understanding the structure of visual working memory. Is working memory performance limited by a maximum number of objects that can be remembered, or by the quality of the memories? Does context affect how we remember objects? This symposium brings together some of the leading thinkers in this field to discuss these central theoretical issues. More…

Contextual and top-down influences in vision

Organizer: Uri Polat, Tel-Aviv University
Time/Room: Friday, May 10, 1:00 – 3:00 pm, Royal 4-5

Vision is an active process. The properties of cortical neurons are subject to learning and to top-down influences of attention, expectation and perceptual task. Even at early cortical stages of visual processing neurons are subject to contextual influences that play a role in our vision, These influences are not fixed but are subject to experience, enabling neurons to encode learned information. In the symposia we will present anatomical, physiological and psychophysical data showing contextual effects in almost every visual task. We will show that visual perception involves both instantaneous pre-attentive and attentive processes that enhance the visual perception. More…

Active Perception: The synergy between perception and action

Organizer: Michele Rucci, Boston University & Eli Brenner, VU University
Time/Room: Friday, May 10, 1:00 – 3:00 pm, Royal 6-8

Visual perception is often studied in a passive manner without consideration of motor activity. Like many other species, however, humans are not passively exposed to the incoming flow of sensory data. They actively seek useful information by coordinating sensory processing with motor activity. In fact, behavior is a key component of sensory perception, as it enables control of sensory signals in ways that simplify perceptual tasks. This workshop will focus on recent findings which have further emphasized the tight link between perception and action. More…

ARVO@VSS: Visual Development

Organizers: Susana Chung, University of California, Berkeley and Anthony Norcia, Stanford University
Time/Room: Friday, May 10, 3:30 – 5:30 pm, Royal 1-3

Many visual functions continue to develop and reach adult levels only in late childhood. The successful development of normal visual functions requires ‘normal’ visual experience. The speakers of this symposium will review the time courses of normal visual development of selected visual functions, and discuss the consequences of abnormal visual experience during development on these visual functions. The prospect of recovering visual functions in adults who experienced abnormal visual experience during development will also be discussed, along with the advances made in the assessment of visual functions in children with abnormal visual development due to damage to the visual cortex and the posterior visual pathways. More…

Decoding and the spatial scale of cortical organization

Organizer: Jeremy Freeman, New York University; Elisha P. Merriam, Departments of Psychology and Neural Science, New York University; and Talia Konkle, Department of Psychology, Harvard University
Time/Room: Friday, May 10, 3:30 – 5:30 pm, Royal 4-5

With functional neuroimaging data we have incredible access to a rich landscape of neural responses, but this access comes with challenging questions: Over what expanse of cortex is information meaningfully clustered — in other words, over what scales should we expect neural information to be organized? How should inferences about cortical organization take into account the complex nature of the imaging signal, which reflects neural and non-neural signals at multiple spatial scales? In this symposium, six investigators discuss representational structure at multiple spatial scales across the cortex, highlighting the inferential strengths and weaknesses of cutting-edge analyses across multiple experimental techniques. More…

Does appearance matter?

Organizer: Sarah R. Allred, Rutgers–The State University of New Jersey
Time/Room: Friday, May 10, 3:30 – 5:30 pm, Royal 6-8

Vision science originated with questions about how and why things look the way do, but phenomenology is sometimes given short shrift in the field as a whole. We discuss objective methods that capture what we mean by appearance and examine the criteria for behaviors that are best thought of as mediated by reasoning about appearances. By utilizing phenomenology, we provide a parsimonious understanding of many empirical phenomena, including instructional effects in lightness perception, contextual effects on color constancy, systematic biases in egocentric distance perception and predicting 3D shape from orientation flows. We also discuss contemporary interactions between appearance, physiology, and neural models. More…

Vision Sciences Society