Visual motion perception is critical to animals. It guides a wide range of behaviors, from navigation and predator avoidance to mating. This project proposes to dissect a newfound visual motion computation in the fruit fly Drosophila, a model organism with powerful genetic tools that can define the roles of individual neurons in neural computations. With these tools, this research will map out the neurons that compute the new motion signal and the way they compute it. This work is significant for two reasons. First, this new motion computation is a different algorithm from classical motion estimation. Classical estimation is well-described by the Hassenstein-Reichardt correlator (HRC), an influential model that has long guided research on visual motion detection in insects, and whose descendants guide motion perception research in animals from mice to primates. The motion-guided behavior studied here appears to be fundamental, but the motion signal is computed in a qualitatively different way from the HRC model's predictions. This new algorithm and its implementation will add to the suite of potential visual motion detectors. Because of the parallels in visual computations between flies and vertebrates, it is likely that vertebrate visual systems make use of a similar algorithm. Second, animals compute motion at several places in their visual systems, but it remains unclear how the different computations are used. Analysis in the compact fly brain will uncover principles for understanding the larger brains of mammals. In the fly, genetic tools and behavioral measurements can be combined to investigate three questions about its motion detection: How do its two motion systems compute different signals? How do they use overlapping circuitry? And how do they guide different behaviors? These questions lead to three main aims of this research.
Aim 1 will characterize the algorithm that computes the new motion signal. Behavioral measurements and targeted visual stimuli will constrain or rule out potential models and provide a mathematical analysis of how this motion estimate is extracted from visual inputs.
Aim 2 will identify neurons required for the new motion computation. Genetic tools will be used to silence specific neurons in the visual system in order to identify which neurons are required for the new computation.
Aim 3 will measure the functional response properties of visual neurons in the new motion circuit. Measurements of neuron responses to stimuli will show how these neurons combine signals to generate the observed motion signals. On completion, these studies will result in a detailed understanding of the algorithm and the neural mechanisms that implement the new motion computation. The new computation and its interactions with the classical motion detector will provide a template for understanding how multiple motion signals are generated and used by the brain.

Public Health Relevance

This research investigates the neural basis for computing visual motion and how different motion signals in the brain are used for different purposes. It is essential to know the full range of early visual computations, including their neural implementation and uses, if one is to understand and treat visual impairments and blindness. Understanding these neural computations will also contribute to improving artificial visual computations, such as those in retinal prosthetics

Agency
National Institute of Health (NIH)
Institute
National Eye Institute (NEI)
Type
Research Project (R01)
Project #
5R01EY026555-04
Application #
9686745
Study Section
Neurotransporters, Receptors, and Calcium Signaling Study Section (NTRC)
Program Officer
Flanders, Martha C
Project Start
2016-04-01
Project End
2021-03-31
Budget Start
2019-04-01
Budget End
2020-03-31
Support Year
4
Fiscal Year
2019
Total Cost
Indirect Cost
Name
Yale University
Department
Biochemistry
Type
Schools of Arts and Sciences
DUNS #
043207562
City
New Haven
State
CT
Country
United States
Zip Code
06520
Astigarraga, Sergio; Douthit, Jessica; Tarnogorska, Dorota et al. (2018) Drosophila Sidekick is required in developing photoreceptors to enable visual motion detection. Development 145:
Salazar-Gatzimas, Emilio; Agrochao, Margarida; Fitzgerald, James E et al. (2018) The Neuronal Basis of an Illusory Motion Percept Is Explained by Decorrelation of Parallel Motion Pathways. Curr Biol 28:3748-3762.e8
Creamer, Matthew S; Mano, Omer; Clark, Damon A (2018) Visual Control of Walking Speed in Drosophila. Neuron :
Demb, Jonathan B; Clark, Damon A (2017) Vision: These retinas are made for walkin'. Nature 546:476-477
Mano, Omer; Clark, Damon A (2017) Graphics Processing Unit-Accelerated Code for Computing Second-Order Wiener Kernels and Spike-Triggered Covariance. PLoS One 12:e0169842
Clark, Damon A; Demb, Jonathan B (2016) Parallel Computations in Insect and Mammalian Visual Motion Processing. Curr Biol 26:R1062-R1072
Salazar-Gatzimas, Emilio; Chen, Juyue; Creamer, Matthew S et al. (2016) Direct Measurement of Correlation Responses in Drosophila Elementary Motion Detectors Reveals Fast Timescale Tuning. Neuron 92:227-239