a. Functions of the module. This module will serve the expanding computing needs of CVS Core users. The past five years have seen many exciting developments in vision science and translational research in ophthalmology at Rochester and with those developments, the need for new software and new computing technologies has increased. Perhaps most important among these changes has been the development of new multi-user research resources, and the proliferation of several sets of common technologies to a number of different research communities within CVS. Included among the multi-user resources are a new virtual reality lab in CVS for the study of perceptual learning and multi-sensory integration in naturalistic environments, a 52-processor computing cluster within CVS for data analysis and for modeling and simulation of large-scale neural networks, and the Rochester Center for Brain imaging, devoted to applying fMRI to studying vision and cognition. Included among the common technologies are a number of common imaging devices (e.g., adaptive optics systems) used by both the retinal imaging and vision correction communities, common multi-electrode recording systems for electrophysiology, and common stimulus display systems used by both neurophysiology and behavioral labs. These common resources require significant support from the computing module. Going forward, the computing module will make important contributions to projects that include: massively parallel simulations of neural networks, new virtual-reality software for multi-sensory integration experiments in both human and non-human primates, super-resolution fMRI for studying the human LGN, and imaging the living human eye using adaptive optics. The computing module is currently staffed by four highly qualified full-time technicians, who assist the CVS investigators using the Module.

National Institute of Health (NIH)
National Eye Institute (NEI)
Center Core Grants (P30)
Project #
Application #
Study Section
Special Emphasis Panel (ZEY1)
Project Start
Project End
Budget Start
Budget End
Support Year
Fiscal Year
Total Cost
Indirect Cost
University of Rochester
United States
Zip Code
Zinszer, Benjamin D; Bayet, Laurie; Emberson, Lauren L et al. (2018) Decoding semantic representations from functional near-infrared spectroscopy signals. Neurophotonics 5:011003
McGregor, Juliette E; Yin, Lu; Yang, Qiang et al. (2018) Functional architecture of the foveola revealed in the living primate. PLoS One 13:e0207102
Lockwood, Colin T; Vaughn, William; Duffy, Charles J (2018) Attentional ERPs distinguish aging and early Alzheimer's dementia. Neurobiol Aging 70:51-58
Chernoff, Benjamin L; Teghipco, Alex; Garcea, Frank E et al. (2018) A Role for the Frontal Aslant Tract in Speech Planning: A Neurosurgical Case Study. J Cogn Neurosci 30:752-769
Alarcon-Martinez, Luis; Yilmaz-Ozcan, Sinem; Yemisci, Muge et al. (2018) Capillary pericytes express ?-smooth muscle actin, which requires prevention of filamentous-actin depolymerization for detection. Elife 7:
Jeon, Kye-Im; Hindman, Holly B; Bubel, Tracy et al. (2018) Corneal myofibroblasts inhibit regenerating nerves during wound healing. Sci Rep 8:12945
Weiss, Menachem Y; Kuriyan, Ajay E (2018) Acute Monocular Vision Loss in a Young Adult. JAMA Ophthalmol 136:297-298
Chapman, Robert M; Gardner, Margaret N; Klorman, Rafael et al. (2018) Temporospatial components of brain ERPs as biomarkers for Alzheimer's disease. Alzheimers Dement (Amst) 10:604-614
Prentiss, Emily K; Schneider, Colleen L; Williams, Zoƫ R et al. (2018) Spontaneous in-flight accommodation of hand orientation to unseen grasp targets: A case of action blindsight. Cogn Neuropsychol 35:343-351
Bosen, Adam K; Fleming, Justin T; Allen, Paul D et al. (2018) Multiple time scales of the ventriloquism aftereffect. PLoS One 13:e0200930

Showing the most recent 10 out of 211 publications