Richard D. Mooney, Ph.D., PI
We study the neurobiology of learned vocal communication, especially experience-dependent vocal learning, and the synaptic basis for vocal learning and communication. Our approaches include in vivo multi photon imaging of neurons, optogenetics, and intracellular and extracellular recordings from freely behaving birds and mice.
Our research aims to identify the neural substrates for vocal learning and communication. We use both songbird and rodents to achieve these aims. Songbirds are one of the few non-human animals that learn to vocalize and serve as the preeminent model in which to identify neural mechanisms for vocal learning. The songbird is ideal for this purpose because of its well-described capacity to vocally imitate the songs of other birds, and because its brain has a constellation of discrete, interconnected brain regions (i.e., song control nuclei, referred to collectively as the song system) that function in the patterning, perception, learning and maintenance of song.
There are two major foci to our songbird studies: elucidating how and where auditory and motor information about learned vocalizations is encoded in the brain; identifying the mechanisms via which auditory experience modifies vocal output, as occurs during sensitive periods for vocal learning. We also study the neurobiology of vocal communication in mice. Although mice do not appear to be vocal learners, they do vocalize to communicate. A major focus of our current research is to understand how vocal premotor and auditory areas interact during self-generated vocalizations and whether they interact when the individual listens to the vocalizations of others. We are using both wild type and genetically modified mice to identify the central neural mechanisms that underlie this form of sensorimotor integration.
We use a wide range of techniques in our research, including in vivo multiphoton neuronal imaging, chronic recording of neural activity in freely behaving animals, in vivo and in vitro intracellular recordings from identified neurons, and manipulation of neuronal activity using either electrical microstimulation or optogenetic methods. Our group also has extensive experience with viral transgenic methods and with behavioral analysis, especially in quantifying acoustic features of vocalizations. Together, these methods provide a broad technical approach to understanding how the brain harnesses sensory information to adaptively modify behavior.