The Orpella Lab aims to further our understanding of speech production, auditory perception, and language learning. We use available cognitive neuroscience methods (e.g., magnetoencephalography, magnetic resonance imaging, transcranial magnetic stimulation) and techniques (e.g., psychophysics, modeling, machine learning) to advance theoretically, computationally, and biologically grounded accounts of how these cognitive feats may be achieved and implemented in typical as well as atypical brains (e.g., following stroke or traumatic injury).
Much of the work in the lab develops within the aforementioned areas of study. For example, we recently used MEG and covert speech to examine the spatiotemporal neural dynamics of speech during planning and (internal) production and to provide evidence for distinct sensory and motor neural codes.Â
Oftentimes, however, we must proceed at the intersection of these cognitive functions. For example, a series of studies examines speech auditory-motor synchronization as a neural mechanism for language statistical learning.
Starting in the fall, a novel and exciting new line of inquiry will investigate how different types of information (e.g., speech prosody, phonology, syntax, timing) are processed, interact, and are integrated during speech perception.