Be on the lookout for a new paper to appear in Cognition. The title is “Hearing is believing: Lexically guided perceptual learning is graded to reflect the quantity of evidence in speech input.” This work was conducted with Shawn Cummings, a Ph.D. trainee in the SLaP Lab. Data/script and preprint are available on the Open Science Framework. Bravo, Shawn!
Paper accepted at Cognition
Be on the lookout for a new paper to appear in Cognition. The title is “Do individual differences in lexical reliance reflect states or traits?” This work was conducted with Nikole Giovannone, a Ph.D. trainee in the SLaP Lab. Data/script and preprint are available on the Open Science Framework. Bravo, Nikole!
Paper accepted at Journal of the Acoustical Society of America
Be on the lookout for a new paper to appear in Journal of the Acoustical Society of America. The title is “Re-examining the left ear, right hemisphere processing advantage for phonetic cues to talker identification.” This work was conducted with Lee Drown (a Ph.D. trainee in the lab), Betsy Philip (an undergraduate Honors student in the lab), and Dr. Alex Francis at Purdue. Data/script and preprint are available on the Open Science Framework.
Paper accepted at Journal of Speech, Language, and Hearing Research
Be on the lookout for a new paper to appear in the Journal of Speech, Language, and Hearing Research. The title is “Role of sleep consolidation on perceptual learning of noise-vocoded speech.” This work was conducted with SLaP Lab trainees Julia Drouin and Victoria Zysk, and friend of the lab Emily Myers. Data/script and preprint are available on the Open Science Framework.
Paper accepted at Attention, Perception, & Psychophysics
Be on the lookout for a new paper to appear in Attention, Perception, & Psychophysics. The title is “Perceptual learning of multiple talkers: Determinants, characteristics, and limitations.” This study was Shawn’s first-year project in the lab, way to go, Shawn! Data/script and preprint are available on the Open Science Framework.
Paper accepted at Journal of the Acoustical Society of America
Be on the lookout for a new paper to appear in Journal of the Acoustical Society of America. The title is “Many tasks, same outcome: Role of training task on perceptual learning and maintenance of noise-vocoded speech.” This work was conducted with Julia Drouin while she was a Ph.D. student in the lab. Data/script and preprint are available on the Open Science Framework.
Dr. Theodore presents at LabPhon 18
Dr. Theodore presented at the LabPhon 18 meeting. She was the plenary speaker for the thematic session, “Changes in space: Online experimentation.” A copy of the slide deck and talk recording is provided below.
Paper accepted at Journal of the Acoustical Society of America
Be on the lookout for a new paper to appear in Journal of the Acoustical Society of America. The title is “Context effects in perception of vowels differentiated by F1 are not influenced by variability in talkers’ mean F1 or F3.” This work was led by Dr. Christian Stilp’s lab at the University of Louisville. Data/script and preprint are available on the Open Science Framework.
Dr. Theodore receives NSF award
Dr. Theodore has received a 3-year grant from the Division of Behavioral and Cognitive Sciences of the National Science Foundation titled “Determinants of perceptual learning for speech perception.” The activities will be completed in collaboration with Dr. Matt Winn at the University of Minnesota. The public abstract is shown below.
Abstract: One of the great mysteries in the cognitive sciences is how humans understand speech despite extreme variation in the actual acoustics of the sounds produced. Speakers rarely produce identical versions of the same speech sound, because everyone has a personal way of talking depending on their dialect, gender, and age. Other factors, such as fatigue or excitement, might change moment-to-moment. Although listeners are usually able to adapt to these differences, research has yet to discover exactly how this adaptation works. This project establishes a new tool for creating experimental stimuli that can generate the wide variation that occurs in natural speech and uses this tool to examine how listeners adapt to talker differences. The experiments test the ways that listeners incorporate new evidence from a talker’s voice, how they use existing knowledge about the speech sounds in their language, how they learn what to ignore, and how individuals may differ in their ability to adapt to a talker. The project has potentially broad societal benefits because it will provide foundational knowledge on how people communicate with each other and provide hands-on education for undergraduate and graduate trainees who will develop skills for experimental design, computer programming, data analysis, and science communication. The project will also create tools for other researchers, including an open-access stimulus corpus and brief lecture videos and assignments that can be used to teach students and others about the physics of sound, thus promoting scientific literacy among the public. All materials associated with this project will be made publicly available.
This interdisciplinary project unites psycholinguistic experimentation, signal processing, and computational modeling to test three key parts of a Bayesian belief-updating model of speech adaptation: (1) beliefs reflect listeners’ knowledge of the relationship between cue distributions and phonetic categories, (2) adaptation reflects integration of observed evidence with prior beliefs that can be driven by both unsupervised and supervised learning signals, and (3) belief-updating is context-specific (e.g., conditioned on talker). The project will establish a novel method for creating fricative continua, create an extensive stimulus corpus, and compare learning between the stimulus corpus and natural stimuli. The novel method of stimulus creation will be used to generate sounds that independently manipulate frequency scaling (reflecting vocal tract size) and gain properties (reflecting specific speech articulations) of fricative cues to discover which specific cues guide learning and how that information generalizes (or not) across talkers. The experiments will also examine how listeners integrate unsupervised and lexically supervised learning signals over time, providing insight into the influence of prior exposure on lexically guided learning and individual differences in learning. These are important innovations because they address two weaknesses in the existing perceptual learning literature, including the practice of superimposing two sounds to create ambiguous input that might be unnatural and the failure to measure incremental change in perception over time. Theoretical benefits include a strong test of the Bayesian belief-updating model of speech adaptation, identification of specific cues that drive perceptual learning, and unpacking how perceptual learning integrates unsupervised and lexically supervised signals.
Dr. Theodore speaks at Acoustics in Focus
This repository contains materials associated with presentation 2pSCa1 at the 2021 Acoustics in Focus meeting (Rachel M. Theodore, “Conducting speech perception experiments remotely: Some tools, successes, and challenges”).
Don’t hesitate to reach out with questions and/or if you’d like access to any of our experiments on Gorilla, preprints/postprints, or data (rachel.theodore@uconn.edu).