What an infant hears during sleep has an immediate and profound impact on his or her brain activity, potentially shaping language learning later in life, suggests a new University of baby直播app Boulder study of slumbering babies.
鈥淲e found that even while babies sleep, they are still processing information about their acoustic environment, and their brains are using that information to develop pathways for learning,鈥 said lead author Phillip Gilley, PhD, principal investigator of the Neurodynamics Laboratory at the Institute of Cognitive Science (ICS).
Research dating back to the 1970s suggests that newborns can already recognize their mother鈥檚 voice. What has remained unclear, however, is how early, and to what degree infants can distinguish between the rapid-firing sounds鈥攕uch as long or slow vowels or consonants鈥攖hat serve as the building blocks of human language.
The science of sleeping babies
To find out, Gilley and his colleagues enlisted the parents of 24 healthy infants under the age of 5 months to bring their newborns to a lab. Each infant had electroencephalogram (EEG) electrodes attached to their heads, then fell asleep. Researchers tested sleeping infants because they are more still and infants spend up to 80 percent of their time asleep. A speaker nearby played a sequence of repeated sounds鈥斺渁h, ah, ah鈥 or 鈥渂ah, bah, bah鈥濃攊nterspersed with an occasional oddball sound鈥斺渆eh鈥 or 鈥渄ah,鈥 respectively. Meanwhile, EEG measured the child鈥檚 brainwaves.
Afterward, Gilley used an algorithm developed in his laboratory to identify and measure these different brain patterns.
The findings, published in March in the journal BMC Neuroscience, came as a surprise.
When the 鈥渟tandard鈥 sound hummed along鈥攁h, ah, ah鈥攖he infant鈥檚 brainwaves remained primarily in 听a theta, or low frequency, wave. But within a few milliseconds of hearing the oddball sound the brainwave pattern shifted to a complex blend of gamma, beta and theta frequencies, a signal of neurons in different regions of the brain oscillating and harmonizing.
In essence, the brain quickly learned what the expected sound was, anticipated it, and reacted with surprise to a different one. In the process, Gilley notes, new neuronal pathways key to discriminating sounds were likely formed. That鈥檚 important, because knowing how to discriminate between distinct sounds is key to learning speech and language.
鈥淭he most surprising finding here is how quickly these infants鈥 brains are able to make those predictions. Within the span of one test, their brain learns a pattern and begins to respond to it.鈥
The paper is the first of a series the group will be rolling out as part of a five-year, multi-center grant from the National Institute on Disability, Independent Living, and Rehabilitation Research.
The goal is twofold: To use EEG to better understand precisely when and how a child鈥檚 brain begins to process sound; and to aid development of better diagnostic tests.
Helping babies with hearing impairments
Today, while EEG is used to determine鈥攁t the brainstem level鈥攚hether a newborn responds to sound, it can鈥檛 show how well they distinguish between sounds. That can make it hard to tune hearing aids for babies with hearing loss. Studies show 50 percent of hearing aids may not be properly calibrated for young children.
鈥淲ithin the next five years, a clinician could put a few small sensors on a newborn鈥檚 head for less than 30 minutes and detect not only that an infant can hear speech, but also how well they tell the difference between speech sounds,鈥 says Gilley.
What should infants be listening to? It鈥檚 too early to say, Gilley notes, although he admits to preferring ocean sounds and other soothing repetitive sounds for his children. Gilley does recommend that if a child has a hearing impairment, they wear their hearing aids as much as possible.
鈥淥therwise, they may be missing out on essential environmental sounds that help their brain to learn.鈥
Kristin Uhler, an assistant professor at the CU Anschutz School of Medicine, Kaylee Watson, a doctor of audiology candidate in the CU Boulder department of Speech, Language, and Hearing Sciences, and Christine Yoshinaga-Itano, a professor with ICS, co-authored the study.