Close Menu
    Facebook X (Twitter) Instagram
    SciTechDaily
    • Biology
    • Chemistry
    • Earth
    • Health
    • Physics
    • Science
    • Space
    • Technology
    Facebook X (Twitter) Pinterest YouTube RSS
    SciTechDaily
    Home»Technology»Machine-Learning System Replicates Human Auditory Behavior, Predicts Brain Responses
    Technology

    Machine-Learning System Replicates Human Auditory Behavior, Predicts Brain Responses

    By Anne Trafton, Massachusetts Institute of TechnologyApril 24, 2018No Comments7 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn WhatsApp Email Reddit
    Share
    Facebook Twitter LinkedIn Pinterest Telegram Email Reddit
    Machine-Learning System Processes Sounds Like Humans
    MIT researchers have created the first model that can replicate human performance on auditory tasks such as identifying a musical genre. Credit: Chelsea Turner/MIT

    Using a machine-learning system known as a deep neural network, MIT researchers have created the first model that can replicate human performance on auditory tasks such as identifying a musical genre.

    This model, which consists of many layers of information-processing units that can be trained on huge volumes of data to perform specific tasks, was used by the researchers to shed light on how the human brain may be performing the same tasks.

    “What these models give us, for the first time, is machine systems that can perform sensory tasks that matter to humans and that do so at human levels,” says Josh McDermott, the Frederick A. and Carole J. Middleton Assistant Professor of Neuroscience in the Department of Brain and Cognitive Sciences at MIT and the senior author of the study. “Historically, this type of sensory processing has been difficult to understand, in part because we haven’t really had a very clear theoretical foundation and a good way to develop models of what might be going on.”

    The study, which appears in the April 19 issue of Neuron, also offers evidence that the human auditory cortex is arranged in a hierarchical organization, much like the visual cortex. In this type of arrangement, sensory information passes through successive stages of processing, with basic information processed earlier and more advanced features such as word meaning extracted in later stages.

    MIT graduate student Alexander Kell and Stanford University Assistant Professor Daniel Yamins are the paper’s lead authors. Other authors are former MIT visiting student Erica Shook and former MIT postdoc Sam Norman-Haignere.

    Modeling the brain

    When deep neural networks were first developed in the 1980s, neuroscientists hoped that such systems could be used to model the human brain. However, computers from that era were not powerful enough to build models large enough to perform real-world tasks such as object recognition or speech recognition.

    Over the past five years, advances in computing power and neural network technology have made it possible to use neural networks to perform difficult real-world tasks, and they have become the standard approach in many engineering applications. In parallel, some neuroscientists have revisited the possibility that these systems might be used to model the human brain.

    “That’s been an exciting opportunity for neuroscience, in that we can actually create systems that can do some of the things people can do, and we can then interrogate the models and compare them to the brain,” Kell says.

    The MIT researchers trained their neural network to perform two auditory tasks, one involving speech and the other involving music. For the speech task, the researchers gave the model thousands of two-second recordings of a person talking. The task was to identify the word in the middle of the clip. For the music task, the model was asked to identify the genre of a two-second clip of music. Each clip also included background noise to make the task more realistic (and more difficult).

    After many thousands of examples, the model learned to perform the task just as accurately as a human listener.

    “The idea is over time the model gets better and better at the task,” Kell says. “The hope is that it’s learning something general, so if you present a new sound that the model has never heard before, it will do well, and in practice that is often the case.”

    The model also tended to make mistakes on the same clips that humans made the most mistakes on.

    The processing units that make up a neural network can be combined in a variety of ways, forming different architectures that affect the performance of the model.

    The MIT team discovered that the best model for these two tasks was one that divided the processing into two sets of stages. The first set of stages was shared between tasks, but after that, it split into two branches for further analysis — one branch for the speech task, and one for the musical genre task.

    Evidence for hierarchy

    The researchers then used their model to explore a longstanding question about the structure of the auditory cortex: whether it is organized hierarchically.

    In a hierarchical system, a series of brain regions performs different types of computation on sensory information as it flows through the system. It has been well documented that the visual cortex has this type of organization. Earlier regions, known as the primary visual cortex, respond to simple features such as color or orientation. Later stages enable more complex tasks such as object recognition.

    However, it has been difficult to test whether this type of organization also exists in the auditory cortex, in part because there haven’t been good models that can replicate human auditory behavior.

    “We thought that if we could construct a model that could do some of the same things that people do, we might then be able to compare different stages of the model to different parts of the brain and get some evidence for whether those parts of the brain might be hierarchically organized,” McDermott says.

    The researchers found that in their model, basic features of sound such as frequency are easier to extract in the early stages. As information is processed and moves farther along the network, it becomes harder to extract frequency but easier to extract higher-level information such as words.

    To see if the model stages might replicate how the human auditory cortex processes sound information, the researchers used functional magnetic resonance imaging (fMRI) to measure different regions of auditory cortex as the brain processes real-world sounds. They then compared the brain responses to the responses in the model when it processed the same sounds.

    They found that the middle stages of the model corresponded best to activity in the primary auditory cortex, and later stages corresponded best to activity outside of the primary cortex. This provides evidence that the auditory cortex might be arranged in a hierarchical fashion, similar to the visual cortex, the researchers say.

    “What we see very clearly is a distinction between primary auditory cortex and everything else,” McDermott says.

    Alex Huth, an assistant professor of neuroscience and computer science at the University of Texas at Austin, says the paper is exciting in part because it offers convincing evidence that the early part of the auditory cortex performs generic sound processing while the higher auditory cortex performs more specialized tasks.

    “This is one of the ongoing mysteries in auditory neuroscience: What distinguishes the early auditory cortex from the higher auditory cortex? This is the first paper I’ve seen that has a computational hypothesis for that,” says Huth, who was not involved in the research.

    The authors now plan to develop models that can perform other types of auditory tasks, such as determining the location from which a particular sound came, to explore whether these tasks can be done by the pathways identified in this model or if they require separate pathways, which could then be investigated in the brain.

    The research was funded by the National Institutes of Health, the National Science Foundation, a Department of Energy Computational Science Graduate Fellowship, and a McDonnell Scholar Award.

    Reference: “A Task-Optimized Neural Network Replicates Human Auditory Behavior, Predicts Brain Responses, and Reveals a Cortical Processing Hierarchy” by Alexander J.E. Kell, Daniel L.K. Yamins, Erica N. Shook, Sam V. Norman-Haignere and Josh H. McDermott, 19 April 2018, Neuron.
    DOI: 10.1016/j.neuron.2018.03.044

    Never miss a breakthrough: Join the SciTechDaily newsletter.
    Follow us on Google and Google News.

    Computer Science Machine Learning MIT Nanoscience Nanotechnology Neuroscience
    Share. Facebook Twitter Pinterest LinkedIn Email Reddit

    Related Articles

    From Pixels to Paradigms: MIT’s Synthetic Leap in AI Training

    Innovative Approach to Controlling Magnetism Opens Route to Ultra-Low-Power Microchips

    On-Chip Optical Filter Provides Cheaper Alternative to “Broadband” Filters

    New Self-Assembly Technique Could Lead to Smaller Microchip Patterns

    New Technique Enlarges Tissue Samples, Making Them Easier to Image

    Newly Designed Nanoparticles Can Deliver Three Cancer Drugs at a Time

    2D Molybdenum Disulfide Shows Potential as Platform for Electronic Devices

    Self-Assembling Polymer Molecules Create Complex Microchip Structures

    All-Carbon Photovoltaic Cell to Harness Infrared Light

    Leave A Reply Cancel Reply

    • Facebook
    • Twitter
    • Pinterest
    • YouTube

    Don't Miss a Discovery

    Subscribe for the Latest in Science & Tech!

    Trending News

    Popular Sugar-Free Sweetener Linked to Liver Disease, Study Warns

    What Is Hantavirus? The Deadly Disease Raising Alarm Worldwide

    Scientists Just Discovered How the Universe Builds Monster Black Holes

    Scientists Unveil New Treatment Strategy That Could Outsmart Cancer

    A Simple Vitamin May Hold the Key to Treating Rare Genetic Diseases

    Scientists Think the Real Fountain of Youth May Be Hiding in Your Gut

    Ravens Don’t Follow Wolves, They Predict Them

    This Common Knee Surgery May Be Doing More Harm Than Good

    Follow SciTechDaily
    • Facebook
    • Twitter
    • YouTube
    • Pinterest
    • Newsletter
    • RSS
    SciTech News
    • Biology News
    • Chemistry News
    • Earth News
    • Health News
    • Physics News
    • Science News
    • Space News
    • Technology News
    Recent Posts
    • This Mars Rock Refused To Let Go of NASA’s Curiosity Rover
    • James Webb Telescope Reveals the Universe’s Hidden Cosmic Web in Stunning Detail
    • Scientists Build a Living AI Device Using Real Brain Cells
    • Why Are So Many New Fathers Dying? Scientists Say the U.S. Has a Dangerous Blind Spot
    • Scientists Identify Simple Supplement That Greatly Reduces Alzheimer’s Damage
    Copyright © 1998 - 2026 SciTechDaily. All Rights Reserved.
    • Science News
    • About
    • Contact
    • Editorial Board
    • Privacy Policy
    • Terms of Use

    Type above and press Enter to search. Press Esc to cancel.