Close Menu
    Facebook X (Twitter) Instagram
    SciTechDaily
    • Biology
    • Chemistry
    • Earth
    • Health
    • Physics
    • Science
    • Space
    • Technology
    Facebook X (Twitter) Pinterest YouTube RSS
    SciTechDaily
    Home»Technology»Machine Learning at Speed: Optimization Code Increases Performance by 5x
    Technology

    Machine Learning at Speed: Optimization Code Increases Performance by 5x

    By King Abdullah University of Science & Technology (KAUST)April 12, 20213 Comments3 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn WhatsApp Email Reddit
    Share
    Facebook Twitter LinkedIn Pinterest Telegram Email Reddit
    Machine Learning on Parallelized Computing Systems
    Technology developed through a KAUST-led collaboration with Intel, Microsoft and the University of Washington can dramatically increase the speed of machine learning on parallelized computing systems. Credit: © 2021 KAUST; Anastasia Serin

    Optimizing network communication accelerates training in large-scale machine-learning models.

    Inserting lightweight optimization code in high-speed network devices has enabled a KAUST-led collaboration to increase the speed of machine learning on parallelized computing systems five-fold.

    This “in-network aggregation” technology, developed with researchers and systems architects at Intel, Microsoft and the University of Washington, can provide dramatic speed improvements using readily available programmable network hardware.

    The fundamental benefit of artificial intelligence (AI) that gives it so much power to “understand” and interact with the world is the machine-learning step, in which the model is trained using large sets of labeled training data. The more data the AI is trained on, the better the model is likely to perform when exposed to new inputs.

    The recent burst of AI applications is largely due to better machine learning and the use of larger models and more diverse datasets. Performing the machine-learning computations, however, is an enormously taxing task that increasingly relies on large arrays of computers running the learning algorithm in parallel.

    “How to train deep-learning models at a large scale is a very challenging problem,” says Marco Canini from the KAUST research team. “The AI models can consist of billions of parameters, and we can use hundreds of processors that need to work efficiently in parallel. In such systems, communication among processors during incremental model updates easily becomes a major performance bottleneck.”

    Programmable Networking Hardware to the Rescue

    The team found a potential solution in new network technology developed by Barefoot Networks, a division of Intel.

    “We use Barefoot Networks’ new programmable dataplane networking hardware to offload part of the work performed during distributed machine-learning training,” explains Amedeo Sapio, a KAUST alumnus who has since joined the Barefoot Networks team at Intel. “Using this new programmable networking hardware, rather than just the network, to move data means that we can perform computations along the network paths.”

    The key innovation of the team’s SwitchML platform is to allow the network hardware to perform the data aggregation task at each synchronization step during the model update phase of the machine-learning process. Not only does this offload part of the computational load, it also significantly reduces the amount of data transmission.

    Balancing Simplicity and Flexibility for Speed Gains

    “Although the programmable switch dataplane can do operations very quickly, the operations it can do are limited,” says Canini. “So our solution had to be simple enough for the hardware and yet flexible enough to solve challenges such as limited onboard memory capacity. SwitchML addresses this challenge by co-designing the communication network and the distributed training algorithm, achieving an acceleration of up to 5.5 times compared to the state-of-the-art approach.” 

    Reference: “Scaling Distributed Machine Learning with In-Network Aggregation” by Amedeo Sapio, Marco Canini, Chen-Yu Ho, Jacob Nelson, Panos Kalnis, Changhoon Kim, Arvind Krishnamurthy, Masoud Moshref, Dan Ports and Peter Richtarik, April 2021, The 18th USENIX Symposium on Networked Systems Design and Implementation (NSDI ’21).
    Link

    Never miss a breakthrough: Join the SciTechDaily newsletter.
    Follow us on Google and Google News.

    Algorithm Artificial Intelligence Computer Science KAUST Machine Learning
    Share. Facebook Twitter Pinterest LinkedIn Email Reddit

    Related Articles

    Scientists Create a “Periodic Table” for Artificial Intelligence

    “AI Is Not Intelligent at All” – Expert Warns of Worldwide Threat to Human Dignity

    Predicting Chaos With AI: The New Frontier in Autonomous Control

    When Algorithms Deliver: The AI Revolution in Logistics

    CausalSim: MIT’s New Tool for Accurately Simulating Complex Systems

    Teaching Physics to AI Can Allow It To Make New Discoveries All on Its Own

    ‘Deep Learning’ Algorithm Brings New Tools to Astronomy

    Halide, A New and Improved Programming Language for Image Processing Software

    New Algorithm Enables Wi-Fi Connected Vehicles to Share Data

    3 Comments

    1. Pedro on August 22, 2021 12:55 am

      optimizing the code in so much is very cool but also optimizing the whole workflow is no less important, so if you are interested in this pay attention to https://www.akveo.com/templates/fleet-management-software it will make your life easier

      Reply
    2. kamir bouchareb st on February 16, 2025 11:51 pm

      thank you

      Reply
    3. Teknik Telekomunikasi on February 25, 2025 1:10 am

      Visit us Magister Akuntansi
      Explain the concept of computer vision.

      Reply
    Leave A Reply Cancel Reply

    • Facebook
    • Twitter
    • Pinterest
    • YouTube

    Don't Miss a Discovery

    Subscribe for the Latest in Science & Tech!

    Trending News

    Artificial Sweeteners May Harm Future Generations, Study Suggests

    Splashdown! NASA Artemis II Returns From Record-Breaking Moon Mission

    What If Consciousness Exists Beyond Your Brain

    Scientists Finally Crack the 100-Million-Year Evolutionary Mystery of Squid and Cuttlefish

    Beyond “Safe Levels”: Study Challenges What We Know About Pesticides and Cancer

    Researchers Have Found a Dietary Compound That Increases Longevity

    Scientists Baffled by Bizarre “Living Fossil” From 275 Million Years Ago

    Your IQ at 23 Could Predict Your Wealth at 27, Study Finds

    Follow SciTechDaily
    • Facebook
    • Twitter
    • YouTube
    • Pinterest
    • Newsletter
    • RSS
    SciTech News
    • Biology News
    • Chemistry News
    • Earth News
    • Health News
    • Physics News
    • Science News
    • Space News
    • Technology News
    Recent Posts
    • Scientists Uncover Hidden Force Powering Yellowstone’s Supervolcano
    • This Metal Melts in Your Hand – and Scientists Just Discovered Something Strange
    • Why Losing Too Much Fat Can Be Just As Dangerous as Obesity
    • Beef vs. Chicken: Surprising Results From New Prediabetes Study
    • Alzheimer’s Breakthrough: Scientists Discover Key Protein May Prevent Toxic Protein Clumps in the Brain
    Copyright © 1998 - 2026 SciTechDaily. All Rights Reserved.
    • Science News
    • About
    • Contact
    • Editorial Board
    • Privacy Policy
    • Terms of Use

    Type above and press Enter to search. Press Esc to cancel.