Close Menu
    Facebook X (Twitter) Instagram
    SciTechDaily
    • Biology
    • Chemistry
    • Earth
    • Health
    • Physics
    • Science
    • Space
    • Technology
    Facebook X (Twitter) Pinterest YouTube RSS
    SciTechDaily
    Home»Technology»MIT and IBM Develop New Tool To Help Choose the Right Method for Evaluating AI Models
    Technology

    MIT and IBM Develop New Tool To Help Choose the Right Method for Evaluating AI Models

    By Adam Zewe, Massachusetts Institute of TechnologyMay 31, 20231 Comment6 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn WhatsApp Email Reddit
    Share
    Facebook Twitter LinkedIn Pinterest Telegram Email Reddit
    Futuristic Cyber Data Storage
    MIT and IBM researchers have created ‘saliency cards’ to aid in the selection of appropriate saliency methods for machine learning models. These cards detail a method’s functionalities and performance characteristics, facilitating users in making informed choices, ultimately improving understanding of their models’ behavior.

    Selecting the right method gives users a more accurate picture of how their model is behaving, so they are better equipped to correctly interpret its predictions.

    When machine-learning models are deployed in real-world situations, perhaps to flag potential diseases in X-rays for a radiologist to review, human users need to know when to trust the model’s predictions.

    But machine-learning models are so large and complex that even the scientists who design them don’t understand exactly how the models make predictions. So, they create techniques known as saliency methods that seek to explain model behavior.

    With new methods being released all the time, researchers from MIT and IBM Research created a tool to help users choose the best saliency method for their particular task. They developed saliency cards, which provide standardized documentation of how a method operates, including its strengths and weaknesses and explanations to help users interpret it correctly.

    They hope that, armed with this information, users can deliberately select an appropriate saliency method for both the type of machine-learning model they are using and the task that model is performing, explains co-lead author Angie Boggust, a graduate student in electrical engineering and computer science at MIT and member of the Visualization Group of the MIT Computer Science and Artificial Intelligence Laboratory (CSAIL).

    Saliency Cards
    New “saliency cards” provide concise summaries of machine-learning saliency methods in terms of 10 user-focused attributes. Credit: iStock

    Interviews with AI researchers and experts from other fields revealed that the cards help people quickly conduct a side-by-side comparison of different methods and pick a task-appropriate technique. Choosing the right method gives users a more accurate picture of how their model is behaving, so they are better equipped to correctly interpret its predictions.

    “Saliency cards are designed to give a quick, glanceable summary of a saliency method and also break it down into the most critical, human-centric attributes. They are really designed for everyone, from machine-learning researchers to lay users who are trying to understand which method to use and choose one for the first time,” says Boggust.

    Joining Boggust on the paper are co-lead author Harini Suresh, an MIT postdoc; Hendrik Strobelt, a senior research scientist at IBM Research; John Guttag, the Dugald C. Jackson Professor of Computer Science and Electrical Engineering at MIT; and senior author Arvind Satyanarayan, associate professor of computer science at MIT who leads the Visualization Group in CSAIL. The research will be presented at the ACM Conference on Fairness, Accountability, and Transparency.

    Picking the Right Method

    The researchers have previously evaluated saliency methods using the notion of faithfulness. In this context, faithfulness captures how accurately a method reflects a model’s decision-making process.

    But faithfulness is not black-and-white, Boggust explains. A method might perform well under one test of faithfulness, but fail another. With so many saliency methods, and so many possible evaluations, users often settle on a method because it is popular or a colleague has used it.

    However, picking the “wrong” method can have serious consequences. For instance, one saliency method, known as integrated gradients, compares the importance of features in an image to a meaningless baseline. The features with the largest importance over the baseline are most meaningful to the model’s prediction. This method typically uses all 0s as the baseline, but if applied to images, all 0s equates to the color black.

    “It will tell you that any black pixels in your image aren’t important, even if they are, because they are identical to that meaningless baseline. This could be a big deal if you are looking at X-rays since black could be meaningful to clinicians,” says Boggust.

    Saliency cards can help users avoid these types of problems by summarizing how a saliency method works in terms of 10 user-focused attributes. The attributes capture the way saliency is calculated, the relationship between the saliency method and the model, and how a user perceives its outputs.

    For example, one attribute is hyperparameter dependence, which measures how sensitive that saliency method is to user-specified parameters. A saliency card for integrated gradients would describe its parameters and how they affect its performance. With the card, a user could quickly see that the default parameters — a baseline of all 0s — might generate misleading results when evaluating X-rays.

    The cards could also be useful for scientists by exposing gaps in the research space. For instance, the MIT researchers were unable to identify a saliency method that was computationally efficient, but could also be applied to any machine-learning model.

    “Can we fill that gap? Is there a saliency method that can do both things? Or maybe these two ideas are theoretically in conflict with one another,” Boggust says.

    Showing Their Cards

    Once they had created several cards, the team conducted a user study with eight domain experts, from computer scientists to a radiologist who was unfamiliar with machine learning. During interviews, all participants said the concise descriptions helped them prioritize attributes and compare methods. And even though he was unfamiliar with machine learning, the radiologist was able to understand the cards and use them to take part in the process of choosing a saliency method, Boggust says.

    The interviews also revealed a few surprises. Researchers often expect that clinicians want a method that is sharp, meaning it focuses on a particular object in a medical image. But the clinician in this study actually preferred some noise in medical images to help them attenuate uncertainty.

    “As we broke it down into these different attributes and asked people, not a single person had the same priorities as anyone else in the study, even when they were in the same role,” she says.

    Moving forward, the researchers want to explore some of the more under-evaluated attributes and perhaps design task-specific saliency methods. They also want to develop a better understanding of how people perceive saliency method outputs, which could lead to better visualizations. In addition, they are hosting their work on a public repository so others can provide feedback that will drive future work, Boggust says.

    “We are really hopeful that these will be living documents that grow as new saliency methods and evaluations are developed. In the end, this is really just the start of a larger conversation around what the attributes of a saliency method are and how those play into different tasks,” she says.

    Reference: “Saliency Cards: A Framework to Characterize and Compare Saliency Methods” by Angie Boggust, Harini Suresh, Hendrik Strobelt, John Guttag and Arvind Satyanarayan.
    PDF

    The research was supported, in part, by the MIT-IBM Watson AI Lab, the U.S. Air Force Research Laboratory, and the U.S. Air Force Artificial Intelligence Accelerator.

    Never miss a breakthrough: Join the SciTechDaily newsletter.
    Follow us on Google and Google News.

    Artificial Intelligence CSAIL Electrical Engineering IBM Machine Learning MIT
    Share. Facebook Twitter Pinterest LinkedIn Email Reddit

    Related Articles

    MIT’s AI Learns Molecular Language for Rapid Material Development and Drug Discovery

    Mastering Uncertainty: An Effective Approach to Training Machines for Real-World Situations

    CausalSim: MIT’s New Tool for Accurately Simulating Complex Systems

    When It Comes to AI, Can We Ditch the Datasets? Using Synthetic Data for Training Machine-Learning Models

    Demystifying Machine-Learning Systems: Automatically Describing Neural Network Components in Natural Language

    Breakthrough AI Technique Enables Real-Time Rendering of Scenes in 3D From 2D Images

    New Artificial Intelligence System Enables Machines That See the World More Like Humans Do

    New Machine-Learning System Gives Robots Social Skills

    Showing Robots How to Do Your Chores – Automated Robots That Learn Just by Watching

    1 Comment

    1. Ralph Johnson on December 15, 2023 8:28 am

      They should develop a algorithm that would present the pros and cons of each choice and then present a method of determining the most positive method for a reasonable outcome.

      Reply
    Leave A Reply Cancel Reply

    • Facebook
    • Twitter
    • Pinterest
    • YouTube

    Don't Miss a Discovery

    Subscribe for the Latest in Science & Tech!

    Trending News

    Scientists May Have Found the Key to Jupiter and Saturn’s Moon Mystery

    Scientists Uncover Brain Changes That Link Pain to Depression

    Saunas May Do More Than Raise Body Temperature – They Activate Your Immune System

    Exercise in a Pill? Metformin Shows Surprising Effects in Cancer Patients

    Hidden Oceans of Magma Could Be Protecting Alien Life

    New Study Challenges Alzheimer’s Theories: It’s Not Just About Plaques

    Artificial Sweeteners May Harm Future Generations, Study Suggests

    Splashdown! NASA Artemis II Returns From Record-Breaking Moon Mission

    Follow SciTechDaily
    • Facebook
    • Twitter
    • YouTube
    • Pinterest
    • Newsletter
    • RSS
    SciTech News
    • Biology News
    • Chemistry News
    • Earth News
    • Health News
    • Physics News
    • Science News
    • Space News
    • Technology News
    Recent Posts
    • Scientists Discover Game-Changing New Way To Treat High Cholesterol
    • Breakthrough Drug Delays Rheumatoid Arthritis for Years After Treatment Ends
    • This Small Change to Your Exercise Routine Could Be the Secret to Living Longer
    • Physicists Discover a Strange New Kind of One-Dimensional Particle
    • Scientists Discover Unexpected Climate Benefit Hidden in Forest Soils
    Copyright © 1998 - 2026 SciTechDaily. All Rights Reserved.
    • Science News
    • About
    • Contact
    • Editorial Board
    • Privacy Policy
    • Terms of Use

    Type above and press Enter to search. Press Esc to cancel.