Close Menu
    Facebook X (Twitter) Instagram
    SciTechDaily
    • Biology
    • Chemistry
    • Earth
    • Health
    • Physics
    • Science
    • Space
    • Technology
    Facebook X (Twitter) Pinterest YouTube RSS
    SciTechDaily
    Home»Technology»More Like Us Than We Realize: ChatGPT Gets Caught Thinking Like a Human
    Technology

    More Like Us Than We Realize: ChatGPT Gets Caught Thinking Like a Human

    By Institute for Operations Research and the Management SciencesApril 10, 20252 Comments4 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn WhatsApp Email Reddit
    Share
    Facebook Twitter LinkedIn Pinterest Telegram Email Reddit
    Artificial Intelligence Data AI Problem Solving
    A new study finds that ChatGPT mirrors human decision-making biases in nearly half of tested scenarios, including overconfidence and the gambler’s fallacy. While it excels at logic and math, the AI struggles with judgment calls, revealing it may think more like us than we realize.

    Groundbreaking research reveals that AI doesn’t just process data, it also makes the same judgment errors as humans.

    Can we really trust AI to make better decisions than humans? According to a recent study, the answer is: not always. Researchers found that OpenAI’s ChatGPT, one of the most advanced and widely used AI models, sometimes makes the same decision-making errors as humans. In certain scenarios, it exhibits familiar cognitive biases, such as overconfidence and the hot-hand (or gambler’s) fallacy. Yet in other cases, it behaves in ways that differ significantly from human reasoning, for example, it tends not to fall for base-rate neglect or the sunk-cost fallacy.

    The study, published in the INFORMS journal Manufacturing & Service Operations Management, suggests that ChatGPT doesn’t simply analyze data, it mirrors aspects of human thinking, including mental shortcuts and systematic errors. These patterns of bias appear relatively consistent across various business contexts, although they may shift as newer versions of the AI are developed.

    AI: A Smart Assistant with Human-Like Flaws

    The study put ChatGPT through 18 different bias tests. The results?

    • AI falls into human decision traps – ChatGPT showed biases like overconfidence or ambiguity aversion, and conjunction fallacy (aka as the “Linda problem”), in nearly half the tests.
    • AI is great at math, but struggles with judgment calls – It excels at logical and probability-based problems but stumbles when decisions require subjective reasoning.
    • Bias isn’t going away – Although the newer GPT-4 model is more analytically accurate than its predecessor, it sometimes displayed stronger biases in judgment-based tasks.

    Why This Matters

    From job hiring to loan approvals, AI is already shaping major decisions in business and government. But if AI mimics human biases, could it be reinforcing bad decisions instead of fixing them?

    “As AI learns from human data, it may also think like a human – biases and all,” says Yang Chen, lead author and assistant professor at Western University. “Our research shows when AI is used to make judgment calls, it sometimes employs the same mental shortcuts as people.”

    The study found that ChatGPT tends to:

    • Play it safe – AI avoids risk, even when riskier choices might yield better results.
    • Overestimate itself – ChatGPT assumes it’s more accurate than it really is.
    • Seek confirmation – AI favors information that supports existing assumptions, rather than challenging them.
    • Avoid ambiguity – AI prefers alternatives with more certain information and less ambiguity.

    “When a decision has a clear right answer, AI nails it – it is better at finding the right formula than most people are,” says Anton Ovchinnikov of Queen’s University. “But when judgment is involved, AI may fall into the same cognitive traps as people.”

    So, Can We Trust AI to Make Big Decisions?

    With governments worldwide working on AI regulations, the study raises an urgent question: Should we rely on AI to make important calls when it can be just as biased as humans?

    “AI isn’t a neutral referee,” says Samuel Kirshner of UNSW Business School. “If left unchecked, it might not fix decision-making problems – it could actually make them worse.”

    The researchers say that’s why businesses and policymakers need to monitor AI’s decisions as closely as they would a human decision-maker.

    “AI should be treated like an employee who makes important decisions – it needs oversight and ethical guidelines,” says Meena Andiappan of McMaster University. “Otherwise, we risk automating flawed thinking instead of improving it.”

    What’s Next?

    The study’s authors recommend regular audits of AI-driven decisions and refining AI systems to reduce biases. With AI’s influence growing, making sure it improves decision-making – rather than just replicating human flaws – will be key.

    “The evolution from GPT-3.5 to 4.0 suggests the latest models are becoming more human in some areas, yet less human but more accurate in others,” says Tracy Jenkin of Queen’s University. “Managers must evaluate how different models perform on their decision-making use cases and regularly re-evaluate to avoid surprises. Some use cases will need significant model refinement.”

    Reference: “A Manager and an AI Walk into a Bar: Does ChatGPT Make Biased Decisions Like We Do?” by Yang Chen, Samuel N. Kirshner, Anton Ovchinnikov, Meena Andiappan and Tracy Jenkin, 31 January 2025, Manufacturing & Service Operations Management.
    DOI: 10.1287/msom.2023.0279

    Never miss a breakthrough: Join the SciTechDaily newsletter.
    Follow us on Google and Google News.

    Artificial Intelligence Computer Science
    Share. Facebook Twitter Pinterest LinkedIn Email Reddit

    Related Articles

    New General-Purpose Technique Sheds Light on Inner Workings of Neural Nets

    “Data Science Machine” Replaces Human Intuition with Algorithms

    AI Framework Predicts Better Patient Health Care and Reduces Cost

    Algorithm Analyzes Information From Medical Images to Identify Disease

    Halide, A New and Improved Programming Language for Image Processing Software

    New Algorithm Enables Wi-Fi Connected Vehicles to Share Data

    Algorithm Enables Robots to Learn and Adapt to Help Complete Tasks

    New Approach Uses Mathematics to Improve Automated Security Monitoring

    Mathematical Framework Formalizes Loop Perforation Technique

    2 Comments

    1. Old Vet on April 10, 2025 11:35 am

      Isn’t it comforting to know that with all of our inherent flaws we’re still better than a machine, thank you GOD.

      Reply
    2. Robert Welch on April 11, 2025 9:46 am

      When A.I. spends hours scrolling thru screen after screen on the ‘net, then it’ll be Human.

      Reply
    Leave A Reply Cancel Reply

    • Facebook
    • Twitter
    • Pinterest
    • YouTube

    Don't Miss a Discovery

    Subscribe for the Latest in Science & Tech!

    Trending News

    Scientists Discover 132-Million-Year-Old Dinosaur Tracks on South Africa’s Coast

    Scientists Uncover the Secret Ingredient Behind the Spark That May Have Started Life on Earth

    Physicists Observe Matter in Two Places at Once in Mind-Bending Quantum Experiment

    Stanford Scientists Discover Hidden Brain Circuit That Fuels Chronic Pain

    New Study Reveals Why Ozempic Works Better for Some People Than Others

    Climate Change Is Altering a Key Greenhouse Gas in a Way Scientists Didn’t Expect

    New Study Suggests Gravitational Waves May Have Created Dark Matter

    Scientists Discover Why the Brain Gets Stuck in Schizophrenia

    Follow SciTechDaily
    • Facebook
    • Twitter
    • YouTube
    • Pinterest
    • Newsletter
    • RSS
    SciTech News
    • Biology News
    • Chemistry News
    • Earth News
    • Health News
    • Physics News
    • Science News
    • Space News
    • Technology News
    Recent Posts
    • The Ideal Temperature for Storing Mangoes Isn’t What You Think
    • Groundbreaking Enzyme Atlas Rewrites Decades of Biology Research
    • New “Nanozyme Hypothesis” Could Rewrite the Story of Life’s Origins
    • Anatomy Isn’t Finished: The Human Body Still Holds Secrets
    • Researchers Discover Long-Lost Words of Ancient Greek Philosopher After 2,000 Years
    Copyright © 1998 - 2026 SciTechDaily. All Rights Reserved.
    • Science News
    • About
    • Contact
    • Editorial Board
    • Privacy Policy
    • Terms of Use

    Type above and press Enter to search. Press Esc to cancel.