AI Learns To Think Like Humans: A Game-Changer in Machine Learning

AI Memristor Technolgy Artificial Neural Network Concept

Researchers at Georgia Tech are advancing neural networks to mimic human decision-making by training them to exhibit variability and confidence in their choices, similar to how humans operate, as demonstrated in their study published in Nature Human Behaviour. Their model, RTNet, not only matches human performance in recognizing noisy digits but also applies human-like traits such as confidence and evidence accumulation, enhancing both accuracy and reliability. Credit: SciTechDaily.com

Georgia Tech researchers have developed a neural network, RTNet, that mimics human decision-making processes, including confidence and variability, improving its reliability and accuracy in tasks like digit recognition.

Humans make nearly 35,000 decisions each day, ranging from determining if it’s safe to cross the road to choosing what to have for lunch. Each decision involves evaluating options, recalling similar past situations, and feeling reasonably confident about the right choice. What might appear to be a snap decision actually results from gathering evidence from the environment. Additionally, the same person might make different decisions in identical scenarios at different times.

Neural networks do the opposite, making the same decisions each time. Now, Georgia Tech researchers in Associate Professor Dobromir Rahnev’s lab are training them to make decisions more like humans. This science of human decision-making is only just being applied to machine learning, but developing a neural network even closer to the actual human brain may make it more reliable, according to the researchers.

In a paper in Nature Human Behaviour, a team from the School of Psychology reveals a new neural network trained to make decisions similar to humans.

Decoding Decision

“Neural networks make a decision without telling you whether or not they are confident about their decision,” said Farshad Rafiei, who earned his Ph.D. in psychology at Georgia Tech. “This is one of the essential differences from how people make decisions.”

Large language models (LLM), for example, are prone to hallucinations. When an LLM is asked a question it doesn’t know the answer to, it will make up something without acknowledging the artifice. By contrast, most humans in the same situation will admit they don’t know the answer. Building a more human-like neural network can prevent this duplicity and lead to more accurate answers.

Making the Model

The team trained their neural network on handwritten digits from a famous computer science dataset called MNIST and asked it to decipher each number. To determine the model’s accuracy, they ran it with the original dataset and then added noise to the digits to make it harder for humans to discern. To compare the model performance against humans, they trained their model (as well as three other models: CNet, BLNet, and MSDNet) on the original MNIST dataset without noise, but tested them on the noisy version used in the experiments and compared results from the two datasets.

The researchers’ model relied on two key components: a Bayesian neural network (BNN), which uses probability to make decisions, and an evidence accumulation process that keeps track of the evidence for each choice. The BNN produces responses that are slightly different each time. As it gathers more evidence, the accumulation process can sometimes favor one choice and sometimes another. Once there is enough evidence to decide, the RTNet stops the accumulation process and makes a decision.

The researchers also timed the model’s decision-making speed to see whether it follows a psychological phenomenon called the “speed-accuracy trade-off” that dictates that humans are less accurate when they must make decisions quickly.

Once they had the model’s results, they compared them to humans’ results. Sixty Georgia Tech students viewed the same dataset and shared their confidence in their decisions, and the researchers found the accuracy rate, response time, and confidence patterns were similar between the humans and the neural network.

“Generally speaking, we don’t have enough human data in existing computer science literature, so we don’t know how people will behave when they are exposed to these images. This limitation hinders the development of models that accurately replicate human decision-making,” Rafiei said. “This work provides one of the biggest datasets of humans responding to MNIST.”

Not only did the team’s model outperform all rival deterministic models, but it also was more accurate in higher-speed scenarios due to another fundamental element of human psychology: RTNet behaves like humans. For example, people feel more confident when they make correct decisions. Without even having to train the model specifically to favor confidence, the model automatically applied it, Rafiei noted.

“If we try to make our models closer to the human brain, it will show in the behavior itself without fine-tuning,” he said.

The research team hopes to train the neural network on more varied datasets to test its potential. They also expect to apply this BNN model to other neural networks to enable them to rationalize more like humans. Eventually, algorithms won’t just be able to emulate our decision-making abilities, but could even help offload some of the cognitive burden of those 35,000 decisions we make daily.

Reference: “The neural network RTNet exhibits the signatures of human perceptual decision-making” by Farshad Rafiei, Medha Shekhar and Dobromir Rahnev, 12 July 2024, Nature Human Behaviour.
DOI: 10.1038/s41562-024-01914-8

Be the first to comment on "AI Learns To Think Like Humans: A Game-Changer in Machine Learning"

Leave a comment

Email address is optional. If provided, your email will not be published or shared.