How Human Expertise Enhances AI Decision-Making: Study Finds Improved Accuracy with Human Input

Date:

RICHLAND, Wash. — Can AI be trusted? The question pops up wherever AI is used or discussed — which, these days, is everywhere.

It’s a question that even some AI systems ask themselves.

Many machine-learning systems create what experts call a confidence score, a value that reflects how confident the system is in its decisions. A low score tells the human user that there is some uncertainty about the recommendation; a high score indicates to the human user that the system, at least, is quite sure of its decisions. Savvy humans know to check the confidence score when deciding whether to trust the recommendation of a machine-learning system.

Scientists at the Department of Energy’s Pacific Northwest National Laboratory have put forth a new way to evaluate an AI system’s recommendations. They bring human experts into the loop to view how the ML performed on a set of data. The expert learns which types of data the machine-learning system typically classifies correctly, and which data types lead to confusion and system errors. Armed with this knowledge, the experts then offer their own confidence score on future system recommendations.

The result of having a human look over the shoulder of the AI system? Humans predicted the AI system’s performance more accurately.

Minimal human effort — just a few hours — evaluating some of the decisions made by the AI program allowed researchers to vastly improve on the AI program’s ability to assess its decisions. In some analyses by the team, the accuracy of the confidence score doubled when a human provided the score.

See also  New Study Reveals Key Genes for Drought Adaptation in Tomatoes

If you didn’t develop the machine-learning algorithm in the first place, then it can seem like a black box, said Corey Fallon, the lead author of the study and an expert in human-machine interaction. In some cases, the decisions seem fine. In other cases, you might get a recommendation that is a real head-scratcher. You may not understand why it’s making the decisions it is.

It’s a dilemma that power engineers working with the electric grid face. Their decisions based on reams of data that change every instant keep the lights on and the nation running. But power engineers may be reluctant to turn over decision-making authority to machine-learning systems.

There are hundreds of research papers about the use of machine learning in power systems, but almost none of them are applied in the real world. Many operators simply don’t trust ML. They have domain experience — something that ML can’t learn, said coauthor Tianzhixi Tim Yin.

The researchers at PNNL, which has a world-class team modernizing the grid, took a closer look at one machine-learning algorithm applied to power systems. They trained the SVM (support-vector machine) algorithm on real data from the grid’s Eastern Interconnection in the U.S. The program looked at 124 events, deciding whether a generator was malfunctioning, or whether the data was showing other types of events that are less noteworthy.

The algorithm was 85% reliable in its decisions. Many of its errors occurred when there were complex power bumps or frequency shifts. Confidence scores created with a human in the loop were a marked improvement over the system’s assessment of its own decisions. The human expert’s input predicted the algorithm’s decisions with much greater accuracy.

See also  AI Revolutionizing ESG Compliance for Sustainable Impact

Fallon and Yin call the new score an Expert-Derived Confidence score, or EDC score.

They found that, on average, when humans weighed in on the data, their EDC scores predicted model behavior that the algorithm’s confidence scores couldn’t predict.

The human expert fills in gaps in the ML’s knowledge, said Yin. The human provides information that the ML did not have, and we show that that information is significant. The bottom line is that we’ve shown that if you add human expertise to the ML results, you get much better confidence.

The work by Fallon and Yin was funded by PNNL through an initiative known as MARS — Mathematics for Artificial Reasoning in Science. The effort is part of a broader effort in artificial intelligence at PNNL. The initiative brought together Fallon, an expert on human-machine teaming and human factors research, and Yin, a data scientist and an expert on machine learning.

This is the type of research needed to prepare and equip an AI-ready workforce, said Fallon. If people don’t trust the tool, then you’ve wasted your time and money. You’ve got to know what will happen when you take a machine learning model out of the laboratory and put it to work in the real world.

I’m a big fan of human expertise and of human-machine teaming. Our EDC scores allow the human to better assess the situation and make the ultimate decision.

Frequently Asked Questions (FAQs) Related to the Above News

What is the purpose of the study conducted by scientists at Pacific Northwest National Laboratory?

The purpose of the study was to evaluate an AI system's recommendations by incorporating human experts into the decision-making process.

How did human experts improve the accuracy of the AI system?

Human experts provided their own confidence score on the system's recommendations, which greatly improved the accuracy of the AI system's assessments.

Why do power engineers often hesitate to trust machine learning in power systems?

Power engineers have domain experience that machine learning cannot learn, making them hesitant to trust the decisions made by machine-learning systems.

What type of algorithm was used in the study?

The study utilized the SVM (support-vector machine) algorithm trained on real data from the grid's Eastern Interconnection in the U.S.

What is the Expert-Derived Confidence score (EDC score)?

The EDC score is a new score developed by the researchers that incorporates human expertise to predict the behavior and decisions of the machine-learning algorithm more accurately.

What initiative funded the research conducted by PNNL scientists?

The research was funded by PNNL through an initiative called MARS (Mathematics for Artificial Reasoning in Science).

What is the importance of incorporating human expertise into AI decision-making?

Incorporating human expertise helps fill knowledge gaps and provides additional information that the machine learning algorithm may not have, resulting in better decision-making and increased confidence in the AI system.

How does this research contribute to the development of an AI-ready workforce?

This research helps in preparing and equipping an AI-ready workforce by emphasizing the importance of trust in machine learning tools and the need for human involvement in decision-making processes.

What are some potential implications of these findings?

These findings highlight the potential for improving AI decision-making accuracy through the incorporation of human expertise. It also emphasizes the importance of trust and understanding when deploying machine learning models in real-world scenarios.

Please note that the FAQs provided on this page are based on the news article published. While we strive to provide accurate and up-to-date information, it is always recommended to consult relevant authorities or professionals before making any decisions or taking action based on the FAQs or the news article.

Share post:

Subscribe

Popular

More like this
Related

AI Revolutionizing Software Engineering: Industry Insights Revealed

Discover how AI is revolutionizing software engineering with industry insights. Learn how AI agents are transforming coding and development processes.

AI Virus Leveraging ChatGPT Spreading Through Human-Like Emails

Stay informed about the AI Virus leveraging ChatGPT to spread through human-like emails and the impact on cybersecurity defenses.

OpenAI’s ChatGPT Mac App Update Ensures Privacy with Encrypted Chats

Stay protected with OpenAI's ChatGPT Mac app update that encrypts chats to enhance user privacy and security. Get the latest version now!

The Rise of AI in Ukraine’s War: A Threat to Human Control

The rise of AI in Ukraine's war poses a threat to human control as drones advance towards fully autonomous weapons.