Skip to main content

Verified by Psychology Today

Artificial Intelligence

Why AI Lacks Judgment

Human involvement is necessary for decision-making.

Key points

  • Despite significant progress in recent years, AI programs lack judgment.
  • Judgment involves the wise resolution of new problems, striking a balance between different values.
  • AI programs cannot take responsibility and thus cannot function independently as decision-makers.

Artificial intelligence (AI) has undoubtedly made remarkable progress in recent years, demonstrating its ability to process vast amounts of data and perform complex tasks. However, it is crucial to recognize that AI lacks a fundamental quality that humans possess: judgment.

Aristotle distinguishes between three types of knowledge: episteme, techne and phronesis. Episteme is the theoretical knowledge of how humans and their environment work. This form of knowledge is the one most commonly taught in schools and universities.

Techne includes skills and abilities and is about how things are done. Knowing a craft is a good example. Technology is a later form of such knowledge.

Phronesis is the judgmental knowledge that enables one to make informed decisions in concrete situations. This form of knowledge can be translated into practical wisdom or good judgment.

AI provides us with tools to create new knowledge. The question is what kind of knowledge AI gives us. To answer, we must first find out what AI can do.

One of the first AI areas was various games that require human intelligence. For example, the chess program Deep Blue beat the then-world champion Garri Kasparov in 1997; since then, several gaming programs have evolved to a level far beyond human ability. They are, however, highly specialized and cannot be modified to solve other tasks.

Gaming programs deal more with problem-solving than creating new knowledge. AI programs that provide new knowledge fall into the episteme.

Recently, the focus has been on the programming method known as deep learning, where an artificial neural network with many layers (hence "deep") is trained with large amounts of data to recognize different types of patterns—for example, faces. Such training can make the program categorize or identify better than humans.

Deep learning has been successfully used in medicine to, for example, interpret X-ray images. The AlphaFold program has shown that, based on the sequence of amino acids in a protein, it can describe the three-dimensional structure of the protein in almost 90 percent of cases. Such tasks have previously been very difficult for researchers to solve and the AI program opens up great opportunities for medical research and the pharmaceutical industry.

Similar methods are used in speech recognition, as used by Siri and Alexa, among others. When it comes to automatic translation, AI programs are mainly based on statistical methods.

However, newer large language models, such as the recent GPT series, which automatically generates texts, are also based on deep learning. The programs are enormous (they contain hundreds of billions of variables) and are trained on huge amounts of text examples from the Internet. The large language models generate good texts, but they are typically deemed lacking in creativity.

The success of different types of AI programs makes it easy to forget what they cannot do. The performance of AI systems heavily relies on the datasets they are trained on. If an exceptional case arises, they have no way of adapting.

A poignant example occurred when Amazon's AI inventory system faltered under the unexpected surge in orders for toilet paper and face masks during the onset of the COVID-19 pandemic. The programs cannot find solutions to new types of problems. Here, we humans are clearly superior to AI systems.

Training data can also distort results. AI programs have been used in courts to adjudicate certain routine cases, such as traffic violations. The program had been trained on a large number of previous cases and learned which factors are relevant and which sentences should be given.

However, it turned out that the program sentenced Black people more harshly than white people because the previous cases also had such a bias. It is therefore very important to use appropriate training data to obtain fair assessments in programs that determine who should receive, for example, health care benefits, social assistance, or bank loans.

Robots and AI are often confused, but they are two different fields. The word "robot," introduced by author Karel Čapek, means "work." Thus, a robot is something that performs concrete actions.

We have been fooled by science fiction movies into thinking that intelligent and capable robots exist. But the robots that exist in reality—industrial robots, robot vacuum cleaners, and robot lawnmowers—have no techne. They can't go outside the narrow range of actions they've been programmed for, nor can they explain why they behave the way they do. They are utterly stupid and, in most cases, clumsy.

Creating robots that can independently solve new practical problems requires a great deal of technical and cognitive knowledge, and progress is slow. A skilled robot carpenter is a long way off.

To return to Aristotle's types of knowledge, AI can exhibit a certain level of episteme but the programs have no techne. What about phronesis, or judgment?

While it is possible to incorporate rule-based systems into AI, the challenge lies in determining the applicability of rules in individual cases. Professionals across various fields, including nurses, doctors, healthcare practitioners, social workers, lawyers, and teachers, understand that relying solely on a rulebook is inadequate. There are always borderline cases and unique combinations of circumstances that demand novel approaches. This is precisely when human judgment becomes indispensable.

To comprehend why AI lacks judgment, it is crucial to grasp the concept itself. Phronesis involves the wise resolution of problems in specific situations, striking a balance between different values based on the knowledge available about the situation. It does not imply a conflict between emotions and reason, but rather an intricate interplay between the two.

However, AI systems lack emotions and morality, rendering them incapable of assuming responsibility. Regardless of how extensive a rulebook may be, it is insufficient to understand how to adapt values to specific circumstances. Professionals with judgment possess the ability to do so, even though the exact mechanisms of their decision-making process remain elusive. Unfortunately, the concept of judgment receives relatively little attention in the humanities and social sciences.

Judgment will also be necessary for future robots when interacting with humans. Self-driving cars require judgment not only to determine whether a cyclist will turn left or when a pedestrian should cross the street, but especially in situations where an accident is imminent. Decisions must be made regarding prioritizing the safety of the car's occupants or individuals outside the vehicle.

In summary, AI systems possess a degree of episteme but lack both techne and phronesis. Consequently, these systems cannot currently function independently as decision-makers. Human involvement, particularly from individuals with experience, is necessary to handle exceptional cases that AI programs are not trained for and to prevent biased decision-making.

To create AI systems with broader capabilities, expertise in programming and technology alone is insufficient. Knowledge of human behavior, societies, and values is equally essential. Therefore, a multidisciplinary approach that combines technical expertise with insights from the humanities and social sciences is crucial for the advancement of AI and its responsible integration into various domains of society.

advertisement
More from Peter Gärdenfors Ph.D.
More from Psychology Today
More from Peter Gärdenfors Ph.D.
More from Psychology Today