In a previous blog post we discussed why explainable AI is so important and how different AI models fall short of helping humans understand the basis for their decisions. How can our modern-day AI marvels be so good at making predictions but so bad at helping us understand the basis for these decisions? Nowhere is this problem more pronounced than with the most popular AI model in existence today – the artificial neural network (ANN). Fortunately, and somewhat ironically, the reason neural networks are so hard to explain is quite easy to explain.
First, it is important to understand that an explanation for the output of any system can only be expressed in terms of the input variables to the system. You can have many inputs to a system, and some may have a greater impact on the output than others, but the explanation of the output must be expressed in terms of these inputs. For this reason, it is imperative that the inputs are preserved as they make their way to through system.
Neural Networks: Making a Smoothie with Data
A primary reason that neural networks are hard to explain is because they are destructive to input information. Let’s take a look at what is happening in a little more detail.
The image above shows the processing that goes on within each node of a neural network.
- First, every input node will sum all of the inputs to the system.
- Second, each node applies an activation function to the calculated sum. Ponder this for a moment. A single node takes all inputs, adds them together and then passes the sum to a function that changes the output into some other value depending on the magnitude of the sum.
- Finally, weights are calculated that determine how much of this nodes’ value should be passed to subsequent down-stream nodes.
The end result is that neural networks are very destructive to data – a sort of data blender. Input data is mixed in various proportions. It is very difficult to preserve the contribution of any single ingredient.
But wait, it gets worse. Much worse. As these blended results move through the multiple layers of the neural network, they face more data blenders. Every node in a neural network blends data and passes this blended result, in various proportions, onto other nodes that create even more complex blends of data. By the time you reach the output layer, the contents of your blenders are indecipherable.
Note: See this article for more on how neural networks work
The Output Layer with a Neural Network
The final stage of a neural network is the output layer. The output layer is quite simple in operation. For every possible prediction, there is a corresponding node in the output layer. Once all the blending has been completed, the output node with the highest value becomes the prediction of the system.
Think of the output layer as a series of lightbulbs. The brightest lightbulb becomes the prediction of the machine. Unfortunately, the slurry of data that is left in the blender is practically useless in helping understand why the system made this prediction.
The Output Layer with Pattern-based ML: Preserving the data
At Natural Intelligence, we are focused on pattern-based AI for some very good reasons. One of the most important is that pattern-based AI is non-destructive to the input data. In other words, we preserve the data all the way from the input of the system to the output prediction. We not only get a prediction, but we also get a ‘recipe’ of input ingredients for that prediction. Within this recipe is the underlying basis for the prediction.
How we preserve this data integrity through our system is beyond the scope of this blog post, but rest assured we have taken great care to implement the same fundamental processing concepts exhibited by the human neocortex and how it responds to patterns. With this understanding it should be no surprise that our system exhibits more human like characteristics than mathematically based AI models.
Output Layer Capacity
Let’s compare the output layer of a neural network with that of a pattern-based ML model. Much like a neural network, the output layer of a pattern-based AI system is a collection of nodes,. However, this is where the similarity ends.
As discussed previously, the output node of a neural network is like a dimmable lightbulb, and the node with the highest (brightest) value becomes the winning node. The number of objects that a neural network can detect is limited to the number of nodes in the output layer. If your neural net has 20 nodes in the output layer, you have capacity to learn up to 20 objects – no more.
An output node in a pattern-based ML model has a much higher capacity for detecting objects. Rather than each node representing a prediction, we look at the pattern of nodes that are represented in the output layer. A pattern-based system with 20 output nodes, of which 5 are active, has a capacity to detect more than 15,000 objects. In probability theory this is referred to as 20C5 *. Compare this to the minuscule capacity of a neural network with the same number of output nodes and you will begin to understand why pattern-based AI is so powerful.
This 20-node example is quite small. Our pattern-based AI systems are often configured with 100 output nodes or more. In these configurations the number of objects that can be learned and detected becomes astronomical. You can see that pattern-based AI has a much high capacity to learn and detect objects.
Increased learning capacity is critically important to delivering Explainable AI. Classifications in machine learning are often based on subtle differences between inputs. High learning capacity provides a correspondingly large number of ways to express those subtle differences between input classes.
Even better, the critical information describing these subtle differences is delivered directly to the data scientist. Gone are the days of trying to unwind the hidden layers of the neural net in an attempt to understand the subtle differences between classifications.
Higher Capacity Learning means Greater Understanding
While the higher capacity of a pattern-based AI system is a good start, we can go much further. In the image above we see five nodes in the output layer of the pattern-based system have been activated. This exact combination (pattern) of nodes corresponds to a specific object class that was input into the system.
For example, this pattern of active nodes might relate to a football in a system designed to classify sporting goods products. But this is just a pattern of nodes. Where is the explanation for why these nodes have turned on? On what basis did the pattern-based system come to a prediction that this pattern of nodes represents a football?
Here is where the magic begins and why it is so important to preserve the integrity of input data in an AI system. We can look inside each node and understand why it turned on. Because pattern-based AI eliminates the problem of ‘blending’ the input values, we can trace backwards from each output node to the very input data responsible for turning the node on.
Furthermore, each node can represent multiple input values in various combinations. Looking again at the output layer of our pattern-based system, we can see the input conditions that caused that node to turn on by backtracking this output node all the way back to the input values.
This single node is sensitive to input values related to inflation pressure, texture, and weight. Similarly, the other nodes would be related to other properties that are commonly associated with footballs, for example, size (dimensions), color, laces, and overall shape.
The Natural Intelligence pattern-based AI has more capacity and provides greater insights than neural network models. This is a powerful combination that can be used in many different ways to help data scientists understand their AI models better and to increase the confidence in AI systems overall.
Compared to how neural networks destroy input data and the simplistic brightest bulb concept used by neural networks, the power and value of pattern-based AI is clear to see.
* Capacity of system (xCy) = x!/((x-y)!y where x=number of nodes, y=number of active nodes
Download Explainability White Paper
To learn more about Natural Intelligence and Explainable AI, please fill in the form below to download our white paper: