In the first part of this post, I gave the basic intuition behind Bayesian belief networks (or just Bayesian networks) — what they are, what they’re used for, and how information is exchanged between their nodes.
In this post, I’m going to show the math underlying everything I talked about in the previous one. It’s going to be a bit more technical, but I’m going to try to give the intuition behind the relevant equations.
If you stick to the end, I promise you’ll get a much deeper understanding of Bayesian networks. To the point of actually being able to use them for real-world calculations.
I’m going to start with a recap of what I talked about in the previous post. Then, I’m going to show how to mathematically represent Bayesian networks and make inferences about the states of its nodes.
Reading two of my previous posts would be very helpful along the way, especially if you’re new to the topic.
In the Calculating Compound Event Probabilities post (specifically, the “General formulas for compound events” section) you’ll find the derivation of the formula for the joint probability distribution of multiple events (which I’m going to use here).
The post The Anatomy Of Bayes’ Theorem gives a really good intuition about the way an observation updates the probabilities of rival hypotheses that try to explain it. I specifically recommend the section explaining the evidence term of Bayes’ theorem because I’m going to use a very similar method in explaining the math of information propagation in Bayesian networks.
Bayesian networks consist of nodes connected by arrows. You usually graphically illustrate the nodes as circles.
Each node represents the probability distribution of a set of mutually exclusive outcomes. For example, a node can represent the outcome of rolling a die, with each side having a probability of 1/6 to be on top. Or the outcome of flipping of a coin, with each side having a probability of 1/2.
More generally, each node represents the possible states of a variable and holds their respective probabilities.
The arrows hold the information about the conditional probabilities between the nodes they connect. The node the arrow is pointing to (the child node) depends on the one the arrow originates from (the parent node). This means that the probability of a child node taking a particular state will depend on the state of its parent.
States of nodes in the network
In the first post, I came up with this example for a Bayesian network:
In this example, the “Season” node’s possible states are:
The “Rain” node’s states are:
- It’s raining
- It’s not raining
You get the idea.
If you haven’t made any observations related to the network, then each node will be in one of its possible states with a certain probability. For example, if you pick a random day of the year (but you haven’t seen the day you’ve picked yet), the probability that you’ve picked any of the 4 possible seasons is 1/4 (assuming the seasons are of equal length). So, the “Season” node’s probability distribution is:
- P(Spring) = 1/4
- P(Summer) = 1/4
- P(Fall) = 1/4
- P(Winter) = 1/4
If you observe the value of a node, its state is no longer probabilistic. For example, if you already know that the day you’ve picked is July 17, the probabilities for the “Season” node become:
- P(Spring) = 0
- P(Summer) = 1
- P(Fall) = 0
- P(Winter) = 0
This simply reflects the fact that once you’ve observed a node, there’s no longer uncertainty about which state it’s in.
Dependence between states
If there’s an arrow between two nodes, the state of the child node will probabilistically depend on the state of its parent. This is expressed as a conditional probability of one node given another:
- P(Node 2 = State X | Node 1 = State Y).
In other words, this is the probability of Node 2 being in State X, if Node 1 is State Y.
For example, if the season is fall, the probability of rain will be higher than if the season is summer:
- P(Rain | Fall) > P(Rain | Summer)
By the way, this dependence between nodes is precisely what allows information propagation within Bayesian networks. Observing the state of a node updates the probabilities of its children and its parents. Then, the newly updated nodes update their children/parents, and so on.
The state of the whole network
A very useful thing to be able to represent is the states of all network nodes simultaneously. Here’s an example:
You see that each node is now in a particular state, instead of being in a state of uncertainty (the pinkish labels below the node names indicate the current state).
You can think of a set of node states as the state of the whole network. This state is known with certainty if you’ve observed the states of each node in the network.
The total number of network states is equal to the number of all combinations of node states (usually a very high number for most networks). Each network state has a probability of its own, which is the joint probability over all node states. And the list of all joint probabilities is the joint probability distribution of the network.
(I’m assuming you’re already familiar with joint probabilities. If not, please take a look at the 2 posts I linked to in the beginning.)
The joint probability of the example state above is written as:
- P(Spring, Low atmospheric pressure, Wet grass, Cat stressed…)
Notice that the expression above specifies the states of the nodes. A more explicit notation would look like:
- P(Season = Spring, Atmospheric pressure = Low, Grass = Wet, Cat mood = stressed…)
More generally, you can express the joint probability distribution just by writing the names of the nodes:
- P(Spring, Atmospheric pressure, Grass, Cat mood…)
The names are simply placeholders and can be replaced by particular states — either hypothetically or if the states are actually observed.
The joint probability distribution of a network
Expanding the joint probability above is done according to this formula. I’m not going to write the full expression here because it is very long. Instead, let’s look at a smaller part of the network:
According to the formula, the joint probability distribution of this small network would be:
- P(Dog bark, Cat mood, Cat hide) = P(Dog bark) * P(Cat mood | Dog bark) * P(Cat hide | Dog bark, Cat mood)
But here’s the beautiful part. The absence of an arrow between nodes means that the nodes are independent (it’s what you’re essentially assuming about 2 nodes if you don’t put an arrow between them). And this allows you to drop the dependency between those nodes in any further calculations.
In this case, there’s no arrow between the “Dog bark” and “Cat mood” nodes, which slightly simplifies the expression on the right-hand side to:
- P(Dog bark, Cat mood, Cat hide) = P(Dog bark) * P(Cat mood) * P(Cat hide | Dog bark, Cat mood)
By not having an arrow between two nodes, you’re basically assuming that the nodes are independent. But in reality they might not be and that would be a flaw of your network.
So, the presence/absence of arrows determines the dependency structure of the network. Once you have that, you also need to define the exact dependencies between the nodes with arrows, as well as the prior probabilities of the nodes without arrows. And you’re done: this defines the entire Bayesian network!
The final expression
Now, with this information in mind, I can actually write the joint distribution of the entire network from the example above, because it’s no longer that long:
- P(Network) = P(Season) * P(Atmospheric pressure) * P(Allergies | Season) * P(Rain | Season, Atmospheric pressure) * P(Grass | Rain) * P(Umbrellas | Season, Rain) * P(Dog bark | Rain) * P(Cat mood) * P(Cat hide | Dog bark, Cat mood)
I’m using the shorthand notation P(Network) to refer to the joint probability distribution over the nodes. And I’m going to use the same notion in the rest of this post.
The conditional probabilities are colored in red and the unconditional (prior) probabilities are colored in green. Notice how each node is only conditioned on its parents. This really shortened an otherwise extremely long and difficult to read expression! Hover over here to see what I’m talking about.
To summarize, the joint probability distribution of Bayesian networks is simply the product of the probabilities of its nodes. Each probability is conditioned only on the parents of the respective node and nodes that have no parents only have a prior probability.
Making inferences about the states of nodes
So, we finally come to the interesting part. Imagine you have a network with a specific joint probability distribution, P(Network). As you already know, if the state of one node is observed, this is going to change the probabilities of the states of (some) other nodes in the network.
In this section, I’m going to show how to mathematically calculate these new probabilities, called posterior probabilities,.
Conditioning the joint distribution on observations
Before any node is observed, each will have a particular prior probability distribution over its states. If a particular node’s state is observed, its probability distribution is changed to 1 for the observed state and 0 for the other states.
The posterior probability distribution of the remaining nodes is then calculated according to Bayes’ theorem:
You’ll remember this particular representation of Bayes’ theorem from my post The Anatomy Of Bayes’ Theorem I linked to in the beginning. More specifically, from the “More intuition about the evidence” section.
So, you see that all it takes to calculate the joint posterior distribution of the network is to divide its joint prior distribution by the marginal probability of the observed node. Yes, it’s really that simple!
And if you’ve observed the states of multiple nodes, you simply condition the full joint distribution on all observed nodes. On the right-hand side of the equation you will divide by the joint distribution of the observed nodes.
The intuition behind conditioning on observed states
I’m going to borrow an example from The Anatomy Of Bayes’ Theorem post for this section.
Imagine there are 4 mutually exclusive hypotheses that attempt to explain something general about the world. Assume that in the beginning each hypothesis is equally likely. Here’s a visual representation of the sample space of this hypothetical world:
Each of the 4 squares represents one of the hypotheses and all squares have the same area (because we said they have equal prior probabilities). So, each hypothesis has a probability of 1/4.
Now, imagine that you learn something about the world by making an observation of some of its states. Let’s say the new observation covers a specific part of the sample space:
The dark square represents the probability of the observation you just made (the probability it had before you made it, actually). And now that you’ve made the observation, you can exclude all the states outside of it. This simply means that the dark square becomes your new sample space:
In other words, you now know that that part of the sample space is actually the world you’re living in: the only parts of the old sample space that are consistent with the observation.
This new sample space represents the posterior probability distribution over the 4 hypotheses. Notice how their probabilities have changed and they’re no longer equally probable. This is because some of them were more consistent with the observation than others.
This may still sound a little abstract, but I just wanted to give a quick reminder of how probabilities are updated using Bayes’ theorem after new observations. In the next section, I’m going to show an application to a specific example that should make everything I’ve talked about so far much clearer.
So, let’s consider a small part of the Bayesian network I’ve been working with so far:
This was actually the very first example I gave in the previous post. When it rains, the dog tends to bark at the window. And when the dog barks, the cat tends to hide under the couch.
Now, consider only the top two nodes. The specific probability distribution for the “Rain” and “Dog bark” nodes I came up with last time was:
Okay, let’s start building the sample space. First of all, there is a 1/4 probability that it rains and 3/4 probability that it doesn’t rain:
Notice how the blue area (representing rain) covers 1/4 of the sample space.
Now, let’s superimpose the sample space of the “Dog bark” node onto the sample space of the “Rain” node:
Now, notice how 9 out of 12 “Rain” rectangles are also gray. This corresponds to the top left cell in the probability table above:
- P(Rains, Dog barks) = 9/48
Similarly, 18 out of the 36 “No rain” rectangles are gray, which corresponds to the probability:
- P(Doesn’t rain, Dog barks) = 18/48
In the same manner, you can calculate the remaining probabilities from the table.
Okay, what about the “Cat hide” node? Its probability table was:
Can we superimpose it on the same sample space? Yes, we can.
Notice the beige colored rectangles representing the “Cat hides = True” event. I’ve carefully colored all rectangles so that this 3-way joint sample space doesn’t contradict any of the probability tables.
Now, imagine you’ve actually looked at the weather outside and you know it’s currently raining. The new (posterior) joint distribution of the network is calculated by conditioning the distribution on the “Rains” event:
Expanding the right-hand side of the equation, we get:
- P(Dog barks, Cat hides | Rains) = P(Cat hides | Dog barks) * P(Dog barks | Rains) * P(Rains) / P(Rains)
The P(Rains) terms cancel out and the final posterior distribution becomes:
- P(Dog barks, Cat hides | Rains) = P(Cat hides | Dog barks) * P(Dog barks | Rains)
So, how does the posterior sample space look like? Remember, because we know it’s currently raining, we ignore the parts of the sample space representing the “Doesn’t rain” event. We only take the 12 blue rectangles on the left and make them the new sample space:
Can you see which of these 12 rectangles corresponds to which of the 12 rectangles from the left-most column from the prior sample space? Take some time to compare the two.
Out of these 12 rectangles:
- 3 represent the “Dog barks & Cat doesn’t hide” event
- 6 represent the “Dog barks & Cat hides” event
- 2 represent the “Dog doesn’t bark & Cat doesn’t hide” event
- 1 represents the “Dog doesn’t bark & Cat hides” event
Now we can answer all sorts of probabilistic questions about the individual events. For example, there are 9 “Dog barks” rectangles in total, which means the posterior probability of the dog barking is:
- P(Dog barks | Rains) = 9/12 = 0.75
Remember, the prior probability P(Dog barks) was 27/48 (about 0.56).
Similarly, there are 7 “Cat hides” rectangles, so the posterior probability of this event is:
- P(Cat hides | Rains) = 7/12 ≈ 0.58
Which is also higher than the prior 25/48 ≈ 0.52.
The big picture
To calculate the posterior probability distribution of Bayesian networks, you divide the prior probability distribution by the probability of the observed event(s). The theory is as simple as that.
I tried to give the intuition behind this calculation by visually showing how the sample space changes after a particular observation. But you obviously don’t have to draw sample spaces every time you want to update the probability distribution of a Bayesian network. In practice, you just apply the formula.
So, the mathematical expression for the example I discussed was:
- P(Dog barks, Cat hides | Rains) = P(Cat hides | Dog barks) * P(Dog barks | Rains)
Take a look at the prior sample space again.
Again, 9 out of the 12 “Rain” rectangles are also colored in gray (which represent the “Dog barks” event.) This means:
- P(Dog barks | Rains) = 9/12
Similarly, 18 out of the 27 “Dog barks” rectangles are also colored in beige (which represents the “Cat hides” event). Therefore:
- P(Cat hides | Dog barks) = 18/27
So, we finally get:
- P(Dog barks, Cat hides | Rains) = P(Cat hides | Dog barks) * P(Dog barks | Rains) = 18/27 * 9/12 = 1/2
If you go back to the posterior sample space, you’ll notice that the “Dog barks & Cat hides” compound event is represented by exactly 6 out of the 12 rectangles:
- P(Dog barks, Cat hides | Rains) = 6/12 = 1/2
As expected, the result is the same!
To calculate a specific event’s posterior probability, like P(Dog barks | Cat hides), you simply calculate the marginal probability of that event from the full posterior distribution.
In practice, things aren’t always as smooth. For more complex networks, calculating the evidence term (the probability of the event(s) you’re conditioning on) or the marginal probability of individual events can be a mathematical challenge. These more challenging cases are topics for future posts.
In this post, I gave the mathematical intuition behind Bayesian networks which I had introduced in Part 1.
Whether you’re making future predictions or inferences about possible explanations of observations, all you need to do to calculate the posterior distribution of the network is to condition it on the observed nodes. Which simply means to divide the prior joint probability distribution of the network by the joint probability of the observed states.
As I mentioned at the end of the last section, in practice many Bayesian belief networks are too complex and there is no exact mathematical solution for their posterior distributions. In such cases, people use different numerical methods to get an approximate solution. These are relatively advanced topics which I plan to cover in future posts, but you don’t have to worry about them for now. If you feel comfortable with the concepts and methods I introduced in these two posts, you will be able to build simple networks and calculate probabilities with them. And if you have a programming background, you will probably even be able to write your own simple networks to make calculations faster.
In some of my next posts, I’m going to show applications of Bayesian belief networks to some real-world problems. I think this will give even better intuition on how useful this tool really is.
One topic that I wanted to cover in this post, but didn’t, was the concept of conditional dependence and independence between nodes. In short, two nodes that were otherwise dependent can become independent if a particular third node’s state is observed. Similarly, nodes that were otherwise independent can become dependent, again after observing a particular third node. The conditional (in)dependence between nodes, given another node, depends on the particular dependency structure (the arrows) between the nodes in question.
Conditional (in)dependence is a central topic to Bayesian networks, so I certainly don’t want to omit it. However, this post already got long enough that I decided to leave this part for a separate post. In fact, this is going to be the very next post I write.