:::{note} These slides are also available as a book page. :::
We illustrate the use of Bayesian networks in ProbLog using the famous Earthquake example.
Suppose there is a burglary in our house with probability 0.7 and an earthquake with probability 0.2. Whether our alarm will ring depends on both burglary and earthquake:
To model this as a Bayesian network, one would use three random variables, burglary, earthquake and alarm, with burglary and earthquake being parents of alarm. To model this in ProbLog, there are two possible solutions: using {ref}'plain' ProbLog <problog:probabilistic_facts>
or using some {ref}syntactic sugar <problog:probabilistic_clauses>
called probabilistic clauses and annotated disjunctions. We now explain both solutions.
digraph alarm1 { burglary -> alarm; earthquake -> alarm; }
ProbLog syntax documentation
In 'plain' ProbLog, we can encode the Bayesian network as follows.
alarm :- burglary, earthquake, p_alarm1
, with p_alarm1
an auxiliary atom defined by means of the probabilistic fact 0.9::p_alarm1
. The point of adding this atom is to ensure that the probability of alarm in this case will be 0.9 as required.We obtain the following {ref}ProbLog program <problog:probabilistic_facts>
.
```{problog} problog:probabilistic_facts 0.7::burglary. 0.2::earthquake. 0.9::p_alarm1. 0.8::p_alarm2. 0.1::p_alarm3.
alarm :- burglary, earthquake, p_alarm1. alarm :- burglary, +earthquake, p_alarm2. alarm :- +burglary, earthquake, p_alarm3.
evidence(alarm,true).
query(burglary). query(earthquake). ```
When pressing 'Evaluate', ProbLog2 calculates the probability of there being a burglary or an earthquake, given the evidence that the alarm rang. The resulting marginals are: $P(burglary)=0.9896$ and $P(earthquake)=0.2275$.
While the above is a correct encoding of the given Bayesian network, it is perhaps not very intuitive due to the auxiliary atoms. Fortunately, ProbLog2 offers some syntactic sugar called probabilistic clauses to encode this in a more readable way. Above, we encoded the information that the conditional probability of an alarm given a burglary and an earthquake equals 0.9 using the rule alarm :- burglary, earthquake, p_alarm1
, plus the probabilistic fact 0.9::p_alarm1
. We can replace both with a single probabilistic clause of the form 0.9::alarm :- burglary, earthquake
. This should be read as: if burglary and earthquake are true, this causes alarm to become true with probability 0.9 if there is a burglary and an earthquake. As this example illustrates, a probabilistic clause has a body, just like regular ProbLog rules, and a head. The difference is that now, the head is annotated with a probability. By also using probabilistic clauses for the other rules in the ProbLog encoding of the Bayesian network, we get {ref}the following program <problog:probabilistic_clauses>
.
```{problog} problog:probabilistic_clauses 0.7::burglary. 0.2::earthquake.
0.9::alarm :- burglary, earthquake. 0.8::alarm :- burglary, +earthquake. 0.1::alarm :- +burglary, earthquake.
evidence(alarm,true). query(burglary). query(earthquake). ```
As you can verify by pressing 'Evaluate', this returns the same marginals as the {ref}'plain' ProbLog encoding <problog:probabilistic_facts>
: $P(burglary)=0.9896$ and $P(earthquake)=0.2275$. This is not a coincidence: the two programs are equivalent (but the program with probabilistic clauses has the advantage of not needing any auxiliary atoms).
To illustrate the use of first-order ProbLog programs, we show below a {ref}first-order extension <problog:first_order>
of the Alarm example.
digraph alarm2 { burglary -> alarm; earthquake -> alarm; alarm -> "calls(john)"; alarm -> "calls(...)"; alarm -> "calls(mary)"; }
Suppose there are $N$ people and each person independently calls the police with a certain probability, depending on the alarm ringing or not: if the alarm rings, the probability of calling is 0.8, otherwise it is 0.1. This can be modelled as follows. We again use probabilistic clauses and show the case $N=2$ (two people).
```{problog} problog:first_order person(john). person(mary).
0.7::burglary. 0.2::earthquake.
0.9::alarm :- burglary, earthquake. 0.8::alarm :- burglary, +earthquake. 0.1::alarm :- +burglary, earthquake.
0.8::calls(X) :- alarm, person(X). 0.1::calls(X) :- +alarm, person(X).
evidence(calls(john),true). evidence(calls(mary),true).
query(burglary). query(earthquake). ```
When pressing 'Evaluate', ProbLog2 calculates the probability of there being a burglary or an earthquake, given the evidence that both john and mary called. We obtain $P(burglary)=0.981939$ and $P(earthquake)=0.226851$.
In general, any Boolean Bayesian network can be modeled in ProbLog using the above methodology. This can also be extended to non-Boolean Bayesian networks (in which some variables can take more than two possible values), by using annotated disjunctions with multiple atoms in the head.
Since the random variables in the Bayesian network are all Boolean, we only need a single literal in the head of the rules. We can extend the Bayesian network to have a multi-valued variable by indicating the severity of the earthquake. The literal earthquake
now has three possible values none
, mild
, heavy
instead of previously two (no or yes). The probabilities of each value is denoted by the annotated disjunction in 0.01::earthquake(heavy); 0.19::earthquake(mild); 0.8::earthquake(none)
. An annotated disjunction is similar to a probabilistic disjunction, but with a different head. Instead of it being one atom annotated with a probability, it is now a disjunction of atoms each annotated with a probability.
{problog} problog:annotated_disjunctions