5 Propositions and Inference

The third edition of Artificial Intelligence: foundations of computational agents, Cambridge University Press, 2023 is now available (including full text).

5.8 Causal Models

A primitive atom is an atom that is defined using facts. A derived atom is an atom that is defined using rules. Typically, the designer writes axioms for the derived atoms and then expects a user to specify which primitive atoms are true. Thus, a derived atom will be inferred as necessary from the primitive atoms and other atoms that can be derived.

The designer of an agent must make many decisions when designing a knowledge base for a domain. For example, consider two propositions, a and b, both of which are true. There are many choices of how to write this. A designer could specify both a and b as atomic clauses, treating both as primitive. A designer could have a as primitive and b as derived, stating a as an atomic clause and giving the rule ba. Alternatively, the designer could specify the atomic clause b and the rule ab, treating b as primitive and a as derived. These representations are logically equivalent; they cannot be distinguished logically. However, they have different effects when the knowledge base is changed. Suppose a was no longer true for some reason. In the first and third representations, b would still be true, and in the second representation b would no longer true.

A causal model, or a model of causality, is a representation of a domain that predicts the results of interventions. An intervention is an action that forces a variable to have a particular value. That is, an intervention changes the value in some way other than manipulating other variables in the model.

To predict the effect of interventions, a causal model represents how the cause implies its effect. When the cause is changed, its effect should be changed. An evidential model represents a domain in the other direction – from effect to cause. Note that we do not assume that there is “the cause” of an effect; rather there are propositions, which together may cause the effect to become true.

Example 5.36.

In the electrical domain depicted in Figure 5.2, consider the relationship between switches s1 and s2 and light l1. Assume all components are working properly. Light l1 is lit whenever both switches are up or both switches are down. Thus,

lit_l1(up_s1up_s2). (5.1)

This is logically equivalent to

up_s1(lit_l1up_s2).

This formula is symmetric between the three propositions; it is true if and only if an odd number of the propositions are true. However, in the world, the relationship between these propositions is not symmetric. Suppose both switches were up and the light was lit. Putting s1 down does not make s2 go down to preserve lit_l1. Instead, putting s1 down makes lit_l1 false, and up_s2 remains true. Thus, to predict the result of interventions, we require more than proposition (5.1) above.

A causal model is

lit_l1up_s1up_s2.
lit_l1up_s1up_s2.

The completion of this is equivalent to proposition (5.1); however, it makes reasonable predictions when one of the values is changed. Changing one of the switch positions changes whether the light is lit, but changing whether the light is lit (by some other mechanism) does not change whether the switches are up or down.

An evidential model is

up_s1lit_l1up_s2.
up_s1lit_l1up_s2.

This can be used to answer questions about whether s1 is up based on the position of s2 and whether l1 is lit. Its completion is also equivalent to formula (5.1). However, it does not accurately predict the effect of interventions.

For most purposes, it is preferable to use a causal model of the world as it is more transparent, stable and modular than an evidential model.