Logic and Probability

First published Thu Mar 7, 2013

Logic and probability theory are two of the main tools in the formal study of reasoning, and have been fruitfully applied in areas as diverse as philosophy, artificial intelligence, cognitive science and mathematics. This entry discusses the major proposals to combine logic and probability theory, and attempts to provide a classification of the various approaches in this rapidly developing field.

1. Combining Logic and Probability Theory

The very idea of combining logic and probability might look strange at first sight (Hájek 2001). After all, logic is concerned with absolutely certain truths and inferences, whereas probability theory deals with uncertainties. Furthermore, logic offers a qualitative (structural) perspective on inference (the deductive validity of an argument is based on the argument's formal structure), whereas probabilities are quantitative (numerical) in nature. However, as will be shown in the next section, there are natural senses in which probability theory presupposes and extends classical logic. Furthermore, historically speaking, several distinguished theorists such as De Morgan (1847), Boole (1854), Ramsey (1926), de Finetti (1937), Carnap (1950), Jeffrey (1992) and Howson (2003, 2007, 2009) have emphasized the tight connections between logic and probability, or even considered their work on probability as a part of logic itself.

By integrating the complementary perspectives of qualitative logic and numerical probability theory, probability logics are able to offer highly expressive accounts of inference. It should therefore come as no surprise that they have been applied in all fields that study reasoning mechanisms, such as philosophy, artificial intelligence, cognitive science and mathematics. The downside to this cross-disciplinary popularity is that terms such as “probability logic” are used by different researchers in different, non-equivalent ways. Therefore, before moving on to the actual discussion of the various approaches, we will first delineate the subject matter of this entry.

The most important distinction is that between probability logic and inductive logic. Classically, an argument is said to be (deductively) valid if and only if it is impossible that the premises of A are all true, while its conclusion is false. In other words, deductive validity amounts to truth preservation: in a valid argument, the truth of the premises guarantees the truth of the conclusion. In some arguments, however, the truth of the premises does not fully guarantee the truth of the conclusion, but it still renders it highly likely. A typical example is the argument with premises “The first swan I saw was white”, …, “The 1000th swan I saw was white”, and conclusion “All swans are white”. Such arguments are studied in inductive logic, which makes extensive use of probabilistic notions, and is therefore considered by some authors to be related to probability logic. There is some discussion about the exact relation between inductive logic and probability logic, which is summarized in the introduction of Kyburg (1994). The dominant position (defended by Adams and Levine (1975), among others), which is also adopted here, is that probability logic entirely belongs to deductive logic, and hence should not be concerned with inductive reasoning. Still, most work on inductive logic falls within the “probability preservation” approach, and is thus closely connected to the systems discussed in Section 2. For more on inductive logic, the reader can consult Fitelson (2006), Romeijn (2011), and the entries on the problem of induction and inductive logic of this encyclopedia.

We will also steer clear of the philosophical debate over the exact nature of probability. The formal systems discussed here are compatible with all of the common interpretations of probability, but obviously, in concrete applications, certain interpretations of probability will fit more naturally than others. For example, the modal probability logics discussed in Section 4 are, by themselves, neutral about the nature of probability, but when they are used to describe the behavior of a transition system, their probabilities are typically interpreted in an objective way, whereas modeling multi-agent scenarios is accompanied most naturally by a subjective interpretation of probabilities (as agents' degrees of belief). This topic is covered in detail in Gillies (2000), Eagle (2010), and the entry on interpretations of probability of this encyclopedia.

Finally, although the success of probability logic is largely due to its various applications, we will not deal with these applications in any detail. For example, we will not assess the use of probability as a formal representation of belief in philosophy (Bayesian epistemology) or artificial intelligence (knowledge representation), and its advantages and disadvantages with respect to alternative representations, such as generalized probability theory (for quantum theory) and fuzzy logic. For more information about these topics, the reader can consult Gerla (1994), Vennekens et al. (2009), Hájek and Hartmann (2010), Hartmann and Sprenger (2010) and the entries on formal representations of belief, Bayesian epistemology, defeasible reasoning, quantum logic and probability theory, and fuzzy logic of this encyclopedia.

With these clarifications in place, we are now ready to look at what will be discussed in this entry. The most common strategy to obtain a concrete system of probability logic is to start with a classical (propositional/modal/etc.) system of logic and to “probabilify” it in one way or another, by adding probabilistic features to it. There are various ways in which this probabilification can be implemented. One can study probabilistic semantics for classical languages (which do not have any explicit probabilistic operators), in which case the consequence relation itself gets a probabilistic flavor: deductive validity becomes “probability preservation”, rather than “truth preservation”. This direction will be discussed in Section 2. Alternatively, one can add various kinds of probabilistic operators to the syntax of the logic. In Section 3 we will discuss some initial, rather basic examples of probabilistic operators. The full expressivity of modal probabilistic operators will be explored in Section 4. Finally, languages with first-order probabilistic operators will be discussed in Section 5.

2. Propositional Probability Logics

In this section, we will present a first family of probability logics, which are used to study questions of “probability preservation” (or dually, “uncertainty propagation”). These systems do not extend the language with any probabilistic operators, but rather deal with a “classical” propositional language L, which has a countable set of atomic propositions, and the usual truth-functional (Boolean) connectives.

The main idea is that the premises of a valid argument can be uncertain, in which case (deductive) validity imposes no conditions on the (un)certainty of the conclusion. For example, the argument with premises “if it will rain tomorrow, I will get wet” and “it will rain tomorrow”, and conclusion “I will get wet” is valid, but if its second premise is uncertain, its conclusion will typically also be uncertain. Propositional probability logics represent such uncertainties as probabilities, and study how they “flow” from the premises to the conclusion; in other words, they do not study truth preservation, but rather probability preservation. The following three subsections discuss systems that deal with increasingly more general versions of this issue.

2.1 Probabilistic Semantics

We begin by recalling the notion of a probability function for the propositional language L. (In mathematics, probability functions are usually defined for a σ-algebra of subsets of a given set Ω, and required to satisfy countable additivity; cf. Subsection 4.3 of this entry. In logical contexts, however, it is often more natural to define probability functions “immediately” for the logic's object language (Williamson 2002). Because this language is finitary—all its formulas have finite length—, it also suffices to require finite additivity.) A probability function (for L) is a function P : L→ ℝ satisfying the following constraints:

Non-negativity.
P(ϕ) ≥ 0 for all ϕ ∈ L.
Tautologies.
If ⊧ ϕ, then P(ϕ) = 1.
Finite additivity.
If ⊧ ¬(ϕ ∧ ψ), then P(ϕ ∨ ψ) = P(ϕ) + P(ψ).

In the second and third constraint, the ⊧-symbol denotes (semantic) validity in classical propositional logic. The definition of probability functions thus requires notions from classical logic, and in this sense probability theory can be said to presuppose classical logic (Adams 1998, 22). It can easily be shown that if P satisfies these constraints, then P(ϕ) ∈ [0,1] for all formulas ϕ ∈ L, and P(ϕ) = P(ψ) for all formulas ϕ, ψ ∈ L that are logically equivalent (i.e., such that ⊧ ϕ ↔ ψ).

We now turn to probabilistic semantics, as defined in Leblanc (1983). An argument with premises Γ and conclusion ϕ—henceforth denoted as (Γ, ϕ)—is said to be probabilistically valid, written Γ ⊧p ϕ, if and only if

for all probability functions P : L→ ℝ:
if P(γ) = 1 for all γ ∈ Γ, then also P(ϕ) = 1.

Probabilistic semantics thus replaces the valuations v : L→{0,1} of classical propositional logic with probability functions P : L→ ℝ, which take values in the real unit interval [0,1]. The classical truth values of true (1) and false (0) can thus be regarded as the endpoints of the unit interval [0,1], and likewise, valuations v : L→{0,1} can be regarded as degenerate probability functions P : L→ [0,1]. In this sense, classical logic is a special case of probability logic, or equivalently, probability logic is an extension of classical logic.

It can be shown that classical propositional logic is (strongly) sound and complete with respect to probabilistic semantics:

Γ ⊧p ϕ if and only if Γ ⊢ ϕ.

Some authors interpret probabilities as generalized truth values (Reichenbach 1949; Leblanc 1983). According to this view, probability logic is just a particular kind of many-valued logic, and probabilistic validity boils down to “truth preservation”: truth (i.e., probability 1) carries over from the premises to the conclusion. Other logicians, such as Tarski (1936) and Adams (1998, 15), have noted that probabilities cannot be seen as generalized truth values, because probability functions are not “extensional”; for example, P(ϕ ∧ ψ) cannot be expressed as a function of P(ϕ) and P(ψ). More discussion on this topic can be found in Hailperin (1984).

Another possibility is to interpret a sentence's probability as a measure of its (un)certainty. For example, the sentence “Jones is in Spain at the moment” can have any degree of certainty, ranging from 0 (maximal uncertainty) to 1 (maximal certainty). (Note that 0 is actually a kind of certainty, viz. certainty about falsity; however, in this entry we follow Adams' terminology (1998, 31) and interpret 0 as maximal uncertainty.) According to this interpretation, the following theorem follows from the strong soundness and completeness of probabilistic semantics:

Theorem 1. Consider a deductively valid argument (Γ, ϕ). If all premises in Γ have probability 1, then the conclusion ϕ also has probability 1.

This theorem can be seen as a first, very partial clarification of the issue of probability preservation (or uncertainty propagation). It says that if there is no uncertainty whatsoever about the premises, then there cannot be any uncertainty about the conclusion either. In the next two subsections we will consider more interesting cases, when there is non-zero uncertainty about the premises, and ask how it carries over to the conclusion.

Finally, it should be noted that although this subsection only discussed probabilistic semantics for classical propositional logic, there are also probabilistic semantics for a variety of other logics, such as intuitionistic propositional logic (van Fraassen 1981b; Morgan and Leblanc 1983), modal logics (Morgan 1982a,b, 1983; Cross 1993), classical first-order logic (Leblanc 1979, 1984; van Fraassen 1981b), relevant logic (van Fraassen 1983) and nonmonotonic logic (Pearl 1991). All of these systems share a key feature: the logic's semantics is probabilistic in nature, but probabilities are not explicitly represented in the object language; hence, they are much closer in nature to the propositional probability logics discussed here than to the systems presented in later sections.

Most of these systems are not based on unary probabilities P(ϕ), but rather on conditional probabilities P(ϕ, ψ). The conditional probability P(ϕ, ψ) is taken as primitive (rather than being defined as P(ϕ ∧ ψ) ∕ P(ψ), as is usually done) to avoid problems when P(ψ) = 0. Goosens (1979) provides an overview of various axiomatizations of probability theory in terms of such primitive notions of conditional probability.

2.2 Adams' Probability Logic

In the previous subsection we discussed a first principle of probability preservation, which says that if all premises have probability 1, then the conclusion also has probability 1. Of course, more interesting cases arise when the premises are less than absolutely certain. Consider the valid argument with premises pq and pq, and conclusion q (the symbol “→” denotes the truth-conditional material conditional). One can easily show that

P(q) = P(pq) + P(pq) − 1.

In other words, if we know the probabilities of the argument's premises, then we can calculate the exact probability of its conclusion, and thus provide a complete answer to the question of probability preservation for this particular argument (for example, if P(pq) = 6∕7 and P(pq) = 5∕7, then P(q) = 4∕7). In general, however, it will not be possible to calculate the exact probability of the conclusion, given the probabilities of the premises; rather, the best we can hope for is a (tight) upper and/or lower bound for the conclusion's probability. We will now discuss Adams' (1998) methods to compute such bounds.

Adams' results can be stated more easily in terms of uncertainty rather than certainty (probability). Given a probability function P : L→ [0,1], the corresponding uncertainty function UP is defined as

UP : L → [0,1] : ϕ ↦ UP(ϕ) := 1 − P(ϕ).

If the probability function P is clear from the context, we will often simply write U instead of UP. In the remainder of this subsection (and in the next one as well) we will assume that all arguments have only finitely many premises (which is not a significant restriction, given the compactness property of classical propositional logic). Adams' first main result, which was originally established by Suppes (1966), can now be stated as follows:

Theorem 2. Consider a valid argument (Γ, ϕ) and a probability function P. Then the uncertainty of the conclusion ϕ cannot exceed the sum of the uncertainties of the premises γ ∈ Γ. Formally:

U(ϕ) ≤ γ∈Γ U(γ).

First of all, note that this theorem subsumes Theorem 1 as a special case: if P(γ) = 1 for all γ ∈ Γ, then U(γ) = 0 for all γ ∈ Γ, so U(ϕ) ≤ ∑ U(γ) = 0 and thus P(ϕ) = 1. Furthermore, note that the upper bound on the uncertainty of the conclusion depends on |Γ|, i.e., on the number of premises. If a valid argument has a small number of premises, each of which only has a small uncertainty (i.e., a high certainty), then its conclusion will also have a reasonably small uncertainty (i.e., a reasonably high certainty). Conversely, if a valid argument has premises with small uncertainties, then its conclusion can only be highly uncertain if the argument has a large number of premises (a famous illustration of this converse principle is Kyburg's (1965) lottery paradox, which is discussed in the entry on epistemic paradoxes of this encyclopedia). To put the matter more concretely, note that if a valid argument has three premises which each have uncertainty 1/11, then adding a premise which also has uncertainty 1/11 will not influence the argument's validity, but it will raise the upper bound on the conclusion's uncertainty from 3/11 to 4/11—thus allowing the conclusion to be more uncertain than was originally the case. Finally, the upper bound provided by Theorem 2 is optimal, in the sense that (under the right conditions) the uncertainty of the conclusion can coincide with its upper bound ∑U(γ):

Theorem 3. Consider a valid argument (Γ, ϕ), and assume that the premise set Γ is consistent, and that every premise γ ∈ Γ is relevant (i.e., Γ −{γ}⊭ϕ). Then there exists a probability function P : L→ [0,1] such that

UP(ϕ) = γ∈Γ UP(γ).

The upper bound provided by Theorem 2 can also be used to define a probabilistic notion of validity. An argument (Γ, ϕ) is said to be Adams-probabilistically valid, written Γ ⊧a ϕ, if and only if

for all probability functions P : L→ ℝ: UP(ϕ) ≤ γ∈Γ UP(γ).

Adams-probabilistic validity has an alternative, equivalent characterization in terms of probabilities rather than uncertainties. This characterization says that (Γ, ϕ) is Adams-probabilistically valid if and only if the conclusion's probability can get arbitrarily close to 1 if the premises' probabilities are sufficiently high. Formally: Γ ⊧a ϕ if and only if

for all ϵ > 0 there exists a δ > 0 such that for all probability functions P:

if P(γ) > 1 − δ for all γ ∈ Γ, then P(ϕ) > 1 − ϵ.

It can be shown that classical propositional logic is (strongly) sound and complete with respect to Adams' probabilistic semantics:

Γ ⊧a ϕ if and only if Γ ⊢ ϕ.

Adams (1998, 154) also defines another logic for which his probabilistic semantics is sound and complete. However, this system involves a non-truth-functional connective (the probability conditional), and therefore falls outside the scope of this section. (For more on probabilistic interpretations of conditionals, the reader can consult the entries on conditionals and the logic of conditionals of this encyclopedia.)

Consider the following example. The argument A with premises p, q, r, s and conclusion p ∧ (qr) is valid. Assume that P(p) = 10∕11, P(q) = P(r) = 9∕11 and P(s) = 7∕11. Then Theorem 2 says that

U(p ∧ (qr)) ≤ 1∕11 + 2∕11 + 2∕11 + 4∕11 = 9∕11.

This upper bound on the uncertainty of the conclusion is rather disappointing, and it exposes the main weakness of Theorem 2. One of the reasons why the upper bound is so high, is that to compute it we took into account the premise s, which has a rather high uncertainty (4∕11). However, this premise is irrelevant, in the sense that the conclusion already follows from the other three premises. Hence we can regard p ∧ (qr) not only as the conclusion of the valid argument A, but also as the conclusion of the (equally valid) argument A′, which has premises p, q, r. In the latter case Theorem 2 yields an upper bound of 1∕11 + 2∕11 + 2∕11 = 5∕11, which is already much lower.

The weakness of Theorem 2 is thus that it takes into account (the uncertainty of) irrelevant or inessential premises. To obtain an improved version of this theorem, a more fine-grained notion of “essentialness” is necessary. In argument A in the example above, premise s is absolutely irrelevant. Similarly, premise p is absolutely relevant, in the sense that without this premise, the conclusion p ∧ (qr) is no longer derivable. Finally, the premise subset {q, r} is “in between”: together q and r are relevant (if both premises are left out, the conclusion is no longer derivable), but each of them separately can be left out (while keeping the conclusion derivable).

The notion of essentialness is formalized as follows:

Essential premise set.
Given a valid argument (Γ, ϕ), a set Γ' ⊆ Γ is essential iff Γ − Γ′⊭ϕ.
Degree of essentialness.
Given a valid argument (Γ, ϕ) and a premise γ ∈ Γ, the degree of essentialness of γ, written E(γ), is 1 ∕ |Sγ|, where |Sγ| is the cardinality of the smallest essential premise set that contains γ. If γ does not belong to any minimal essential premise set, then the degree of essentialness of γ is 0.

With these definitions, a refined version of Theorem 2 can be established:

Theorem 4. Consider a valid argument (Γ, ϕ). Then the uncertainty of the conclusion ϕ cannot exceed the weighted sum of the uncertainties of the premises γ ∈ Γ, with the degrees of essentialness as weights. Formally:

U(ϕ) ≤ γ∈Γ E(γ)U(γ).

The proof of Theorem 4 is significantly more difficult than that of Theorem 2: Theorem 2 requires only basic probability theory, whereas Theorem 4 is proved using methods from linear programming (Adams and Levine 1975; Goldman and Tucker 1956). Theorem 4 subsumes Theorem 2 as a special case: if all premises are relevant (i.e., have degree of essentialness 1), then Theorem 4 yields the same upper bound as Theorem 2. Furthermore, Theorem 4 does not take into account irrelevant premises (i.e., premises with degree of essentialness 0) to compute this upper bound; hence if a premise is irrelevant for the validity of the argument, then its uncertainty will not carry over to the conclusion. Finally, note that since E(γ) ∈ [0,1] for all γ ∈ Γ, it holds that

γ∈Γ E(γ) U(γ) ≤ γ∈Γ U(γ),

i.e., Theorem 4 yields in general a tighter upper bound than Theorem 2. To illustrate this, consider again the argument with premises p, q, r, s and conclusion p ∧ (qr). Recall that P(p) = 10∕11, P(q) = P(r) = 9∕11 and P(s) = 7∕11. One can calculate the degrees of essentialness of the premises: E(p) = 1, E(q) = E(r) = 1∕2 and E(s) = 0. Hence Theorem 4 yields that

U(p ∧ (qr)) ≤ (1 × 1∕11) + (1∕2 × 2∕11) + (1∕2 × 2∕11) + (0 × 4∕11) = 3∕11,

which is a tighter upper bound for the uncertainty of p ∧ (qr) than any of the bounds obtained above via Theorem 2 (viz. 9∕11 and 5∕11).

2.3 Further Generalizations

Given the uncertainties (and degrees of essentialness) of the premises of a valid argument, Adams' theorems allow us to compute an upper bound for the uncertainty of the conclusion. Of course these results can also be expressed in terms of probabilities rather than uncertainties; they then yield a lower bound for the probability of the conclusion. For example, when expressed in terms of probabilities rather than uncertainties, Theorem 4 looks as follows:

P(ϕ) ≥ 1 − γ∈Γ E(γ)(1 − P(γ)).

Adams' results are restricted in at least two ways:

  • They only provide a lower bound for the probability of the conclusion (given the probabilities of the premises). In a sense this is the most important bound: it represents the conclusion's probability in the “worst-case scenario”, which might be useful information in practical applications. However, in some applications it might also be informative to have an upper bound for the conclusion's probability. For example, if one knows that this probability has an upper bound of 0.4, then one might decide to refrain from certain actions (that one would have performed if this upper bound were (known to be) 0.9).

  • They presuppose that the premises' exact probabilities are known. In practical applications, however, there might only be partial information about the probability of a premise γ: its exact value is not known, but it is known to have a lower bound a and an upper bound b (Walley 1991). In such applications it would be useful to have a method to calculate (optimal) lower and upper bounds for the probability of the conclusion in terms of the upper and lower bounds of the probabilities of the premises.

Hailperin (1965, 1984, 1986, 1996) and Nilsson (1986) use methods from linear programming to show that these two restrictions can be overcome. Their most important result is the following:

Theorem 5. Consider an argument (Γ, ϕ), with |Γ| = n. There exist functions LΓ, ϕ : ℝ2n → ℝ and UΓ, ϕ : ℝ2n → ℝ such that for any probability function P, the following holds: if aiPi) ≤ bi for 1 ≤ in, then:

  1. LΓ, ϕ(a1, …, an, b1, …, bn) ≤ P(ϕ) ≤ UΓ, ϕ(a1, …, an, b1, …, bn).
  2. The bounds in item 1 are optimal, in the sense that there exist probability functions PL and PU such that aiPLi), PUi) ≤ bi for 1 ≤ in, and LΓ, ϕ(a1, …, an, b1, …, bn) = PL(ϕ) and PU(ϕ) = UΓ, ϕ(a1, …, an, b1, …, bn).
  3. The functions LΓ, ϕ and UΓ, ϕ are effectively determinable from the Boolean structure of the sentences in Γ ∪ {ϕ}.

This result can also be used to define yet another probabilistic notion of validity, which we will call Hailperin-probabilistic validity or simply h-validity. This notion is not defined with respect to formulas, but rather with respect to pairs consisting of a formula and a subinterval of [0,1]. If Xi is the interval associated with premise γi ∈ Γ and Y is the interval associated with the conclusion ϕ, then the argument (Γ, ϕ) is said to be h -valid, written Γ ⊧h ϕ, if and only if

for all probability functions P : if Pi) ∈ Xi for 1 ≤ in, then P(ϕ) ∈ Y

In Haenni et al. (2011) this is written as

γ1 X1, …, γn Xn |≈ ϕY

and called the standard probabilistic semantics.

Nilsson's work on probabilistic logic (1986, 1993) has sparked a lot of research on probabilistic reasoning in artificial intelligence (Hansen and Jaumard 2000; chapter 2 of Haenni et al. 2011). However, it should be noted that although Theorem 5 states that the functions LΓ, ϕ and UΓ, ϕ are effectively determinable from the sentences in Γ ∪ {ϕ}, the computational complexity of this problem is quite high (Georgakopoulos et al. 1988, Kavvadias and Papadimitriou 1990), and thus finding these functions quickly becomes computationally unfeasible in real-world applications. Contemporary approaches based on probabilistic argumentation systems and probabilistic networks are better capable of handling these computational challenges. Furthermore, probabilistic argumentation systems are closely related to Dempster-Shafer theory (Dempster 1968; Shafer 1976; Haenni and Lehmann 2003). However, an extended discussion of these approaches is beyond the scope of (the current version of) this entry; see (Haenni et al. 2011) for a recent survey.

3. Basic Probability Operators

In this section we will study probability logics that extend the propositional language L with rather basic probability operators. Subsection 2.1 discusses qualitative probability operators; Subsection 2.2 discusses quantitative probability operators.

3.1 Qualitative Representations of Uncertainty

There are several applications in which qualitative theories of probability might be useful, or even necessary. In some situations there are no frequencies available to use as estimates for the probabilities, or it might be practically impossible to obtain those frequencies. Furthermore, people are often willing to compare the probabilities of two statements (‘ϕ is more probable than ψ’), without being able to assign explicit probabilities to each of the statements individually (Szolovits and Pauker 1978; Halpern and Rabin 1987). In such situations qualitative probability logics will be useful.

One of the earliest qualitative probability logics is Hamblin's (1959). The language is extended with a unary operator □, which is to be read as “probably”. Hence a formula such as □ϕ is to be read as “probably ϕ”. This notion of “probable” can be formalized as sufficiently high (numerical) probability (i.e., P(ϕ) ≥ t, for some threshold value 1∕2 < t ≤ 1), or alternatively in terms of plausibility, which is a non-metrical generalization of probability. Burgess (1969) further develops these systems, focusing on the “high numerical probability”-interpretation. Both Hamblin and Burgess introduce additional operators into their systems (expressing, for example, metaphysical necessity and/or knowledge), and study the interaction between the “probably”-operator and these other modal operators. However, the “probably”-operator already displays some interesting features on its own (independent from any other operators). If it is interpreted as “sufficiently high probability”, then it fails to satisfy the principle (□ϕ ∧□ψ) → □(ϕ ∧ ψ). This means that it is not a normal modal operator, and cannot be given a Kripke (relational) semantics. Herzig and Longin (2003) and Arló Costa (2005) provide weaker systems of neighborhood semantics for such “probably”-operators, while Yalcin (2010) discusses their behavior from a more linguistically oriented perspective.

Another route is taken by Segerberg (1971) and Gärdenfors (1975a,b). They introduce a binary operator ≥; the formula ϕ ≥ ψ is to be read as “ϕ is at least as probable as ψ” (formally: P(ϕ) ≥ P(ψ)). The key idea is that one can completely characterize the behavior of ≥ without having to use the “underlying” probabilities of the individual formulas. Finally, it should be noted that with comparative probability (a binary operator), one can also express some absolute probabilistic properties (unary operators). For example, ϕ ≥ ⊤ expresses that ϕ has probability 1, and ϕ ≥ ¬ϕ expresses that ϕ has probability at least 1/2.

3.2 Expressing Properties of Probabilities with and without Linear Combinations

The semantics of propositional probability logic involves a probability function P, satisfying certain properties. Here we consider P as an operator in the object language. Such a language might simply add probability formulas such as P(ϕ) ≥ q, where ϕ is a propositional formula, to propositional logic. But note that one of the conditions for a probability function is additivity: P(ϕ ∨ ψ) = P(ϕ) + P(ψ) whenever ¬(ϕ ∧ ψ) is a tautology. It is thus natural to involve addition (and more generally, linear combinations) in a probability language with probability operators. But we will see that much can be expressed without linear combinations explicitly in the language.

It is often desirable to have as few definitions as primitive and to generate further definitions from the primitive definitions. This allows us to specify the language more concisely. Let us first look at what can be expressed using linear combinations of a basic primitive form. Take as primitive formulas of the form a1P1) + ⋯ + anPn) ≥ b, where n is a positive integer that may differ from formula to formula, and a1, …, an, and b are all rational numbers. Here are some examples of what can be expressed.

  • P(ϕ) ≤ q by − P(ϕ) ≥−q,
  • P(ϕ) < q by ¬(P(ϕ) ≥ q),
  • P(ϕ) = q by P(ϕ) ≥ qP(ϕ) ≤ q.

Now consider the language restricted to formulas of the form P(ϕ) ≥ q for some propositional formula ϕ and rational q. Note that we do not even consider coefficients of the probability term. We can define

  • P(ϕ) ≤ q by P(¬ϕ) ≥ 1 − q,

which is reasonable considering that the probability of the complement of a proposition is equal to 1 minus the probability of the proposition. The formulas P(ϕ) < q and P(ϕ) = q can be defined without linear combinations as we did above. Using this restricted probability language, we can reason about additivity in a less direct way. The formula

[P(ϕ ∧ ψ) = aP(ϕ ∧¬ψ) = b] → P(ϕ) = a + b

states that if the probability of ϕ ∧ ψ is a and the probability of ϕ ∧¬ψ is b, then the probability of the disjunction of the formulas (ϕ ∧ ψ) ∨ (ϕ ∧¬ψ)—which is equivalent to ϕ—is a + b. However, while the use of linear combinations allows us to assert that the probabilities of ϕ ∧ ψ and ϕ ∧¬ψ are additive by using the formula P(ϕ ∧ ψ) + P(ϕ ∧¬ψ) = P(ϕ), the formula without linear combinations above only does so if we choose the correct numbers a and b.

In Fagin et al. (1990), a sound and complete proof system is given for a logic involving linear combinations, where axioms are given for linear combinations. In Heifetz and Mongin (2001), a sound and complete proof system is given for a logic without linear combinations, where additivity is broken up into implications

[P(ϕ ∧ ψ) ≥ aP(ϕ ∧¬ψ) ≥ b] → P(ϕ) ≥ a + b,

stating that the probability of the union of two disjoint sets is at least the sum of lower bounds of the probabilities of each of the sets, and

[P(ϕ ∧ ψ) < aP(ϕ ∧¬ψ) < b] → P(ϕ) < a + b.

Both of these logics lack the compactness property; for example, every finite subset of {P(p) > 0} ∪ {P(p) ≤ a | a > 0} is satisfiable, but the entire set is not.

4. Modal Probability Logics

Many probability logics are interpreted over a single, but arbitrary probability space. Modal probability logic makes use of many probability spaces, each associated with a possible world or state. This can be viewed as a minor adjustment to the relational semantics of modal logic: rather than associate to every possible world a set of accessible worlds as is done in modal logic, modal probability logic associates to every possible world a probability distribution, a probability space, or a set of probability distributions. The language of modal probability logic allows for embedding of probabilities within probabilities, that is, it can for example reason about the probability that (possibly a different) probability is 1∕2. This modal setting involving multiple probabilities has generally been given a (1) stochastic interpretation, concerning different probabilities over the next states a system might transition into (Larsen and Skou 1991), and (2) a subjective interpretation, concerning different probabilities that different agents may have about a situation or each other's probabilities (Fagin and Halpern 1988). Both interpretations can use exactly the same formal framework.

A basic modal probability logic adds to propositional logic formulas of the form P(ϕ) ≥ q, where q is typically a rational number, and ϕ is any formula of the language, possibly a probability formula. The reading of such a formula is that the probability of ϕ is at least q. This general reading of the formula does not reflect any difference between modal probability logic and other probability logics with the same formula; where the difference lies is in the ability to embed probabilities in the arguments of probability terms and in the semantics. The following subsections provide an overview of the variations of how modal probability logic is modeled. In one case the language is altered slightly (Subsection 4.2), and in other cases, the logic is extended to address interactions between qualitative and quantitative uncertainty (Subsection 4.4) or dynamics (Subsection 4.5).

4.1 Basic Finite Modal Probability Models

Formally, a Basic Finite Modal Probabilistic Model is a tuple M = (W, P, V), where W is a finite set of possible worlds or states, P is a function associating a distribution Pw over W to each world wW, and V is a “valuation function” assigning atomic propositions from a set Φ to each world. The distribution is additively extended from individual worlds to sets of worlds: Pw(S) = ∑sSPw(s). The first two components of a basic modal probabilistic model are effectively the same as a Kripke frame whose relation is decorated with numbers (probability values). Such a structure has different names, such as a directed graph with labelled edges in mathematics, or a probabilistic transition system in computer science. The valuation function, as in a Kripke model, allows us to assign properties to the worlds.

The semantics for formulas are given on pairs (M,w), where M is a model and w is an element of the model. A formula P(ϕ) ≥ q is true at a pair (M,w), written (M,w) ⊧ P(ϕ) ≥ q, if and only if Pw({w′ | (M,w′) ⊧ ϕ}) ≥ q.

4.2 Indexing and Interpretations

The first generalization, which is most common in applications of modal probabilistic logic, is to allow the distributions to be indexed by two sets rather than one. The first set is the set W of worlds (the base set of the model), but the other is an index set A often to be taken as a set of actions, agents, or players of a game. Formally, P associates a distribution Pa,w over W for each wW and aA. For the language, rather than involving formulas of the form P(ϕ) ≥ q, we have Pa(ϕ) ≥ q, and (M,w) ⊧ Pa(ϕ) ≥ q if and only if Pa,w({w′ | (M,w′) ⊧ ϕ}) ≥ q.

Example: Suppose we have an index set A = {a,b}, and a set Φ = {p,q} of atomic propositions. Consider (W, P, V), where

  • W = {w, x, y, z}
  • Pa,w and Pa,x map w to 1∕2, x to 1∕2, y to 0, and z to 0.
    Pa,y and Pa,z map y to 1∕3, z to 2∕3, w to 0, and x to 0.
    Pb,w and Pb,y map w to 1∕2, y to 1∕2, x to 0, and z to 0.
    Pb,x and Pb,z map x to 1∕4, z to 3∕4, w to 0, and y to 0.
  • V (p) = {w,x}
    V (q) = {w,y}.

We depict this example with the following diagram. Inside each circle is a labeling of the truth of each proposition letter for the world whose name is labelled right outside the circle. The arrows indicate the probabilities. For example, an arrow from world x to world z labeled by (b,3∕4) indicates that from x, the probably of z under label b is 3∕4. Probabilities of 0 are not labelled.

[a diagram of four worlds (w[p,q], x[p,-q], y[-p,q], z[-p,-q]) with w and x connected by a double arrowed line[(a,1/2)], with w and y connected by a double arrowed line[(b,1/2)], with an arrow[(b,3/4)] going from x to z and another[(b,1/4)] back from z to x, with an arrow[(a,2/3)] going from y to z and another[(a,1/3)] back from z to y, arrows self-looping to w are labelled (a,1/2) and (b,1/2), arrows self-looping to x are labelled (a,1/2) and (b,1/4), arrows self-looping to y are labelled (a,1/3) and (b,1/2), arrows self-looping to z are labelled (a,2/3) and (b,3/4).]

Stochastic Interpretation: Consider the elements a and b of A to be actions, for example, pressing buttons on a machine. In this case, pressing a button does not have a certain outcome. For instance, if the machine is in state x, there is a 1∕2 probability it will remain in the same state after pressing a, but a 1∕4 probability of remaining in the same state after pressing b. That is,

(M,x) ⊧ Pa(p ∧¬q) = 1∕2 ∧ Pb(p ∧¬q) = 1∕4.

A significant feature of modal logics in general (and this includes modal probabilistic logic) is the ability to support higher-order reasoning, that is, the reasoning about probabilities of probabilities. The importance of higher-order probabilities is clear from the role they play in, for example, Miller's principle, which states that P1(ϕ | P2(ϕ) = b) = b. Here, P1 and P2 are probability functions, which can have various interpretations, such as the probabilities of two agents, logical and statistical probability, or the probabilities of one agent at different moments in time (Miller 1966; Lewis 1980; van Fraassen 1984; Halpern 1991). Higher-order probability also occurs for instance in the Judy Benjamin Problem (van Fraassen 1981a) where one conditionalizes on probabilistic information. Whether one agrees with the principles proposed in the literature on higher-order probabilities or not, the ability to represent them forces one to investigate the principles governing them.

To illustrate higher-order reasoning more concretely, we return to our example and see that at x, there is a 1∕2 probability that after pressing a, there is a 1∕2 probability that after pressing b, it will be the case that ¬p is true, that is,

(M,x) ⊧ Pa(Pbp) = 1∕2) = 1∕2.

Subjective Interpretation: Suppose the elements a and b of A are players of a game. p and ¬p are strategies for player a and q and ¬q are both strategies for player b. In the model, each player is certain of her own strategy; for instance at x, player a is certain that she will play p and player b is certain that she will play ¬q, that is

(M,x) ⊧ Pa(p) = 1 ∧ Pbq) = 1.

But the players randomize over their opponents. For instance at x, the probability that b has for a's probability of ¬q being 1∕2 is 1∕4, that is

(M,x) ⊧ Pb(Pa(q) = 1∕2) = 1∕4.

4.3 Probability Spaces

Probabilities are generally defined as measures in a measure space. A measure space is a set Ω (the sample space) together with a σ-algebra (also called σ-field) A over Ω, which is a non-empty set of subsets of Ω such that AA implies that Ω − AA, and AiA for all natural numbers i, implies that iAiA. A measure is a function μ defined on the σ-algebra A, such that μ(A) ≥ 0 for every set AA and μ(⋃iAi) = ∑i μ(Ai) whenever AiAj = ∅ for each i, j.

The effect of the σ-algebra is to restrict the domain so that not every subset of Ω need have a probability. This is crucial for some probabilities to be defined on uncountably infinite sets; for example, a uniform distribution over a unit interval cannot be defined on all subsets of the interval while also maintaining the countable additivity condition for probability measures.

The same basic language as was used for the basic finite probability logic need not change, but the semantics is slightly different: for every state wW, the component Pw of a modal probabilistic model is replaced by an entire probability space (Ωw, Aw, μw), such that ΩwW and Aw is a σ-algebra over Ωw. The reason we may want entire spaces to differ from one world to another is to reflect uncertainty about what probability space is the right one. For the semantics of probability formulas, (M,w) ⊧ P(ϕ) ≥ q if and only if μw({w′ | (M,w′) ⊧ ϕ}) ≥ q. Such a definition is not well defined in the event that {w′ | (M,w′) ⊧ ϕ} ∉ Aw. Thus constraints are often placed on the models to ensure that such sets are always in the σ-algebras.

4.4 Combining Quantitative and Qualitative Uncertainty

Although probabilities reflect quantitative uncertainty at one level, there can also be qualitative uncertainty about probabilities. We might want to have qualitative and quantitative uncertainty because we may be so uncertain about some situations that we do not want to assign numbers to the probabilities of their events, while there are other situations where we do have a sense of the probabilities of their events; and these situations can interact.

There are many situations in which we might not want to assign numerical values to uncertainties. One example is where a computer selects a bit 0 or 1, and we know nothing about how this bit is selected. Results of coin flips, on the other hand, are often used examples of where we would assign probabilities to individual outcomes.

An example of how these might interact is where the result of the bit determines whether a fair coin or a weighted coin (say, heads with probability 2∕3) be used for a coin flip. Thus there is qualitative uncertainty as to whether the action of flipping a coin yields heads with probability 1∕2 or 2∕3.

One way to formalize the interaction between probability and qualitative uncertainty is by adding another relation to the model and a modal operator to the language as is done in Fagin and Halpern (1988, 1994). Formally, we add to a basic finite probability model a relation R ⊆ W2. Then we add to the language a modal operator □, such that (M,w) ⊧ □ϕ if and only if (M,w′) ⊧ ϕ whenever wRw′.

Consider the following example:

  • W = {(0,H), (0,T), (1,H), (1,T)},
  • Φ = {h,t} is the set of atomic propositions,
  • R = W2,
  • P associates with (0,H) and (0,T) the distribution mapping (0,H) and (0,T) each to 1∕2, and associates with (1,H) and (1,T) the distribution mapping (1,H) to 2∕3 and (1,T) to 1∕3,
  • V maps h to the set {(0,H), (1,H)} and t to the set {(0,T), (1,T)}.

Then the following formula is true at (0,H): ¬□h (¬□P(h) = 1∕2) ∧ (♢P(h) = 1∕2). This can be read as it is not known that h is true, and it is not known that the probability of h is 1∕2, but it is possible that the probability of h is 1∕2.

4.5 Dynamics

We have discussed two views of modal probability logic. One is temporal or stochastic, where the probability distribution associated with each state determines the likelihood of transitioning into other states; another is concerned with subjective perspectives of agents, who may reason about probabilities of other agents. A stochastic system is dynamic in that it represents probabilities of different transitions, and this can be conveyed by the modal probabilistic models themselves. But from a subjective view, the modal probabilistic models are static: the probabilities are concerned with what currently is the case. Although static in their interpretation, the modal probabilistic setting can be put in a dynamic context.

Dynamics in a modal probabilistic setting is generally concerned with simultaneous changes to probabilities in potentially all possible worlds. Intuitively, such a change may be caused by new information that invokes a probabilistic revision at each possible world. The dynamics of subjective probabilities is often modeled using conditional probabilities, such as in Kooi (2003), Baltag and Smets (2008), and van Benthem et al. (2009). The probability of E conditional on F, written P(E | F), is P(EF) ∕ P(F). When updating by a set F, a probability distribution P is replaced by the probability distribution P′, such that P′(E) = P(E | F), so long as P(F) ≠ 0. Let us assume for the rest of this dynamics section that every relevant set considered has positive probability.

Using a probability logic with linear combinations, we can abbreviate the conditional probability P(ϕ | ψ) ≥ q by P(ϕ ∧ ψ) − qP(ψ) ≥ 0. In a modal setting, an operator [!ψ] can be added to the language, such that M,w ⊧ [!ψ]ϕ if and only if M′,w ⊧ ϕ, where M′ is the model obtained from M by revising the probabilities of each world by ψ. Note that [!ψ](P(ϕ) ≥ q) differs from P(ϕ | ψ) ≥ q, in that in [!ψ](P(ϕ) ≥ q), the interpretation of probability terms inside ϕ are affected by the revision by ψ, whereas in P(ϕ | ψ) ≥ q, they are not, which is why P(ϕ | ψ) ≥ q nicely unfolds into another probability formula. However, [!ψ]ϕ does unfold too, but in more steps:

[!ψ](P(ϕ) ≥ q) ↔ (ψ → P([!ψ] ϕ | ψ) ≥ q).

5 First-order Probability Logic

In this section we will discuss first-order probability logics. As was explained in Section 1 of this entry, there are many ways in which a logic can have probabilistic features. The models of the logic can have probabilistic aspects, the notion of consequence can have a probabilistic flavor, or the language of the logic can contain probabilistic operators. In this section we will focus on those logical operators that have a first-order flavor. The first-order flavor is what distinguishes these operators from the probabilistic modal operators of the previous section.

Consider the following example from Bacchus (1990)

More than 75% of all birds fly.

There is a straightforward probabilistic interpretation of this sentence, namely when one randomly selects a bird, then the probability that the selected bird flies is more than 3/4. First-order probabilistic operators are needed to express these sort of statements.

5.1 An Example of a First-order Probability Logic

In this section we will have a closer look at a particular first-order probability logic, whose language is as simple as possible, in order to focus on the probabilistic quantifiers. The language is very much like the language of classical first-order logic, but rather than the familiar universal and existential quantifier, the language contains a probabilistic quantifier.

The language is built on a set of individual variables (denoted by x, y, z, x1, x2, …), a set of function symbols (denoted by f, g, h, f1, …) where an arity is associated with each symbol (nullary function symbols are also called individual constants), and a set of predicate letters (denoted by R, P1, …) where an arity is associated with each symbol. The language contains two kinds of syntactical objects, namely terms and formulas. The terms are defined inductively as follows:

  • Every individual variable x is a term.
  • Every function symbol f of arity n followed by an n-tuple of terms (t1, …, tn) is a term.

Given this definition of terms, the formulas are defined inductively as follows:

  • Every predicate letter R of arity n followed by an n-tuple of terms (t1, …, tn) is a formula.
  • If ϕ is a formula, then so is ¬ϕ.
  • If ϕ and ψ are formulas, then so is (ϕ ∧ ψ).
  • If ϕ is a formula and q is a rational number in the interval [0,1], then so is Px(ϕ) ≥ q.

Formulas of the form Px(ϕ) ≥ q should be read as: “the probability of selecting an x such that x satisfies ϕ is at least q”. The formula Px(ϕ) ≤ q is an abbreviation of Px(¬ϕ) ≥ 1 − q and Px(ϕ) = q is an abbreviation of Px(ϕ) ≥ qPx(ϕ) ≤ q. Every free occurrence of x in ϕ is bound by the operator.

This language is interpreted on very simple first-order models, which are triples M = (D, I, P), where the domain of discourse D is a finite nonempty set of objects, the interpretation I associates an n-ary function on D with every n-ary function symbol occurring in the language, and an n-ary relation on D with every n-ary predicate letter. P is a probability function that assigns a probability P(d) to every element d in D such that dD P(d) = 1.

In order to interpret formulas containing free variables one also needs an assignment g which assigns an element of D to every variable. The interpretation [[t]]M,g of a term t given a model M = (D,I,P) and an assignment g is defined inductively as follows:

  • [[x]]M,g = g(x)
  • [[f(t1, …, tn)]]M,g = I(f)([[t1]], …, [[tn]])

Truth is defined as a relation ⊧ between models with assignments and formulas:

  • M, gR(t1, …, tn) iff ([[t1]], …, [[tn]]) ∈ I(R)
  • M, g ⊧ ¬ϕ iff M, g ⊭ϕ
  • M, g ⊧ (ϕ ∧ ψ) iff M, g ⊧ ϕ and M, g ⊧ ψ
  • M, gPx(ϕ) ≥ q iff ∑d:M,g[x↦d] ⊧ ϕ P(d) ≥ q

As an example, consider a model of a vase containing nine marbles: five are black and four are white. Let us assume that P assigns a probability of 1/9 to each marble, which captures the idea that one is equally likely to pick any marble. Suppose the language contains a unary predicate B whose interpretation is the set of black marbles. The sentence Px(B(x)) = 5∕9 is true in this model regardless of the assignment.

5.2 The Need for Extensions

The logic presented in the previous section is too simple to capture many forms of reasoning about probabilities. We will discuss three extensions here.

5.2.1 Quantifying over More than One Variable

First of all one would like to reason about cases where more than one object is selected from the domain. Consider for example the probability of first picking a black marble, putting it back, and then picking a white marble from the vase. This probability is 5/9 × 4/9 = 20/81, but we cannot express this in the language above. For this we need one operator that deals with multiple variables simultaneously, written as Px1,…xn(ϕ) ≥ q. The semantics for such operators will then have to provide a probability measure on subsets of Dn. The simplest way to do this is by simply taking the product of the probability function P on D, which can be taken as an extension of P to tuples, where P(d1,…dn) = P(d1) ×⋯ × P(dn), which yields the following semantics:

M, gPx1xn(ϕ) ≥ q iff ∑(d1,…, dn):M,g[x1d1,…, xndn] ⊧ ϕ P(d1,…, dn) ≥ q

This approach is taken by Bacchus (1990) and Halpern (1990), corresponding to the idea that selections are independent and with replacements. With these semantics the example above can be formalized as Px,y(B(x) ∧¬B(y)) = 20∕81. There are also more general approaches to extending the measure on the domain to tuples from the domain such as by Hoover (1978) and Keisler (1985).

5.2.2 Conditional Probability

When one considers the initial example that more than 75% of all birds fly, one finds that this cannot be adequately captured in a model where the domain contains objects that are not birds. These objects should not matter to what one wishes to express, but the probability quantifiers, quantify over the whole domain. In order to restrict quantification one must add conditional probability operators Px(ϕ | ψ) ≥ q with the following semantics:

M, gPx(ϕ | ψ) ≥ q iff if there is a dD such that M, g[xd] ⊧ ψ then (∑d:M,g[xd] ⊧ ϕ∧ψ P(d) ∕ d:M,g[xd] ⊧ ψ P(d)) ≥ q.

With these operators, the formula Px(F(x) | B(x)) > 3∕4 expresses that more than 75% of all birds fly.

5.2.3 Probabilities as Terms

When one wants to compare the probability of different events, say of selecting a black ball and selecting a white ball, it may be more convenient to consider probabilities to be terms in their own right. That is, an expression Px(ϕ) is interpreted as referring to some rational number. Then one can extend the language with arithmetical operations such as addition and multiplication, and with operators such as equality and inequalities to compare probability terms. One can then say that one is twice as likely to select a black ball compared to a white ball as Px(B(x)) = 2 × Px(W(x)). Such an extension requires that the language contains two separate classes of terms: one for probabilities, numbers and the results of arithmetical operations on such terms, and one for the domain of discourse which the probabilistic operators quantify over. We will not present such a language and semantics in detail here. One can find such a system in Bacchus (1990).

5.3 Metalogic

Generally it is hard to provide proof systems for first-order probability logics, because the validity problem for these logics is generally undecidable. It is even not the case, as it is the case in classical first-order logic, that if an inference is valid, then one can find out in finite time (see Abadi and Halpern (1994)).

Nonetheless there are many results for first-order probability logic. For instance, Hoover (1978) and Keisler (1985) study completeness results. Bacchus (1990) and Halpern (1990) also provide complete axiomatizations as well as combinations of first-order probability logics and modal probability logics.

Bibliography

  • Abadi, M. and J. Y. Halpern, 1994, “Decidability and Expressiveness for First-Order Logics of Probability,” Information and Computation, 112: 1–36.
  • Adams, E. W. and H. P. Levine, 1975, “On the Uncertainties Transmitted from Premisses to Conclusions in Deductive Inferences,” Synthese, 30: 429–460.
  • Adams, E. W., 1998, A Primer of Probability Logic, Stanford, CA: CSLI Publications.
  • Arló Costa, H., 2005, “Non-Adjunctive Inference and Classical Modalities,” Journal of Philosophical Logic, 34: 581–605.
  • Bacchus, F., 1990, Representing and Reasoning with Probabilistic Knowledge, Cambridge, MA: The MIT Press.
  • Baltag, A. and S. Smets, 2008, “Probabilistic Dynamic Belief Revision,” Synthese, 165: 179–202.
  • van Benthem, J., J. Gerbrandy, and B. Kooi, 2009, “Dynamic Update with Probabilities,” Studia Logica, 93: 67–96.
  • Boole, G., 1854, An Investigation of the Laws of Thought, on which are Founded the Mathematical Theories of Logic and Probabilities, London: Walton and Maberly.
  • Burgess, J., 1969, “Probability Logic,” Journal of Symbolic Logic, 34: 264–274.
  • Carnap, R., 1950, Logical Foundations of Probability, Chicago, IL: University of Chicago Press.
  • Cross, C., 1993, “From Worlds to Probabilities: A Probabilistic Semantics for Modal Logic,” Journal of Philosophical Logic, 22: 169–192.
  • Dempster, A., 1968, “A Generalization of Bayesian Inference,” Journal of the Royal Statistical Society, 30: 205–247.
  • De Morgan, A., 1847, Formal Logic, London: Taylor and Walton.
  • de Finetti, B., 1937, “La Prévision: Ses Lois Logiques, Ses Sources Subjectives”, Annales de l'Institut Henri Poincaré, 7: 168; translated as “Foresight. Its Logical Laws, Its Subjective Sources,” in Studies in Subjective Probability, H. E. Kyburg, Jr. and H. E. Smokler (eds.), Malabar, FL: R. E. Krieger Publishing Company, 1980, pp. 53–118.
  • Eagle, A., 2010, Philosophy of Probability: Contemporary Readings, London: Routledge.
  • Fagin, R. and J. Y. Halpern, 1988, “Reasoning about Knowledge and Probability,” in Proceedings of the 2nd conference on Theoretical aspects of reasoning about knowledge, M. Y. Vardi (ed.), Pacific Grove, CA: Morgan Kaufmann, pp. 277–293.
  • –––, 1994, “Reasoning about Knowledge and Probability,” Journal of the ACM, 41: 340–367.
  • Fagin, R., J. Y. Halpern, and N. Megiddo, 1990, “A Logic for Reasoning about Probabilities,” Information and Computation, 87: 78–128.
  • Fitelson, B., 2006, “Inductive Logic,” in The Philosophy of Science: An Encyclopedia, J. Pfeifer and S. Sarkar (eds.), New York, NY: Routledge, pp. 384–394.
  • Gärdenfors, P., 1975a, “Qualitative Probability as an Intensional Logic,” Journal of Philosophical Logic, 4: 171–185.
  • –––, 1975b, “Some Basic Theorems of Qualitative Probability,” Studia Logica, 34: 257–264.
  • Georgakopoulos, G., D. Kavvadias, and C. H. Papadimitriou, 1988, “Probabilistic Satisfiability,” Journal of Complexity, 4: 1–11.
  • Gerla, G., 1994, “Inferences in Probability Logic,” Artificial Intelligence, 70: 33–52.
  • Gillies, D., 2000, Philosophical Theories of Probability, London: Routledge.
  • Goldman, A. J. and A. W. Tucker, 1956, “Theory of Linear Programming,” in Linear Inequalities and Related Systems. Annals of Mathematics Studies 38, H. W. Kuhn and A. W. Tucker (eds.), Princeton: Princeton University Press, pp. 53–98.
  • Goosens, W. K., 1979, “Alternative Axiomatizations of Elementary Probability Theory,” Notre Dame Journal of Formal Logic, 20: 227–239.
  • Hájek, A., 2001, “Probability, Logic, and Probability Logic,” in The Blackwell Guide to Philosophical Logic, L. Goble (ed.), Oxford: Blackwell, pp. 362–384.
  • Hájek, A. and S. Hartmann, 2010, “Bayesian Epistemology,” in A Companion to Epistemology, J. Dancy, E. Sosa, and M. Steup (eds.), Oxford: Blackwell, pp. 93–106.
  • Haenni, R. and N. Lehmann, 2003, “Probabilistic Argumentation Systems: a New Perspective on Dempster-Shafer Theory,” International Journal of Intelligent Systems, 18: 93–106.
  • Haenni, R., J.-W. Romeijn, G. Wheeler, and J. Williamson, 2011, Probabilistic Logics and Probabilistic Networks, Dordrecht: Springer.
  • Hailperin, T., 1965, “Best Possible Inequalities for the Probability of a Logical Function of Events,” American Mathematical Monthly, 72: 343–359.
  • –––, 1984, “Probability Logic,” Notre Dame Journal of Formal Logic, 25: 198–212.
  • –––, 1986, Boole's Logic and Probability, Amsterdam: North-Holland.
  • –––, 1996, Sentential Probability Logic: Origins, Development, Current Status, and Technical Applications, Bethlehem, PA: Lehigh University Press.
  • Halpern, J. Y. and M. O. Rabin, 1987, “A Logic to Reason about Likelihood”, Artificial Intelligence, 32: 379–405.
  • Halpern, J. Y., 1990, “An analysis of first-order logics of probability”, Artificial Intelligence, 46: 311–350.
  • –––, 1991, “The Relationship between Knowledge, Belief, and Certainty,” Annals of Mathematics and Artificial Intelligence, 4: 301–322. Errata appeared in Annals of Mathematics and Artificial Intelligence, 26 (1999): 59–61.
  • –––, 2003, Reasoning about Uncertainty, Cambridge, MA: The MIT Press.
  • Hamblin, C.L., 1959, “The modal ‘probably’”, Mind, 68: 234–240.
  • Hansen, P. and B. Jaumard, 2000, “Probabilistic Satisfiability,” in Handbook of Defeasible Reasoning and Uncertainty Management Systems. Volume 5: Algorithms for Uncertainty and Defeasible Reasoning, J. Kohlas and S. Moral (eds.), Dordrecht: Kluwer, pp. 321–367.
  • Hartmann, S. and J. Sprenger, 2010, “Bayesian Epistemology,” in Routledge Companion to Epistemology, S. Bernecker and D. Pritchard (eds.), London: Routledge, pp. 609–620.
  • Heifetz, A. and P. Mongin, 2001, “Probability Logic for Type Spaces”, Games and Economic Behavior, 35: 31–53.
  • Herzig, A. and D. Longin, 2003, “On Modal Probability and Belief,” in Proceedings of the 7th European Conference on Symbolic and Quantitative Approaches to Reasoning with Uncertainty (ECSQARU 2003), T.D. Nielsen and N.L. Zhang (eds.), Lecture Notes in Computer Science 2711, Berlin: Springer, pp. 62–73.
  • Hoover, D. N., 1978, “Probability Logic,” Annals of Mathematical Logic, 14: 287–313.
  • Howson, C., 2003, “Probability and Logic,” Journal of Applied Logic, 1: 151–165.
  • –––, 2007, “Logic with Numbers,” Synthese, 156: 491–512.
  • –––, 2009, “Can Logic be Combined with Probability? Probably,” Journal of Applied Logic, 7: 177–187.
  • Jeffrey, R., 1992, Probability and the Art of Judgement, Cambridge: Cambridge University Press.
  • Jonsson, B., K. Larsen, and W. Yi, 2001 “Probabilistic Extensions of Process Algebras,” in Handbook of Process Algebra, J. A. Bergstra, A. Ponse, and S. A. Smolka (eds.), Amsterdam: Elsevier, pp. 685–710.
  • Kavvadias, D. and C. H. Papadimitriou, 1990, “A Linear Programming Approach to Reasoning about Probabilities,” Annals of Mathematics and Artificial Intelligence, 1: 189–205.
  • Keisler, H. J., 1985, “Probability Quantifiers,” in Model-Theoretic Logics, J. Barwise and S. Feferman (eds.), New York, NY: Springer, pp. 509–556.
  • Kooi B. P., 2003, “Probabilistic Dynamic Epistemic Logic,” Journal of Logic, Language and Information, 12: 381–408.
  • Kyburg, H. E., 1965, “Probability, Rationality, and the Rule of Detachment,” in Proceedings of the 1964 International Congress for Logic, Methodology, and Philosophy of Science, Y. Bar-Hillel (ed.), Amsterdam: North-Holland, pp. 301–310.
  • –––, 1994, “Uncertainty Logics, ” in Handbook of Logic in Artificial Intelligence and Logic Programming, D. M. Gabbay, C. J. Hogger, and J. A. Robinson (eds.), Oxford: Oxford University Press, pp. 397–438.
  • Larsen, K. and A. Skou, 1991, “Bisimulation through Probabilistic Testing,” Information and Computation, 94: 1–28.
  • Leblanc, H., 1979, “Probabilistic Semantics for First-Order Logic,” Zeitschrift für mathematische Logik und Grundlagen der Mathematik, 25: 497–509.
  • –––, 1983, “Alternatives to Standard First-Order Semantics,” in Handbook of Philosophical Logic, Volume I, D. Gabbay and F. Guenthner (eds.), Dordrecht: Reidel, pp. 189–274.
  • Lewis, D., 1980, “A Subjectivist's Guide to Objective Chance,” in Studies in Inductive Logic and Probability. Volume 2, R. C. Jeffrey (ed.), Berkeley, CA: University of California Press, pp. 263–293; reprinted in Philosophical Papers. Volume II, Oxford: Oxford University Press, 1987, pp. 83–113.
  • Miller, D., 1966, “A Paradox of Information,” British Journal for the Philosophy of Science, 17: 59–61.
  • Morgan, C., 1982a, “There is a Probabilistic Semantics for Every Extension of Classical Sentence Logic,” Journal of Philosophical Logic, 11: 431–442.
  • –––, 1982b, “Simple Probabilistic Semantics for Propositional K, T, B, S4, and S5,” Journal of Philosophical Logic, 11: 443–458.
  • –––, 1983, “Probabilistic Semantics for Propositional Modal Logics”. in Essays in Epistemology and Semantics, H. Leblanc, R. Gumb, and R. Stern (eds.), New York, NY: Haven Publications, pp. 97–116.
  • Morgan, C. and H. Leblanc, 1983, “Probabilistic Semantics for Intuitionistic Logic,” Notre Dame Journal of Formal Logic, 24: 161–180.
  • Nilsson, N., 1986, “Probabilistic Logic,” Artificial Intelligence, 28: 71–87.
  • –––, 1993, “Probabilistic Logic Revisited,” Artificial Intelligence, 59: 39–42.
  • Paris, J. B., 1994, The Uncertain Reasoner's Companion, A Mathematical Perspective, Cambridge: Cambridge University Press.
  • Parma, A. and R. Segala, 2007, “Logical Characterizations of Bisimulations for Discrete Probabilistic Systems,” in Proceedings of the 10th International Conference on Foundations of Software Science and Computational Structures (FOSSACS), H. Seidl (ed.), Lecture Notes in Computer Science 4423, Berlin: Springer, pp. 287–301.
  • Pearl, J., 1991, “Probabilistic Semantics for Nonmonotonic Reasoning,” in Philosophy and AI: Essays at the Interface, R. Cummins and J. Pollock (eds.), Cambridge, MA: The MIT Press, pp. 157–188.
  • Ramsey, F. P., 1926, “Truth and Probability”, in Foundations of Mathematics and other Essays, R. B. Braithwaite (ed.), London: Routledge and Kegan Paul, 1931, pp. 156–198; reprinted in Studies in Subjective Probability, H. E. Kyburg, Jr. and H. E. Smokler (eds.), 2nd ed., Malabar, FL: R. E. Krieger Publishing Company, 1980, pp. 23–52; reprinted in Philosophical Papers, D. H. Mellor (ed.) Cambridge: Cambridge University Press, 1990, pp. 52–94.
  • Reichenbach, H., 1949, The Theory of Probability, Berkeley, CA: University of California Press.
  • Romeijn, J.-W., 2011, “Statistics as Inductive Logic,” in Handbook for the Philosophy of Science. Vol. 7: Philosophy of Statistics, P. Bandyopadhyay and M. Forster (eds.), Amsterdam: Elsevier, pp. 751–774.
  • Segerberg, K., 1971, “Qualitative Probability in a Modal Setting”, in Proceedings 2nd Scandinavian Logic Symposium, E. Fenstad (ed.), Amsterdam: North-Holland, pp. 341–352.
  • Shafer, G., 1976, A Mathematical Theory of Evidence, Princeton, NJ: Princeton University Press.
  • Suppes, P., 1966, “Probabilistic Inference and the Concept of Total Evidence,” in Aspects of Inductive Logic, J. Hintikka and P. Suppes (eds.), Amsterdam: Elsevier, pp. 49–65.
  • Szolovits, P. and S. G. Pauker, 1978, “Categorical and Probabilistic Reasoning in Medical Diagnosis,” Artificial Intelligence, 11: 115–144.
  • Tarski, A., 1936, “Wahrscheinlichkeitslehre und mehrwertige Logik”, Erkenntnis, 5: 174–175.
  • Van Fraassen, B., 1981a, “A Problem for Relative Information Minimizers in Probability Kinematics,” British Journal for the Philosophy of Science, 32:375–379.
  • –––, 1981b, “Probabilistic Semantics Objectified: I. Postulates and Logics,” Journal of Philosophical Logic, 10: 371–391.
  • –––, 1983, “Gentlemen's Wagers: Relevant Logic and Probability,” Philosophical Studies, 43: 47–61.
  • –––, 1984, “Belief and the Will,” Journal of Philosophy, 81: 235–256.
  • Vennekens, J., M. Denecker, and M. Bruynooghe, 2009, “CP-logic: A Language of Causal Probabilistic Events and its Relation to Logic Programming,” Theory and Practice of Logic Programming, 9: 245–308.
  • Walley, P., 1991, Statistical Reasoning with Imprecise Probabilities, London: Chapman and Hall.
  • Williamson, J., 2002, “Probability Logic,” in Handbook of the Logic of Argument and Inference: the Turn Toward the Practical, D. Gabbay, R. Johnson, H. J. Ohlbach, and J. Woods (eds.), Amsterdam: Elsevier, pp. 397–424.
  • Yalcin, S., 2010, “Probability Operators,” Philosophy Compass, 5: 916–937.

Other Internet Resources

[Please contact the author with suggestions.]

Acknowledgments

We would like to thank Joe Halpern, Jan Heylen, Jan-Willem Romeijn and the anonymous referees for their comments on this entry.

Copyright © 2013 by
Lorenz Demey <lorenz.demey@hiw.kuleuven.be>
Barteld Kooi <b.p.kooi@rug.nl>
Joshua Sack <joshua.sack@gmail.com>

This is a file in the archives of the Stanford Encyclopedia of Philosophy.
Please note that some links may no longer be functional.
[an error occurred while processing the directive]