To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
In Chapter 3 we saw how belief networks are used to represent statements about independence of variables in a probabilistic model. Belief networks are simply one way to unite probability and graphical representation. Many others exist, all under the general heading of ‘graphical models’. Each has specific strengths and weaknesses. Broadly, graphical models fall into two classes: those useful for modelling, such as belief networks, and those useful for inference. This chapter will survey the most popular models from each class.
Graphical models
Graphical Models (GMs) are depictions of independence/dependence relationships for distributions. Each class of GM is a particular union of graph and probability constructs and details the form of independence assumptions represented. Graphical models are useful since they provide a framework for studying a wide class of probabilistic models and associated algorithms. In particular they help to clarify modelling assumptions and provide a unified framework under which inference algorithms in different communities can be related.
It needs to be emphasised that all forms of GM have a limited ability to graphically express conditional (in)dependence statements [281]. As we've seen, belief networks are useful formodelling ancestral conditional independence. In this chapter we'll introduce other types of GM that are more suited to representing different assumptions. Here we'll focus on Markov networks, chain graphs (which marry belief and Markov networks) and factor graphs. There are many more inhabitants of the zoo of graphical models, see [73, 314].
In this paper we extend a classical theorem of Corrádi and Hajnal into the setting of sparse random graphs. We show that if p(n) ≫ (log n/n)1/2, then asymptotically almost surely every subgraph of G(n, p) with minimum degree at least (2/3 + o(1))np contains a triangle packing that covers all but at most O(p−2) vertices. Moreover, the assumption on p is optimal up to the (log n)1/2 factor and the presence of the set of O(p−2) uncovered vertices is indispensable. The main ingredient in the proof, which might be of independent interest, is an embedding theorem which says that if one imposes certain natural regularity conditions on all three pairs in a balanced 3-partite graph, then this graph contains a perfect triangle packing.
In this paper we propose a novel approach for analysing proof nets of Multiplicative Linear Logic (MLL) using coding theory. We define families of proof structures called PS-families and introduce a metric space for each family. In each family:
(1) an MLL proof net is a true code element; and
(2) a proof structure that is not an MLL proof net is a false (or corrupted) code element.
The definition of our metrics elegantly reflects the duality of the multiplicative connectives. We show that in our framework one-error-detection is always possible but one-error-correction is always impossible. We also demonstrate the importance of our main result by presenting two proof-net enumeration algorithms for a given PS-family: the first searches proof nets naively and exhaustively without help from our main result, while the second uses our main result to carry out an intelligent search. In some cases, the first algorithm visits proof structures exponentially, while the second does so only polynomially.
A subgraph of a hypergraph H is even if all its degrees are positive even integers, and b-bounded if it has maximum degree at most b. Let fb(n) denote the maximum number of edges in a linearn-vertex 3-uniform hypergraph which does not contain a b-bounded even subgraph. In this paper, we show that if b ≥ 12, thenfor some absolute constant B, thus establishing fb(n) up to polylogarithmic factors. This leaves open the interesting case b = 2, which is the case of 2-regular subgraphs. We are able to show for some constants c, C > 0 thatWe conjecture that f2(n) = n1 + o(1) as n → ∞.
R. H. Schelp conjectured that if G is a graph with |V(G)| = R(Pn, Pn) such that δ(G) > , then in every 2-colouring of the edges of G there is a monochromatic Pn. In other words, the Ramsey number of a path does not change if the graph to be coloured is not complete but has large minimum degree.
Here we prove Ramsey-type results that imply the conjecture in a weakened form, first replacing the path by a matching, showing that the star-matching–matching Ramsey number satisfying R(Sn, nK2, nK2) = 3n − 1. This extends R(nK2, nK2) = 3n − 1, an old result of Cockayne and Lorimer. Then we extend this further from matchings to connected matchings, and outline how this implies Schelp's conjecture in an asymptotic sense through a standard application of the Regularity Lemma.
It is sad that we are unable to hear Dick Schelp's reaction to our work generated by his conjecture.
Upper and lower bounds are proved for the maximum number of triangles in C2k+1-free graphs. The bounds involve extremal numbers related to appropriate even cycles.
In Part I we discussed inference and showed that for certain models this is computationally tractable. However, for many models of interest, one cannot perform inference exactly and approximations are required.
In Part V we discuss approximate inference methods, beginning with sampling-based approaches. These are popular and well known in many branches of the mathematical sciences, having their origins in chemistry and physics. We also discuss alternative deterministic approximate inference methods which in some cases can have remarkably accurate performance.
It is important to bear in mind that no single algorithm is going to be best on all inference tasks. For this reason, we attempt throughout to explain the assumptions behind the techniques so that one may select an appropriate technique for the problem at hand.
Natural organisms inhabit a dynamical environment and arguably a large part of natural intelligence is in modelling causal relations and consequences of actions. In this sense, modelling temporal data is of fundamental interest. In a more artificial environment, there are many instances where predicting the future is of interest, particularly in areas such as finance and also in tracking of moving objects.
In Part IV, we discuss some of the classical models of timeseries that may be used to represent temporal data and also to make predictions of the future. Many of these models are well known in different branches of science from physics to engineering and are heavily used in areas such as speech recognition, financial prediction and control. We also discuss some more sophisticated models in Chapter 25, which may be skipped at first reading.
As an allusion to the fact that natural organisms inhabit a temporal world, we also address in Chapter 26 some basic models of how information processing might be achieved in distributed systems.
When the distribution is multiply connected it would be useful to have a generic inference approach that is efficient in its reuse of messages. In this chapter we discuss an important structure, the junction tree, that by clustering variables enables one to perform message passing efficiently (although the structure onwhich themessage passing occurs may consist of intractably large clusters). The most important thing is the junction tree itself, based on which different message-passing procedures can be considered. The junction tree helps forge links with the computational complexity of inference in fields from computer science to statistics and physics.
Clustering variables
In Chapter 5 we discussed efficient inference for singly connected graphs, for which variable elimination and message-passing schemes are appropriate. In the multiply connected case, however, one cannot in general perform inference by passing messages only along existing links in the graph. The idea behind the Junction Tree Algorithm (JTA) is to form a new representation of the graph in which variables are clustered together, resulting in a singly connected graph in the cluster variables (albeit on a different graph). The main focus of the development will be on marginal inference, though similar techniques apply to different inferences, such as finding the most probable state of the distribution.
At this stage it is important to point out that the JTA is not a magic method to deal with intractabilities resulting from multiply connected graphs; it is simply a way to perform correct inference on a multiply connected graph by transforming to a singly connected structure.
Let 1 ≤ p ≤ r + 1, with r ≥ 2 an integer, and let G be a graph of order n. Let d(v) denote the degree of a vertex v ∈ V(G). We show that ifthen G has more than(r + 1)-cliques sharing a common edge. From this we deduce that ifthen G contains more thancliques of order r + 1.
In turn, this statement is used to strengthen the Erdős–Stone theorem by using ∑v ∈ V(G)dp(v) instead of the number of edges.
Richard Schelp completed his PhD in lattice theory in 1970 at Kansas State University. However, he did not take a traditional route to a PhD in mathematics and an outstanding career as a professor and a mathematical researcher. He grew up in rural northeast Missouri. He received his BS in mathematics and physics from the University of Central Missouri. After the completion of his master's degree in mathematics from Kansas State University, he assumed a position as an associate mathematician in the Applied Science Laboratory at Johns Hopkins University for five years. To start his PhD programme at Kansas State University, he had to quit a well-paying position. Also, he was already married to his wife Billie (Swopes) Schelp and he had a family – a daughter Lisa and a son Rick. This was a courageous step to take, but it says something about who Dick Schelp was.
Hidden Markov models assume that the underlying process is discrete; linear dynamical systems that the underlying process is continuous. However, there are scenarios in which the underlying system might jump from one continuous regime to another. In this chapter we discuss a class of models that can be used in this situation. Unfortunately the technical demands of this class of models are somewhat more involved than in previous chapters, although the models are correspondingly more powerful.
Introduction
Complex timeseries which are not well described globally by a single linear dynamical system may be divided into segments, each modelled by a potentially different LDS. Such models can handle situations in which the underlying model ‘jumps’ from one parameter setting to another. For example a single LDS might well represent the normal flows in a chemical plant. When a break in a pipeline occurs, the dynamics of the system changes from one set of linear flow equations to another. This scenario can be modelled using a set of two linear systems, each with different parameters. The discrete latent variable at each time st ∈ {normal, pipe broken} indicates which of the LDSs is most appropriate at the current time. This is called a Switching LDS (SLDS) and is used in many disciplines, from econometrics to machine learning [12, 63, 59, 235, 324, 189].