To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Given a weighted undirected graph, the maximum matching problem is to find a matching with maximum total weight. In his seminal paper, Edmonds [35] described an integral polytope for the matching problem, and the famous Blossom Algorithm for solving the problem in polynomial time.
In this chapter, we will show the integrality of the formulation given by Edmonds [35] using the iterative method. The argument will involve applying uncrossing in an involved manner and hence we provide a detailed proof. Then, using the local ratio method, we will show how to extend the iterative method to obtain approximation algorithms for the hypergraph matching problem, a generalization of the matching problem to hypergraphs.
Graph matching
Matchings in bipartite graphs are considerably simpler than matchings in general graphs; indeed, the linear programming relaxation considered in Chapter 3 for the bipartite matching problem is not integral when applied to general graphs. See Figure 9.1 for a simple example.
Linear programming relaxation
Given an undirected graph G = (V, E) with a weight function w: E → ℛ on the edges, the linear programming relaxation for the maximum matching problem due to Edmonds is given by the following LPM(G). Recall that E(S) denotes the set of edges with both endpoints in S ⊆ V and x(F) is a shorthand for ∑e∈Fxe for F ⊆ E.
Although there are exponentially many inequalities in LPM(G), there is an efficient separation oracle for this linear program, obtained by Padberg and Rao using Gomory-Hu trees.
Even though we mentioned the paper by Jain [75] as the first explicit application of the iterative method to approximation algorithms, several earlier results can be reinterpreted in this light, which is what we set out to do in this chapter. We will first present a result by Beck and Fiala [12] on hypergraph discrepancy, whose proof is closest to other proofs in this book. Then we will present a result by Steinitz [127] on rearrangements of sums in a geometric setting, which is the earliest application that we know of. Then we will present an approximation algorithm by Skutella [123] for the single source unsplittable flow problem. Then we present the additive approximation algorithm for the bin packing problem by Karmarkar and Karp [77], which is still one of the most sophisticated uses of the iterative relaxation method. Finally, we sketch a recent application of the iterative method augmented with randomized rounding to the undirected Steiner tree problem [20] following the simplification due to Chakrabarty et al. [24].
A discrepancy theorem
In this section, we present the Beck–Fiala theorem from discrepancy theory using an iterative method. Given a hypergraph G = (V, E), a 2-coloring of the hypergraph is defined as an assignment ψ: V → {-1, +1} on the vertices. The discrepancy of a hyperedge e is defined as discχ(e) = ∑v∈e ψ(v), and the discrepancy of the hypergraph G is defined as discψ(G) = maxe∈E(G) |{discψ(e)}|.
In this first chapter we motivate our method via the assignment problem. Through this problem, we highlight the basic ingredients and ideas of the method. We then give an outline of how a typical chapter in the rest of the book is structured, and how the remaining chapters are organized.
The assignment problem
Consider the classical assignment problem: Given a bipartite graph G = (V1 ∪ V2, E) with |V1| = |V2| and weight function w: E → ℝ+, the objective is to match every vertex in V1 with a distinct vertex in V2 to minimize the total weight (cost) of the matching. This is also called the minimum weight bipartite perfect matching problem in the literature and is a fundamental problem in combinatorial optimization. See Figure 1.1 for an example of a perfect matching in a bipartite graph.
One approach to the assignment problem is to model it as a linear programming problem. A linear program is a mathematical formulation of the problem with a system of linear constraints that can contain both equalities and inequalities, and also a linear objective function that is to be maximized or minimized. In the assignment problem, we associate a variable xuv for every {u, v} ∈ E. Ideally, we would like the variables to take one of two values, zero or one (hence in the ideal case, they are binary variables). When xuv is set to one, we intend the model to signal that this pair is matched; when xuv is set to zero, we intend the model to signal that this pair is not matched.
In this chapter, we will study the spanning tree problem in undirected graphs. First, we will study an exact linear programming formulation and show its integrality using the iterative method. To do this, we will introduce the uncrossing method, which is a very powerful technique in combinatorial optimization. The uncrossing method will play a crucial role in the proof and will occur at numerous places in later chapters. We will show two different iterative algorithms for the spanning tree problem, each using a different choice of 1-elements to pick in the solution. For the second iterative algorithm, we show three different correctness proofs for the existence of a 1-element in an extreme point solution: a global counting argument, a local integral token counting argument and a local fractional token counting argument. These token counting arguments will be used in many proofs in later chapters.
We then address the degree-bounded minimum-cost spanning tree problem. We show how the methods developed for the exact characterization of the spanning tree polyhedron are useful in designing approximation algorithms for this NP-hard problem. We give two additive approximation algorithm: The first follows the first approach for spanning trees and naturally generalizes to give a simple proof of the additive two approximation result of Goemans [59]; the second follows the second approach for spanning trees and uses the local fractional token counting argument to provide a very simple proof of the additive one approximation result of Singh and Lau [125].
In this chapter, we consider a simple model, based on a directed tree representation of the variables and constraints, called network matrices. We show how this model as well as its dual have integral optima when used as constraint matrices with integral right-hand sides. Finally, we show the applications of these models, especially in proving the integrality of the dual of the matroid intersection problem in Chapter 5, as well as the dual of the submodular flow problem in Chapter 7.
While our treatment of network matrices is based on its relations to uncrossed structures and their representations, they play a crucial role in the characterization of totally unimodular matrices, which are all constraint matrices that yield integral polytopes when used as constraint matrices with integral right-hand sides [121]. Note that total unimodularity of network matrices automatically implies integrality of the dual program when the right-hand sides of the dual are integral.
The integrality of the dual of the matroid intersection and submodular flow polyhedra can be alternately derived by showing the Total Dual Integrality of these systems [121]. Although our proof of these facts uses iterative rounding directly on the dual, there is a close connection between these two lines of proof since both use the underlying structure on span of the constraints defining the extreme points of the corresponding linear program.
Quoting Lovász from his paper “Submodular Functions and Convexity” [94]:
Several recent combinatorial studies involving submodularity fit into the following pattern. Take a classical graph-theoretical result (e.g. the Marriage Theorem, the Max-flow-min-cut Theorem etc.), and replace certain linear functions occurring in the problem (either in the objective function or in the constraints) by submodular functions. Often the generalizations of the original theorems obtained this way remain valid; sometimes even the proofs carry over. What is important here to realize is that these generalizations are by no means l'art pour l'art. In fact, the range of applicability of certain methods can be extended tremendously by this trick.
The submodular flow model is an excellent example to illustrate this point. In this chapter, we introduce the submodular flow problem as a generalization of the minimum cost circulation problem. We then show the integrality of its LP relaxation and its dual using the iterative method. We then discuss many applications of the main result. We also show an application of the iterative method to an NP-hard degree bounded generalization and show some applications of this result as well.
The crux of the integrality of the submodular flow formulations will be the property that a maximal tight set of constraints form a cross-free family. This representation allows an inductive token counting argument to show a 1-element in an optimal extreme point solution. We will see that this representation is precisely the one we will eventually encounter in Chapter 8 on network matrices.
We develop lower bounds on the Hadwiger number h(G) of graphs G with high chromatic number. In particular, if G has n vertices and chromatic number k then h(G) ≥ (4k − n)/3.
Analytic combinatorics aims to enable precise quantitative predictions of the properties of large combinatorial structures. The theory has emerged over recent decades as essential both for the analysis of algorithms and for the study of scientific models in many disciplines, including probability theory, statistical physics, computational biology, and information theory. With a careful combination of symbolic enumeration methods and complex analysis, drawing heavily on generating functions, results of sweeping generality emerge that can be applied in particular to fundamental structures such as permutations, sequences, strings, walks, paths, trees, graphs and maps. This account is the definitive treatment of the topic. The authors give full coverage of the underlying mathematics and a thorough treatment of both classical and modern applications of the theory. The text is complemented with exercises, examples, appendices and notes to aid understanding. The book can be used for an advanced undergraduate or a graduate course, or for self-study.
We pose a new and intriguing question motivated by distributed computing regarding random walks on graphs: How long does it take for several independent random walks, starting from the same vertex, to cover an entire graph? We study the cover time – the expected time required to visit every node in a graph at least once – and we show that for a large collection of interesting graphs, running many random walks in parallel yields a speed-up in the cover time that is linear in the number of parallel walks. We demonstrate that an exponential speed-up is sometimes possible, but that some natural graphs allow only a logarithmic speed-up. A problem related to ours (in which the walks start from some probabilistic distribution on vertices) was previously studied in the context of space efficient algorithms for undirected s–t connectivity and our results yield, in certain cases, an improvement upon some of the earlier bounds.
We first describe a reduction from the problem of lower-bounding the number of distinct distances determined by a set S of s points in the plane to an incidence problem between points and a certain class of helices (or parabolas) in three dimensions. We offer conjectures involving the new set-up, but are still unable to fully resolve them.
Instead, we adapt the recent new algebraic analysis technique of Guth and Katz [9], as further developed by Elekes, Kaplan and Sharir [6], to obtain sharp bounds on the number of incidences between these helices or parabolas and points in ℝ3. Applying these bounds, we obtain, among several other results, the upper bound O(s3) on the number of rotations (rigid motions) which map (at least) three points of S to three other points of S. In fact, we show that the number of such rotations which map at least k ≥ 3 points of S to k other points of S is close to O(s3/k12/7).
One of our unresolved conjectures is that this number is O(s3/k2), for k ≥ 2. If true, it would imply the lower bound Ω(s/logs) on the number of distinct distances in the plane.
It often occurs that local copies of a text are modified by users but that the local modifications are not synchronized (thus allowing the merged text to become the source for later editions) until later when, for instance the network connection is reestablished. Since text editions usually affect a small fraction of the whole content, the history of edit operations provides a compact representation of the modified file. In this paper, we define a normal form for these records which will permit for the comparison of all text files that have been obtained by editing a common source S when the difference between each output file Oi and the source file is given as a sequence Li of edit operations. We show that the normalized sequence is unique for all the equivalent text editions and provide efficient procedures with which to compute this normal form and to obtain the edit sequence LM transforming S into a merged file M which integrates all the local modifications. We also discuss how these normalization can be integrated into the operational transformation paradigm for optimistic replication.
Sets of integers form a monoid, where the product of two sets Aand B is defined as the set containing a+b for all $a\in A$ and$b\in B$. We give a characterization of when a family of finitesets is a code in this monoid, that is when the sets do not satisfyany nontrivial relation. We also extend this result for someinfinite sets, including all infinite rational sets.
Recently, a new measurement – the advice complexity – was introduced for measuring the information content of online problems. The aim is to measure the bitwise information that online algorithms lack, causing them to perform worse than offline algorithms. Among a large number of problems, a well-known scheduling problem, job shop scheduling with unit length tasks, and the paging problem were analyzed within this model. We observe some connections between advice complexity and randomization. Our special focus goes to barely random algorithms, i.e., randomized algorithms that use only a constant number of random bits, regardless of the input size. We adapt the results on advice complexity to obtain efficient barely random algorithms for both the job shop scheduling and the paging problem. Furthermore, so far, it has not yet been investigated for job shop scheduling how good an online algorithm may perform when only using a very small (e.g., constant) number of advice bits. In this paper, we answer this question by giving both lower and upper bounds, and also improve the best known upper bound for optimal algorithms.
In this paper we study the parameterized complexity of approximating theparameterized counting problems contained in the class $\#W[P]$,the parameterized analogue of $\#P$. We prove a parameterized analogue of afamous theorem of Stockmeyer claiming that approximate counting belongs tothe second level of the polynomial hierarchy.