To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
We study the succinctness of monadic second-order logic and a varietyof monadic fixed point logics on trees. All these languages are known to havethe same expressive power on trees, but some can express the samequeries much more succinctly than others. For example, we show that, undersome complexity theoretic assumption, monadic second-order logic isnon-elementarily more succinct than monadic least fixed point logic,which in turn is non-elementarily more succinct than monadic datalog. Succinctness of the languages is closely related to the combined andparameterised complexity of query evaluation for these languages.
We analyse a class of randomized Least Recently Used (LRU) cache replacement algorithms under the independent reference model with generalized Zipf's law request probabilities. The randomization was recently proposed for Web caching as a mechanism that discriminates between different document sizes. In particular, the cache maintains an ordered list of documents in the following way. When a document of size $s$ is requested and found in the cache, then with probability $p_s$ it is moved to the front of the cache; otherwise the cache stays unchanged. Similarly, if the requested document of size $s$ is not found in the cache, the algorithm places it with probability $p_s$ to the front of the cache or leaves the cache unchanged with the complementary probability $(1-p_s)$. The successive randomized decisions are independent and the corresponding success probabilities $p_s$ are completely determined by the size of the currently requested document. In the case of a replacement, the necessary number of documents that are least recently moved to the front of the cache are removed in order to accommodate the newly placed document.
In this framework, we provide explicit asymptotic characterization of the cache fault probability. Using the derived result we prove that the asymptotic performance of this class of algorithms is optimized when the randomization probabilities are chosen to be inversely proportional to document sizes. In addition, for this optimized and easy-to-implement policy, we show that its performance is within a constant factor from the optimal static algorithm.
Let $F(\b{z})=\sum_\b{r} a_\b{r}\b{z^r}$ be a multivariate generating function that is meromorphic in some neighbourhood of the origin of $\mathbb{C}^d$, and let $\sing$ be its set of singularities. Effective asymptotic expansions for the coefficients can be obtained by complex contour integration near points of $\sing$.
In the first article in this series, we treated the case of smooth points of $\sing$. In this article we deal with multiple points of $\sing$. Our results show that the central limit (Ornstein–Zernike) behaviour typical of the smooth case does not hold in the multiple point case. For example, when $\sing$ has a multiple point singularity at $(1, \ldots, 1)$, rather than $a_\b{r}$ decaying as $|\b{r}|^{-1/2}$ as $|\b{r}| \to \infty$, $a_\b{r}$ is very nearly polynomial in a cone of directions.
This special issue is devoted to the Analysis of Algorithms (AofA). Most of the papers are from the Eighth Seminar on Analysis of Algorithms, held in Strobl, Austria, June 23–29, 2002.
Heap ordered trees are planted plane trees, labelled in such a way that the labels always increase from the root to a leaf. We study two parameters, assuming that $p$ of the $n$ nodes are selected at random: the size of the ancestor tree of these nodes and the smallest subtree generated by these nodes. We compute expectation, variance, and also the Gaussian limit distribution, the latter as an application of Hwang's quasi-power theorem.
An additive decomposition of a set $I$ of nonnegative integers is an expression of $I$ as the arithmetic sum of two other such sets. If the smaller of these has $p$ elements, we have a $p$-decomposition. If $I$ is obtained by randomly removing $n^{\alpha}$ integers from $\{0,\dots,n-1\}$, decomposability translates into a balls-and-urns problem, which we start to investigate (for large $n$) by first showing that the number of $p$-decompositions exhibits a threshold phenomenon as $\alpha$ crosses a $p$-dependent critical value. We then study in detail the distribution of the number of 2-decompositions. For this last case we show that the threshold is sharp and we establish the threshold function.
We consider Boolean functions over $n$ variables. Any such function can be represented (and computed) by a complete binary tree with and or or in the internal nodes and a literal in the external nodes, and many different trees can represent the same function, so that a fundamental question is related to the so-called complexity of a Boolean function: $L(f):=$ minimal size of a tree computing $f$.
The existence of a limiting probability distribution $P(\cdot)$ on the set of and/or trees was shown by Lefmann and Savický [8]. We give here an alternative proof, which leads to effective computation in simple cases. We also consider the relationship between the probability $P(f)$ and the complexity $L(f)$ of a Boolean function $f$. A detailed analysis of the functions enumerating some sub-families of trees, and of their radius of convergence, allows us to improve on the upper bound of $P(f)$, established by Lefmann and Savický.
We show that, for a certain class of probabilistic models, the number of internal nodes $S_n$ of a trie built from $n$ independent and identically distributed keys is concentrated around its mean, in the sense that $\Var S_n=\Oh(\EE S_n)$. Keys are sequences of symbols which may be taken from varying alphabets, and the choice of the alphabet from which the $k$th symbol is taken, as well as its distribution, may depend on all the preceding symbols. In the construction of the trie we also allow for bucket sizes greater than 1. The property that characterizes our models is the following: there is a constant $C$ such that for any word $v$ that may occur as a prefix of a key the size $S^v_n$ of a trie built from the suffixes of $n$ independent keys conditioned to have the common prefix $v$ has the property $\EE S^v_n\leq Cn$. This class of models contains memoryless and Markovian source models as well as the probabilistic dynamical source models that were recently introduced and thoroughly developed by Vallée [Algorithmica29 (2001) 262–306], in particular the continued fraction source. Furthermore we study the external path length $L_n$, which obeys $\EE L_n=\Oh(n\ln n)$ and $\Var L_n=\Oh(n\ln^2 n)$.
In this paper, we present several probabilistic transforms related to classical urn models. These transforms render the dependent random variables describing the urn occupancies into independent random variables with appropriate distributions. This simplifies the analysis of a large number of problems for which a function under investigation depends on the urn occupancies. The approach used for constructing the transforms involves generating functions of combinatorial numbers characterizing the urn distributions. We also show, by using Tauberian theorems derived in this paper, that under certain simple conditions the asymptotic expressions of target functions in the transform domain and in the inverse–transform domain are identical. Therefore, asymptotic information about certain statistics can be obtained without evaluating the inverse transform.
In this paper, we investigate the limit law of the inertial moment of Dyck paths with respect to the $x$-axis, that is, the sum of the squares of the altitudes. We find its Laplace transform using Louchard's methodology, rediscovering a result which was in fact well known by probabilists. We give recurrence relations which enable us to compute the moments of the joint limit law of the area and the inertial moments of both Dyck paths and Grand Dyck paths (bilateral Dyck paths).
We give an algorithm that, with high probability, recovers a planted $k$-partition in a random graph, where edges within vertex classes occur with probability $p$ and edges between vertex classes occur with probability $r\ge p+c\sqrt{p\log n/n}$. The algorithm can handle vertex classes of different sizes and, for fixed $k$, runs in linear time. We also give variants of the algorithm for partitioning matrices and hypergraphs.
This article proposes a surprisingly simple framework for the random generation of combinatorial configurations based on what we call Boltzmann models. The idea is to perform random generation of possibly complex structured objects by placing an appropriate measure spread over the whole of a combinatorial class – an object receives a probability essentially proportional to an exponential of its size. As demonstrated here, the resulting algorithms based on real-arithmetic operations often operate in linear time. They can be implemented easily, be analysed mathematically with great precision, and, when suitably tuned, tend to be very efficient in practice.
The Lehmer–Euclid Algorithm is an improvement of the Euclid Algorithm when applied to large integers. The original Lehmer–Euclid Algorithm replaces divisions on multi-precision integers by divisions on single-precision integers. Here we study a slightly different algorithm that replaces computations on $n$-bit integers by computations on $\mu n$-bit integers. This algorithm depends on the truncation degree $\mu\in ]0, 1[$ and is denoted as the ${\mathcal{LE}}_\mu$ algorithm. The original Lehmer–Euclid Algorithm can be viewed as the limit of the ${\mathcal{LE}}_\mu$ algorithms for $\mu \to 0$. We provide here a precise analysis of the ${\mathcal{LE}}_\mu$ algorithm. For this purpose, we are led to study what we call the Interrupted Euclid Algorithm. This algorithm depends on some parameter $\alpha \in [0, 1]$ and is denoted by ${\mathcal E}_{\alpha}$. When running with an input $(a, b)$, it performs the same steps as the usual Euclid Algorithm, but it stops as soon as the current integer is smaller than $a^\alpha$, so that ${\mathcal E}_{0}$ is the classical Euclid Algorithm. We obtain a very precise analysis of the algorithm ${\mathcal E}_{\alpha}$, and describe the behaviour of main parameters (number of iterations, bit complexity) as a function of parameter $\alpha$. Since the Lehmer–Euclid Algorithm ${\mathcal {LE}}_\mu$ when running on $n$-bit integers can be viewed as a sequence of executions of the Interrupted Euclid Algorithm ${\mathcal E}_{1/2}$ on $\mu n $-bit integers, we then come back to the analysis of the ${\mathcal {LE}}_\mu$ algorithm and obtain our results.
In this paper, we show for generalized $M$-ary search trees that the Steiner distance of $p$ randomly chosen nodes in random search trees is asymptotically normally distributed. The special case $p=2$ shows, in particular, that the distribution of the distance between two randomly chosen nodes is asymptotically Gaussian. In the presented generating functions approach, we consider first the size of the ancestor-tree of $p$ randomly chosen nodes. From the obtained Gaussian limiting distribution for this parameter, we deduce the result for the Steiner distance. Since the size of the ancestor-tree is essentially the same as the number of passes in the (generalized) Multiple Quickselect algorithm, the limiting distribution result also holds for this parameter.
We recast dataflow in a modern categorical light using profunctors as a generalisation of relations. The well-known causal anomalies associated with relational semantics of indeterminate dataflow are avoided, but still we preserve much of the intuitions of a relational model. The development fits with the view of categories of models for concurrency and the general treatment of bisimulation they provide. In particular, it fits with the recent categorical formulation of feedback using traced monoidal categories. The payoffs are: (1) explicit relations to existing models and semantics, especially the usual axioms of monotone IO automata are read off from the definition of profunctors; (2) a new definition of bisimulation for dataflow, the proof of the congruence of which benefits from the preservation properties associated with open maps; and (3) a treatment of higher-order dataflow as a biproduct, essentially by following the geometry of interaction programme.
We study typed behavioural equivalences for the $\pi$-calculus in which the type system allows a form of subtyping. This enables processes to distribute different capabilities selectively on communication channels. The equivalences considered include typed versions of testing equivalences and barbed bisimulation equivalences.
We show that each of these equivalences can be characterised via standard techniques applied to a novel labelled transition system of configurations. These consist of a process term together with two related type environments; one constraining the process and the other its computing environment.
The asynchronous $\pi$-calculus has been considered as the basis of experimental programming languages (or proposals for programming languages) like Pict, Join and TyCO. However, on closer inspection, these languages are based on an even simpler calculus, called Localised$\pi$ (L$\pi$), where: (a) only the output capability of names may be transmitted; (b) there is no matching or similar constructs for testing equality between names.
We study the basic operational and algebraic theory of L$\pi$. We focus on bisimulation-based behavioural equivalences, more precisely, on barbed congruence. We prove two coinductive characterisations of barbed congruence in L$\pi$, and some basic algebraic laws. We then show applications of this theory, including: the derivability of the delayed input; the correctness of an optimisation of the encoding of call-by-name $\lambda$-calculus; the validity of some laws for Join; the soundness of Thielecke's axiomatic semantics of the Continuation Passing Style calculus.
The context of this article is a program studying the bicategory of spans of graphs as an algebra of processes, with applications to concurrency theory. The objective here is to study functorial aspects of reachability, minimisation and minimal realisation. The compositionality of minimisation has application to model-checking.
In propositional or predicate logic, formulas are either true, or false, in any model. Propositional logic and predicate logic do not allow for any further possibilities. From many points of view, however, this is inadequate. In natural language, for example, we often distinguish between various ‘modes’ of truth, such as necessarily true, known to be true, believed to be true and true in the future. For example, we would say that, although the sentence
George W. Bush is president of the United States of America.
is currently true, it will not be true at some point in the future. Equally, the sentence
There are nine planets in the solar system.
while true, and maybe true for ever in the future, is not necessarily true, in the sense that it could have been a different number. However, the sentence
The cube root of 27 is 3.
as well as being true is also necessarily true and true in the future. It does not enjoy all modes of truth, however. It may not be known to be true by some people (children, for example); it may not be believed by others (if they are mistaken).
In computer science, it is often useful to reason about modes of truth. In Chapter 3, we studied the logic CTL in which we could distinguish not only between truth at different points in the future, but also between different futures.