To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
We show that the $4$-state anti-ferromagnetic Potts model with interaction parameter $w\in (0,1)$ on the infinite $(d+1)$-regular tree has a unique Gibbs measure if $w\geq 1-\dfrac{4}{d+1_{_{\;}}}$ for all $d\geq 4$. This is tight since it is known that there are multiple Gibbs measures when $0\leq w\lt 1-\dfrac{4}{d+1}$ and $d\geq 4$. We moreover give a new proof of the uniqueness of the Gibbs measure for the $3$-state Potts model on the $(d+1)$-regular tree for $w\geq 1-\dfrac{3}{d+1}$ when $d\geq 3$ and for $w\in (0,1)$ when $d=2$.
This article considers the link removal problem in a strongly connected directed network with the goal of minimizing the dominant eigenvalue of the network’s adjacency matrix while maintaining its strong connectivity. Due to the complexity of the problem, this article focuses on computing a suboptimal solution. Furthermore, it is assumed that the knowledge of the overall network topology is not available. This calls for distributed algorithms which rely solely on the local information available to each individual node and information exchange between each node and its neighbors. Two different strategies based on matrix perturbation analysis are presented, namely simultaneous and iterative link removal strategies. Key ingredients in implementing both strategies include novel distributed algorithms for estimating the dominant eigenvectors of an adjacency matrix and for verifying strong connectivity of a directed network under link removal. It is shown via numerical simulations on different type of networks that in general the iterative link removal strategy yields a better suboptimal solution. However, it comes at a price of higher communication cost in comparison to the simultaneous link removal strategy.
The transition to open data practices is straightforward albeit surprisingly challenging to implement largely due to cultural and policy issues. A general data sharing framework is presented along with two case studies that highlight these challenges and offer practical solutions that can be adjusted depending on the type of data collected, the country in which the study is initiated, and the prevailing research culture. Embracing the constraints imposed by data privacy considerations, especially for biomedical data, must be emphasized for data outside of the United States until data privacy law(s) are established at the Federal and/or State level.
The study of threshold functions has a long history in random graph theory. It is known that the thresholds for minimum degree k, k-connectivity, as well as k-robustness coincide for a binomial random graph. In this paper we consider an inhomogeneous random graph model, which is obtained by including each possible edge independently with an individual probability. Based on an intuitive concept of neighborhood density, we show two sufficient conditions guaranteeing k-connectivity and k-robustness, respectively, which are asymptotically equivalent. Our framework sheds some light on extending uniform threshold values in homogeneous random graphs to threshold landscapes in inhomogeneous random graphs.
Numerical estimators of differential entropy and mutual information can be slow to converge as sample size increases. The offset Kozachenko–Leonenko (KLo) method described here implements an offset version of the Kozachenko–Leonenko estimator that can markedly improve convergence. Its use is illustrated in applications to the comparison of trivariate data from successive scene color images and the comparison of univariate data from stereophonic music tracks. Publicly available code for KLo estimation of both differential entropy and mutual information is provided for R, Python, and MATLAB computing environments at https://github.com/imarinfr/klo.
Matryoshka dolls, the traditional Russian nesting figurines, are known worldwide for each doll’s encapsulation of a sequence of smaller dolls. In this paper, we exploit the structure of a new sequence of nested matrices we call matryoshkan matrices in order to compute the moments of the one-dimensional polynomial processes, a large class of Markov processes. We characterize the salient properties of matryoshkan matrices that allow us to compute these moments in closed form at a specific time without computing the entire path of the process. This simplifies the computation of the polynomial process moments significantly. Through our method, we derive explicit expressions for both transient and steady-state moments of this class of Markov processes. We demonstrate the applicability of this method through explicit examples such as shot noise processes, growth–collapse processes, ephemerally self-exciting processes, and affine stochastic differential equations from the finance literature. We also show that we can derive explicit expressions for the self-exciting Hawkes process, for which finding closed-form moment expressions has been an open problem since their introduction in 1971. In general, our techniques can be used for any Markov process for which the infinitesimal generator of an arbitrary polynomial is itself a polynomial of equal or lower order.
We present a Markov chain on the n-dimensional hypercube $\{0,1\}^n$ which satisfies $t_{{\rm mix}}^{(n)}(\varepsilon) = n[1 + o(1)]$. This Markov chain alternates between random and deterministic moves, and we prove that the chain has a cutoff with a window of size at most $O(n^{0.5+\delta})$, where $\delta>0$. The deterministic moves correspond to a linear shift register.
Let $(\xi_k,\eta_k)_{k\in\mathbb{N}}$ be independent identically distributed random vectors with arbitrarily dependent positive components. We call a (globally) perturbed random walk a random sequence $T\,{:\!=}\, (T_k)_{k\in\mathbb{N}}$ defined by $T_k\,{:\!=}\, \xi_1+\cdots+\xi_{k-1}+\eta_k$ for $k\in\mathbb{N}$. Consider a general branching process generated by T and let $N_j(t)$ denote the number of the jth generation individuals with birth times $\leq t$. We treat early generations, that is, fixed generations j which do not depend on t. In this setting we prove counterparts for $\mathbb{E}N_j$ of the Blackwell theorem and the key renewal theorem, prove a strong law of large numbers for $N_j$, and find the first-order asymptotics for the variance of $N_j$. Also, we prove a functional limit theorem for the vector-valued process $(N_1(ut),\ldots, N_j(ut))_{u\geq0}$, properly normalized and centered, as $t\to\infty$. The limit is a vector-valued Gaussian process whose components are integrated Brownian motions.
The standard coalescent is widely used in evolutionary biology and population genetics to model the ancestral history of a sample of molecular sequences as a rooted and ranked binary tree. In this paper we present a representation of the space of ranked trees as a space of constrained ordered matched pairs. We use this representation to define ergodic Markov chains on labeled and unlabeled ranked tree shapes analogously to transposition chains on the space of permutations. We show that an adjacent-swap chain on labeled and unlabeled ranked tree shapes has a mixing time at least of order $n^3$, and at most of order $n^{4}$. Bayesian inference methods rely on Markov chain Monte Carlo methods on the space of trees. Thus it is important to define good Markov chains which are easy to simulate and for which rates of convergence can be studied.
In this article we consider the estimation of the log-normalization constant associated to a class of continuous-time filtering models. In particular, we consider ensemble Kalman–Bucy filter estimates based upon several nonlinear Kalman–Bucy diffusions. Using new conditional bias results for the mean of the aforementioned methods, we analyze the empirical log-scale normalization constants in terms of their $\mathbb{L}_n$-errors and $\mathbb{L}_n$-conditional bias. Depending on the type of nonlinear Kalman–Bucy diffusion, we show that these are bounded above by terms such as $\mathsf{C}(n)\left[t^{1/2}/N^{1/2} + t/N\right]$ or $\mathsf{C}(n)/N^{1/2}$ ($\mathbb{L}_n$-errors) and $\mathsf{C}(n)\left[t+t^{1/2}\right]/N$ or $\mathsf{C}(n)/N$ ($\mathbb{L}_n$-conditional bias), where t is the time horizon, N is the ensemble size, and $\mathsf{C}(n)$ is a constant that depends only on n, not on N or t. Finally, we use these results for online static parameter estimation for the above filtering models and implement the methodology for both linear and nonlinear models.
Today’s conflicts are becoming increasingly complex, fluid, and fragmented, often involving a host of national and international actors with multiple and often divergent interests. This development poses significant challenges for conflict mediation, as mediators struggle to make sense of conflict dynamics, such as the range of conflict parties and the evolution of their political positions, the distinction between relevant and less relevant actors in peace-making, or the identification of key conflict issues and their interdependence. International peace efforts appear ill-equipped to successfully address these challenges. While technology is already being experimented with and used in a range of conflict related fields, such as conflict predicting or information gathering, less attention has been given to how technology can contribute to conflict mediation. This case study contributes to emerging research on the use of state-of-the-art machine learning technologies and techniques in conflict mediation processes. Using dialogue transcripts from peace negotiations in Yemen, this study shows how machine-learning can effectively support mediating teams by providing them with tools for knowledge management, extraction and conflict analysis. Apart from illustrating the potential of machine learning tools in conflict mediation, the article also emphasizes the importance of interdisciplinary and participatory, cocreation methodology for the development of context-sensitive and targeted tools and to ensure meaningful and responsible implementation.
We obtain series expansions of the q-scale functions of arbitrary spectrally negative Lévy processes, including processes with infinite jump activity, and use these to derive various new examples of explicit q-scale functions. Moreover, we study smoothness properties of the q-scale functions of spectrally negative Lévy processes with infinite jump activity. This complements previous results of Chan et al. (Prob. Theory Relat. Fields150, 2011) for spectrally negative Lévy processes with Gaussian component or bounded variation.
In this study, we propose a nonlinear Bayesian extension of the Lee–Carter (LC) model using a single-stage procedure with a dimensionality reduction neural network (NN). LC is originally estimated using a two-stage procedure: dimensionality reduction of data by singular value decomposition followed by a time series model fitting. To address the limitations of LC, which are attributed to the two-stage estimation and insufficient model fitness to data, single-stage procedures using the Bayesian state-space (BSS) approaches and extensions of flexibility in modeling by NNs have been proposed. As a fusion of these two approaches, we propose a NN extension of LC with a variational autoencoder that performs the variational Bayesian estimation of a state-space model and dimensionality reduction by autoencoding. Despite being a NN model that performs single-stage estimation of parameters, our model has excellent interpretability and the ability to forecast with confidence intervals, as with the BSS models, without using Markov chain Monte Carlo methods.
In the context of life insurance with profit participation, the future discretionary benefits (FDB), which are a central item for Solvency II reporting, are generally calculated by computationally expensive Monte Carlo algorithms. We derive analytic formulas to estimate lower and upper bounds for the FDB. This yields an estimation interval for the FDB, and the average of lower and upper bound is a simple estimator. These formulae are designed for real world applications, and we compare the results to publicly available reporting data.
We propose a generalized Cramér–Lundberg model of the risk theory of non-life insurance and study its ruin probability. Our model is an extension of that of Dubey (1977) to the case of multiple insureds, where the counting process is a mixed Poisson process and the continuously varying premium rate is determined by a Bayesian rule on the number of claims. We use two proofs to show that, for each fixed value of the safety loading, the ruin probability is the same as that of the classical Cramér–Lundberg model and does not depend on either the distribution of the mixing variable of the driving mixed Poisson process or the number of claim contracts.
Customer churn, which insurance companies use to describe the non-renewal of existing customers, is a widespread and expensive problem in general insurance, particularly because contracts are usually short-term and are renewed periodically. Traditionally, customer churn analyses have employed models which utilise only a binary outcome (churn or not churn) in one period. However, real business relationships are multi-period, and policyholders may reside and transition between a wider range of states beyond that of the simply churn/not churn throughout this relationship. To better encapsulate the richness of policyholder behaviours through time, we propose multi-state customer churn analysis, which aims to model behaviour over a larger number of states (defined by different combinations of insurance coverage taken) and across multiple periods (thereby making use of readily available longitudinal data). Using multinomial logistic regression (MLR) with a second-order Markov assumption, we demonstrate how multi-state customer churn analysis offers deeper insights into how a policyholder’s transition history is associated with their decision making, whether that be to retain the current set of policies, churn, or add/drop a coverage. Applying this model to commercial insurance data from the Wisconsin Local Government Property Insurance Fund, we illustrate how transition probabilities between states are affected by differing sets of explanatory variables and that a multi-state analysis can potentially offer stronger predictive performance and more accurate calculations of customer lifetime value (say), compared to the traditional customer churn analysis techniques.
Modelling loss reserve data in run-off triangles is challenging due to the complex but unknown dynamics in the claim/loss process. Popular loss reserve models describe the mean process through development year, accident year, and calendar year effects using the analysis of variance and covariance (ANCOVA) models. We propose to include in the mean function the persistence terms in the conditional autoregressive range model for modelling the persistence of claim across development years. In the ANCOVA model, we adopt linear trends for the accident and calendar year effects and a quadratic trend for the development year effect. We investigate linear or log-transformed mean functions and four distributions, namely generalised beta type 2, generalised gamma, Weibull, and exponential extension, with positive support to enhance the model flexibility. The proposed models are implemented using the Bayesian user-friendly package Stan running in the R environment. Results show that the models with log-transformed mean function and persistence terms provide better model fits. Lastly, the best model is applied to forecast partial loss reserve and calendar year reserve for three years.