Hostname: page-component-6b88cc9666-cdmm5 Total loading time: 0 Render date: 2026-02-13T09:28:40.951Z Has data issue: false hasContentIssue false

Fluid limit of a distributed ledger model with random delay

Published online by Cambridge University Press:  13 February 2026

Jiewei Feng*
Affiliation:
Northeastern University
Christopher King*
Affiliation:
Northeastern University
*
*Postal address: Department of Mathematics, Northeastern University, Boston, MA 02115, USA.
*Postal address: Department of Mathematics, Northeastern University, Boston, MA 02115, USA.
Rights & Permissions [Opens in a new window]

Abstract

Distributed ledgers, including blockchain and other decentralized databases, are designed to store information online where all trusted network members can update the data with transparency. The dynamics of a ledger’s development can be mathematically represented by a directed acyclic graph (DAG). In this paper, we study a DAG model that considers batch arrivals and random delay of attachment. We analyze the asymptotic behavior of this model by letting the arrival rate go to infinity and the inter-arrival time go to zero. We establish that the number of leaves in the DAG, as well as various random variables characterizing the vertices in the DAG, can be approximated by its fluid limit, represented as the solution to a set of delayed partial differential equations. Furthermore, we establish the stable state of this fluid limit and validate our findings through simulations.

Information

Type
Original Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2026. Published by Cambridge University Press on behalf of Applied Probability Trust

1. Introduction

A distributed ledger (DL) is a decentralized database where transactions are stored on a directed acyclic graph (DAG). The goal of any DL is to provide a secure and consistent record of transactions. Due to the widespread adoption of DL technologies for cryptocurrencies, there has recently been growing interest in formally establishing the properties of the ledger.

In the DAG associated with a DL, each vertex represents a block of information. A vertex arrives when a user begins creating a new block locally by collecting data. However, this vertex cannot be attached to the existing DAG until the user completes a proof-of-work (POW) computation [Reference Back2]. Only after solving the POW can the block be broadcast to the network and added as a new vertex in the graph, with edges connecting it to one or more existing blocks according to a specified attachment rule. This time delay between arrival and attachment, imposed by the POW requirement, complicates the system’s dynamics.

To better understand POW, consider the following description: When a new vertex A arrives, one or more existing vertices in the DAG will be chosen; these are called the parents of A. Using the information in A and its parents, a problem is then generated, which is typically implemented with a hash function as described in Section 3 in [Reference Back2]. The user trying to publish the vertex A will have to solve the problem, and this process is called POW. Vertex A is accepted into the DAG only after the POW is completed, with edges connecting it to its chosen parents. The solution to the POW problem is stored in A, ensuring that any alteration to the data in A or its parents changes the problem. If this happens, the stored solution becomes invalid, enabling users to detect tampered data. This mechanism helps users to verify the data in the ledger and protect its record from being doctored. Once a vertex is attached by any future vertex, the transactions stored in the corresponding block are considered verified since at least one POW is finished to secure the data. Figure 1 provides an example of a DAG.

Figure 1. Example of a DAG. A solid directed edge implies that the associated POW has been completed and the data has been accepted into the ledger. For example, vertex 3 has selected 1 as its parent and finished its POW. A dashed vertex with outgoing dashed edge implies the POW has started but not yet been finished. For example, vertex 5 has selected 2 and 3 as its parents but its POW has not yet been finished. We generally refer to the solid vertices that have no solid edge pointing toward them as tips, which means the vertices have been accepted to the distributed ledger but have not yet been attached by any other vertices. For example, vertices 2 and 3 are tips because there is no solid edge pointing toward them, while vertices 0 and 1 are not tips because they have been attached by other solid vertices. Also, vertices 4 and 5 are not considered to be tips because they have not been accepted to the system, as indicated by the fact that their POW have not been finished

In 2008, blockchain [Reference Nakamoto18, Reference Zheng28] was introduced as the first fully decentralized DL. In blockchain, each arriving block only attaches to the one existing block that is the furthest from the root in the DAG, which is known as the longest chain protocol, hence leading to a linear chain structure. When there are multiple blocks with ongoing POWs, only one of them will be accepted into the system to maintain the chain structure and prevent double spending. Such a protocol requires a significant amount of computational power for POW, and hence normally requires a transaction fee to compensate participants who contribute computational resources to solving the POW.

The first version of IOTA [Reference Popov21] was an alternative to blockchain where the chain structure is no longer required, and hence all honest blocks can be accepted into the system as long as their POWs are completed. Without the chain structure constraint, IOTA and other similar protocols are more challenging to analyze, particularly when accounting for POW delays. Tangle 2.0 [Reference Müller17] was later introduced as a fully decentralized version of IOTA. This version includes a voting mechanism that replaces POW for achieving consensus. However, in Tangle 2.0, POW still exists in the sense that blocks are created using hash functions, as described in Section IV.A of [Reference Müller17].

Distributed ledger technologies have evolved significantly in recent years and continue to be a dynamic and rapidly growing field. As a result, there is increasing interest in rigorously establishing the mathematical properties related to their security and stability. Recent research has explored several key aspects of DLs. One aspect is known as the consensus, where all users should agree on the same set of non-conflicting information. This aspect has been extensively analyzed for blockchain designs; see, for example, [Reference Ameen, Sankagiri and Hajek1, Reference Pahlajani, Kshirsagar and Pachghare19, Reference Zhang and Lee27]. In particular, the work [Reference Dey and Gopalan4] identifies the relationship between the consensus of DL and the one-endedness property of the corresponding DAG model for blockchain. Later work [Reference Gopalan, Sankararaman, Walid and Vishwanath13] extends the analysis to other protocols and shows that IOTA-like DLs are one-ended when considering communication delay, with [Reference Feng, King and Duffy6] generalizing the results for delays caused by both POW and communication. The work in [Reference Frolkova and Mandjes9] investigates how random communication delay affects broadcasting newly created blocks to all users using a queuing model. The probability bound on the success rate when an adversarial power performs a double spending attack is studied in [Reference Gazi, Ren and Russell10, Reference Goffard12]. Variants of blockchain have also been introduced to improve certain performance measures [Reference Lewenberg, Sompolinsky and Zohar15, Reference Sankagiri, Wang, Kannan and Viswanath23]. The convergence of the long-time average tip count in individual data storage is studied in [Reference Müller, Amigo, Reiffers-Masson and Ruano-Rincón16] with its converging limit derived through simulation.

In this paper, we focus on the IOTA system, where each vertex selects two parents randomly from the tips with equal probability. As previously mentioned, a vertex becomes verified when it is attached by a future vertex. A critical aspect of the security of DLs is the time it takes for an accepted vertex to be attached by others, which is influenced by the number of tips. A fluid limit approximation for the dynamics of the number of tips, assuming a fixed POW duration, was introduced in [Reference King14]. This approximation provides key insights into the stationary number of tips and its convergence rate, which is critical for understanding the verification speed in the ledger. Other applications of the fluid limit introduced in [Reference King14] include the study of the behavior of vertices with conflicting information [Reference Ferraro, King and Shorten7] and the modification of the IOTA attachment mechanism [Reference Ferraro, King and Shorten8].

As an extension of [Reference King14], this paper focuses on analyzing the fluid limit under the assumption that the POW duration is multinoulli distributed. When assuming a fixed duration of POW, the time at which a tip ceases to be a tip is determined by the first time this tip gets selected as a parent. However, when POW duration is random, the time at which a tip will cease to be a tip is uncertain until it is actually attached by others. Therefore, various challenges arise when trying to analyze the dynamics of IOTA with random POW duration. In this paper, we introduce a new model to describe the dynamics by allowing that multiple vertices can arrive simultaneously. A new expression of the fluid limit arises because of this approach and the consideration of random POW duration, which is proved to be a good approximation for the random dynamics of the model. In addition to the number of tips, which was analyzed in [Reference King14], we also further distinguish the tips into different types according to their expected duration before they cease to be a tip. Such consideration of additional variables not only enables us to prove that the fluid limit is a good estimation of the random model, but also provides insight into the proportion of tips that are at different levels of risk from a security aspect.

The remainder of this paper is organized as follows. Section 2 provides a description of the model and the distribution of the random variables. Section 2.6 introduces the fluid limit and discusses the initial conditions that affect the analysis. Our main results are summarized in Section 3, which also presents several applications of the fluid limit model, demonstrating its efficacy in capturing key aspects of the DL’s dynamics. Section 4 offers a high-level intuition for the proof of the main result, with the detailed proof provided in Sections 5 and 6.

2. Model Set-Up

The DAG model used in this paper is a variation of the DAG model introduced in [Reference King14]. The main novelties of the model in this paper include the consideration of random POW durations and simultaneous arrivals of multiple vertices at each time. Before defining the model in detail, Section 2.1 provides a brief introduction to the notation, indicating where each variable is formally defined. Then, we introduce variables that appear in the context of a DL. The distributions of the random variables and the evolution equations for the model will be given in Sections 2.3 and 2.4, respectively.

2.1. Notation

Section 2.2: $\epsilon$ denotes the time between arrivals, N denotes the number of vertices arriving at each time $t_n=n\epsilon$ , and $\lambda=N/\epsilon$ denotes the arrival rate of vertices. The values $h_1 < \cdots < h_M$ represent the possible durations of POW, and $N_i(t_n)$ denotes the number of Type i arrivals at time $t_n$ . The variable $L(t_n)$ denotes the number of tips at $t_n$ , while $F(t_n)$ and $W(t_n)$ , respectively, denote the number of free tips and the number of pending tips. The variable $F_i(t_n)$ counts the number of free tips selected as parents by Type i arrivals but not by Type j arrivals for any $j < i$ . The number of Type i pending tips at time $t_n$ with residual lifetime (RLT) u is recorded by $W_i(t_n,u)$ . The value $J_{i,j}(t_n,u)$ counts the number of Type i pending tips at time $t_n$ with RLT u that change to Type j pending tips. Sometimes we will use $F_\lambda(t)$ and $L_\lambda(t)$ to respectively denote $F(t)/\lambda$ and $L(t)/\lambda$ for simplicity of presentation.

Section 2.5: The values $l,f,w,w_i$ represent the fluid limit of variables $L,F,W,W_i$ .

Section 3: g(t) records the difference between the discrete process and its fluid limits, with T representing the time horizon. The values $\gamma,\xi,m$ are parameters describing the fluid limit, which are given by (3.3)–(3.5).

2.2. Preliminaries

Let $\epsilon$ and N be two constants, and let $t_n=n\epsilon$ for $n\in\mathbb{N}$ . We assume that N vertices arrive simultaneously at each time $t_n$ , with $\epsilon$ representing the inter-arrival time. We call a vertex that arrives at time $t_n$ an arrival at $t_n$ . We define $\lambda\,:\!=\,N/\epsilon$ , which represents the arrival rate of the vertices.

We assume that each new arriving vertex independently chooses its duration of POW from the set $\{h_1,h_2,\ldots,h_M\}\subseteq\mathbb{R}$ with probability $p_1,p_2,\ldots,p_M$ , respectively. Without loss of generality, we assume that $h_1 < h_2 < \cdots < h_M$ , which are not necessarily equidistant, and that any $h_i$ is an integer multiple of $\epsilon$ . Note that while arriving processes are commonly analyzed with $\epsilon = 1$ , where all time variables are normalized by the inter-arrival time, we take $\epsilon \rightarrow 0$ so that we can use fixed $\{h_i\}_{i=1,\ldots,M}$ to represent real-world POW durations (e.g. minutes). The alternative approach of fixing $\epsilon = 1$ and letting $\{h_i\}_{i=1,\ldots,M} \rightarrow \infty$ would be less natural for interpreting results in terms of actual timescales. This current setting captures scenarios where many blocks arrive rapidly but each requires substantial but bounded computation time. Throughout the paper, we call an arriving vertex with POW duration $h_i$ a Type i arrival and the corresponding POW a Type i POW. We define $N_i(t_n)$ to be the number of Type i arrivals that arrive at time $t_n$ ; their distribution will be provided in (2.1). Note that $\sum_{i=1}^{M}N_i(t_n)=N$ because at each time we have N arriving vertices. Figure 2 provides a graphical demonstration of the dynamics of the DAG used in this paper. It is worth noting that the assumption of a deterministic arrival process can be generalized to a binomial arrival process, a classical model closely related to Poisson arrivals. This can be done by assuming that each arrival is independently discarded with probability $p_0$ . While this generalization requires minor modifications to the proof, the main arguments remain intact, as will become clearer in Section 2.5.

Figure 2. A demonstration of a DAG modeling IOTA at some time t. Each time there are $N=2$ arriving vertices, for example, $C_1,C_2$ arrived at the same time while $D_1,D_2$ arrived simultaneously $\epsilon$ time after $C_1,C_2$ arrived. A solid vertex with outgoing solid edge(s) implies that its corresponding POW has been completed and the vertex has been accepted into the system. For example, vertex $C_1$ has selected $A_1$ as its parent and finished its POW, hence it has been accepted into the system. A dashed vertex with outgoing dashed edge(s) implies the POW has not yet been finished. For example, vertex $D_2$ has selected $B_1$ and $A_2$ as its parents but the POW corresponding to $D_2$ has not yet been finished. The dashed vertices are the ones that have not been accepted into the system because their POWs are still in process. A dashed vertex will be included in the system once its corresponding POW is finished, then the dashed vertex and the dashed edge(s) originating from it will become solid

We recall the terminology for tips and pending tips used in [Reference King14]. As in Figure 2, the solid vertices without any solid edge toward them are called tips, which represent the vertices that are accepted into the system and have not yet been attached by other vertices. For example, the tips in Figure 2 are $A_2,B_1,B_2,C_1,C_2$ . At time $t_n$ we let $L(t_n)$ denote the number of tips in the system. Among these, we distinguish between the pending tips and the free tips. A tip is a pending tip at time $t_n$ if it has been selected as a parent by some vertices that arrived at some time $t_j$ with $j < n$ . A pending tip appears as a solid vertex with a dashed edge toward it; for example, the pending tips in Figure 2 are vertices $B_1,A_2$ . A tip is a free tip at time $t_n$ if it is not a pending tip. A free tip appears as a solid vertex without any dashed edge toward it, for example, the free tips in Figure 2 are $B_2,C_1,C_2$ . We define $W(t_n)$ to be the number of pending tips at time $t_n$ and $F(t_n)$ to be the number of free tips at time $t_n$ . It follows that $L(t_n) = W(t_n) + F(t_n)$ . In order to analyze the evolution of random variables $L(t_n)$ and $F(t_n)$ , which will be respectively described in (2.8) and (2.7), we first introduce some other random variables.

Essentially, all DL protocols require each new vertex to have at least one parent. Vertices in blockchain typically have one parent, resulting in a linear graph. In addition to IOTA, there are other protocols where each vertex has multiple parents, including the work in [Reference Sompolinsky, Lewenberg and Zohar24, Reference Sompolinsky, Wyborski and Zohar25]. While this paper assumes that each vertex has two parents, the same procedure can be applied to protocols that require k parents.

To model the parent selection algorithm used in IOTA, we assume that each vertex arriving at time $t_n$ chooses two parents with replacement from the set of tips, whose population is recorded by $L(t_n)$ , and the tips are selected as parents with equal probability. The exact probability distribution will be given in (2.2). Note that it could be the case that the vertex makes the same choice for both selections, resulting in a situation where the vertex has only one parent. In fact, whether we consider replacement does not affect our results, as the difference only appears in lower-order terms. For example, (2.12) considers only the highest-order term, which suffices for our analysis. The selection of parents affects the dynamics of the system, and we assume that the effect resulting from the parent selection will be reflected at the next step. For example, after a free tip is selected as a parent at $t_n$ , it will become a pending tip at $t_{n+1}$ .

In order to analyze the dynamics of the DAG, we now introduce the variables used to record the number of free tips selected at each step. We first provide the definitions of these variables and then give an example for demonstration. We suppose that the Type 1 arrivals first make their selections, and we define $F_1(t_n)$ to be the number of free tips selected by Type 1 arrivals at time $t_n$ . Subsequently, the Type 2 arrivals make their selections, and we define $F_2(t_n)$ to be the number of free tips selected by Type 2 arrivals but not by any Type 1 arrivals at time $t_n$ . We then use the same idea to define $F_i(t_n)$ as the number of free tips that are selected by Type i arrivals but not by any Type j arrivals for all $j < i$ . An example is given in Figure 3. To model the situation where a free tip becomes a pending tip after it is selected as a parent, we require that all the tips counted in $\sum_{i=1}^M F_i(t_n)$ become pending tips at the next step; that is, they are counted as part of $W(t_{n+1})$ . Based on the description of $F_i(t_n)$ , we have the distribution of $F_i(t_n)$ as provided in (2.3)–(2.5).

Figure 3. Example of the random variables $F_i(t_k)$ . In this example, $N=3$ which means there are 3 arrivals at each time. At some time $t_k$ the vertices $D_1,D_2,D_3$ arrive and they have POW durations $h_1,h_2,h_3$ , respectively. The graph on the left represents the graph at $t_k$ when $D_1,D_2,D_3$ arrive while the right graph represents the graph at $t_{k+1}$ . At time $t_k$ , the vertices $A_1,B_1,C_1,C_2$ are free tips since their corresponding POWs have been finished but they have not yet been selected as parents. The free tip $A_1$ is selected as a parent by a Type 1 arrival $D_1$ . The free tip $B_1$ is selected by a Type 1 arrival $D_1$ and a Type 2 arrival $D_2$ . The free tip $C_2$ is selected by a Type 2 arrival $D_2$ and a Type 3 arrival $D_3$ . Then $F_1(t_k)=2$ because both $A_1,B_1$ are selected by a Type 1 arrival. Because $B_1$ is already included in $F_1(t_k)$ , we have $F_2(t_k)=1$ although both $B_1,C_2$ are selected by a Type 2 arrival. Finally, $F_3(t_k)=0$ because $C_2$ is the only free tip that is selected by a Type 3 arrival but $C_2$ has been counted in $F_2(t_k)$ . The selection will be reflected at $t_{k+1}$ and hence $A_1,B_1,C_2$ become pending tips at $t_{k+1}$

Recall that a POW with duration $h_i$ is called a Type i POW, and a pending tip is a tip that has been selected as a parent. If vertex A is selected as a parent of a vertex B, we say that the POW corresponding to B is directed to vertex A. A pending tip will cease to be a tip when a POW directed to it is finished. We define the residual lifetime (RLT) of a pending tip at time $t_n$ as the remaining time (measured in units of $\epsilon$ ) required for finishing one of the ongoing POWs directed to the pending tip at time $t_n$ . For example, a pending tip A at time $t_7$ might have been selected by three vertices: an arrival at time $t_4$ with corresponding POW duration $h_2=7\epsilon$ whose POW will be finished at time $t_4+h_2=t_{11}$ ; an arrival at time $t_5$ with corresponding POW duration $h_3=8\epsilon$ whose POW will be finished at time $t_5+h_3=t_{13}$ ; and an arrival at time $t_6$ with corresponding POW duration $h_5=10\epsilon$ whose POW will be finished at $t_6+h_5=t_{16}$ . Since the three listed POWs will be finished at $t_{11}$ , $t_{13}$ , and $t_{16}$ , respectively, the RLT of this pending tip A at time $t_7$ is $(11-7)\epsilon=4\epsilon$ , which is determined by the POW with duration $h_2=7\epsilon$ . Suppose no arrivals at time $t_7$ select A as a parent. Then at time $t_8$ , the RLT of the pending tip A is $4\epsilon-\epsilon=3\epsilon$ because $\epsilon$ time has passed. However, notice that the RLT at $t_n$ only uses information up to time $t_n$ , and it can jump to a much lower value if the pending tip is selected as a parent by a new arrival. Using the same example, suppose that at time $t_8$ an arrival with POW duration $h_1=2\epsilon$ selects A as its parent. Then at time $t_9$ there are four POWs that are directed to A and the POW that starts at $t_8$ will finish at the earliest at time $t_8+2\epsilon=t_{10}$ , and hence the RLT of A at $t_9$ is $\epsilon$ . When the RLT of a pending tip becomes zero at $t_n$ , the pending tip will cease to be a tip at the next step; that is, the vertex will no longer be included in the set of tips at time $t_{n+1}$ .

Notice that there could be multiple POWs that determine the RLT of a pending tip at a given time. In this set of vertices whose corresponding POWs determine the RLT of a given pending tip at a given time, the vertices might be of different Types, i.e. they have different POW durations. For example, a pending tip A at time $t_9$ was selected by two vertices: a Type 2 arrival at time $t_6$ with POW duration $h_2=4\epsilon$ which will be finished at $t_{10}$ ; and a Type 1 arrival at $t_8$ with POW duration $h_1=2\epsilon$ which will be finished at $t_{10}$ . Then at time $t_9$ , the RLT of A is $\epsilon$ , which is determined by both the Type 1 POW and the Type 2 POW directed to A. If the smallest Type of the POWs that determine the RLT of the pending tip A at time $t_n$ is i, then we call the pending tip A a Type i pending tip. In the example, the smallest Type of POW that determines the RLT of A at $t_9$ is Type 1, and hence A is a Type 1 pending tip at $t_9$ . Note that a pending tip can still be selected as a parent by a new arrival before one of the POWs directed to it has been completed, so a pending tip may change from Type i to Type j for $i>j$ . However, notice that a pending tip cannot change from Type i to Type j if $i<j$ because being selected by a vertex with higher Type will not affect the related set of POWs that determine the RLT of the pending tip.

By the above definitions, we denote $\mathcal{W}_i(t_n, u)$ to be the set of Type i pending tips with RLT equal to u at time $t_n$ . And we denote $W_i(t_n, u)\,:\!=\,|\mathcal{W}_i(t_n, u)|$ as the number of Type i pending tips at time $t_n$ with RLT equal to u. For a Type i pending tip counted in $W_i(t_n,u)$ , it might jump to a lower Type j where $h_j\leq u$ because it gets selected by a Type j arrival at time $t_n$ . If this happens, the pending tip will be counted in $W_j(t_{n+1},h_j-\epsilon)$ as a Type j pending tip in the next moment. We hence define $J_{i,j}(t_n,u)$ to be the number of pending tips in the set ${\mathcal{W}}_i(t_n, u)$ that jump to Type j at the next step. The distribution of $J_{i,j}(t_n,u)$ will be defined in (2.6).

2.3. Distributions of random variables

Based on the intuition of the random variables described in Section 2.2, we first introduce the distribution of the variables, including $N_i(t_n)$ , $F_i(t_n)$ , and $J_{i,j}(t_n,u)$ . Since there are N arrivals at each time and each of them has probability $p_k$ of having POW duration $h_k$ for $k=1,\ldots,M$ , the term $N_i(t_n)$ satisfies a binomial distribution as it counts the number of arrivals at time $t_n$ that have POW duration $h_i$ . This means that

(2.1) \begin{align}N_i(t_n)\sim \text{Binomial}(N,p_i) \quad \text{and} \quad \sum_{i=1}^M N_i(t_n)=N.\end{align}

The random variables $F_i(t_n)$ and $J_{i,j}(t_n,u)$ are generated by selections of parents. To model the parent selection algorithm used in IOTA, we assume that each vertex arriving at time $t_n$ chooses two parents with replacement from the set of tips whose population is recorded by $L(t_n)$ and the tips are selected as parents with equal probability. Hence, for an arrival v at time $t_n$ and a tip v’ at time $t_n$ ,

(2.2) \begin{align}\mathbb{P}\big(v' \text{ is not selected as a parent of }v|L(t_n)\big)\,:\!=\, \left(1-\frac{1}{L(t_n)} \right)^2. \end{align}

The term $F_i(t_n)$ calculates how many free tips are selected by the arrivals with duration $h_i$ but are not selected by any arrival with POW duration less than $h_i$ at time $t_n$ . We suppose that the N arrivals first choose their duration for POW and then select their parents. By (2.2) and that the number of arrivals with POW duration $h_1$ is recorded by $N_1(t_n)$ , each free tip has probability $(1-1/L(t_n))^{2N_1(t_n)}$ of not getting selected by any Type 1 arrivals at $t_n$ . We use binary variables $X_{n,1,j}$ with $j=1,2,\ldots,F(t_n)$ to denote whether each free tip is selected as a parent by any Type 1 arrival at $t_n$ , where $X_{n,1,j}=1$ if the corresponding free tip is selected. The conditional probability that a free tip is not selected as a parent by any Type 1 arrival at $t_n$ is

\begin{align*}\mathbb{P}\big(X_{n,1,j}=0|L(t_n),N_1(t_n),F(t_n) \big)\,:\!=\,\left(1 - \frac{1}{L(t_n)} \right)^{2 N_1(t_n)} \quad\text{for } j=1,2,\ldots,F(t_n).\end{align*}

Therefore, we define

(2.3) \begin{align}F_1(t_n) \,:\!=\,\sum_{j=1}^{F(t_n)} X_{n,1,j} \end{align}

where the random variables $X_{n,1,1},X_{n,1,2},\ldots,X_{n,1,F(t_n)}$ are dependent.

Using the same idea, as well as the assumption that the Type i arrivals select their parents after the Type $i-1$ arrivals, we have that by the time the Type i arrivals select their parents, the number of tips that are possible to be counted toward $F_i(t_n)$ is $F(t_n)-\sum _{k=1}^{i-1}F_k(t_n)$ . Given n and i, we use $X_{n,i,j}$ with $j=1,\ldots,F(t_n)-\sum_{k=1}^{i-1}F_k(t_n)$ to denote whether each free tip that is possible to be counted toward $F_i(t_n)$ is selected by a Type i arrival, with $X_{n,i,j}=1$ if the free tip is selected and $X_{n,i,j}=0$ otherwise:

(2.4) \begin{align}\mathbb{P}\big(X_{n,i,j}=0|L(t_n),F(t_n),N_i(t_n), F_k(t_n)\colon k< i\big)\,:\!=\,\left(1 - \frac{1}{L(t_n)}\right)^{2 N_i(t_n)}. \end{align}

As a generalization of (2.3), we define

(2.5) \begin{align}F_i(t_n) \,:\!=\, \sum_{j=1}^{F(t_n)-\sum_{k=1}^{i-1}F_k(t_n)} X_{n,i,j} \quad\text{for } i=2,3,\ldots,M,\end{align}

where M is the total number of possible durations of POW.

Recall that $W_i(t_n,u)$ denotes the number of Type i pending tips whose RLT is u. For $i>j$ and $u\geq h_j$ , $J_{i,j}(t_n, u)$ counts how many pending tips with RLT u jump from Type i to Type j because they are selected by a Type j arrival. This number is also generated by random selection carried out by the N arrivals at time $t_n$ , and we use $X'_{n,i,u,j,k}$ to denote whether each Type i pending tip with RLT u is selected by a Type j arrival at $t_n$ :

\begin{align*}\mathbb{P}\big(X'_{n,i,u,j,k}=0|W_i(t_n, u),N_j(t_n)\big)\,:\!=\,\left(1 - \frac{1}{L(t_n)}\right)^{2 N_{j}(t_n)},\end{align*}

where $k=1,2,\ldots,W_i(t_n, u)$ . We then define

(2.6) \begin{align}J_{i,j}(t_n, u) \,:\!=\, \sum_{k=1}^{W_i(t_n, u)} X'_{n,i,u,j,k}.\end{align}

2.4. Update rules of the model

In this paper, our primary focus is to analyze the dynamics of the random variables $\{F(t_n),L(t_n), W_{i}(t_n,u)\colon u\in[0,\ h_i],\epsilon\text{ divides } u\}$ . While the initial condition will be introduced in Section 2.6, we now describe how the variables including $L(t_n), F(t_n)$ , and $W_i(t_n,u)$ are updated at each time step. First, consider the number of free tips $F(t_n)$ . This increases because some vertices complete their POWs, and it decreases because some free tips are selected as parents. Therefore we define

(2.7) \begin{align}F(t_{n+1}) - F(t_n) = \sum_{i=1}^{M} N_i(t_n-h_i)-\sum_{i=1}^{M}F_i(t_n).\end{align}

Next, consider the number of tips $L(t_n)$ , which increases because some vertices complete their POWs and decreases because some tips are attached by other vertices and cease to be tips. To model this mechanism, we define

(2.8) \begin{align}L(t_{n+1}) - L(t_n) = &\sum_{i=1}^M N_i(t_n-h_i) -\sum_{i=1}^M F_i(t_n-h_i)\nonumber\\&-\sum_{j < i}\sum_{u=h_j+\epsilon}^{h_i-\epsilon}J_{i,j}(t_n-h_j,u)+\sum_{j < i}\sum_{u=h_j+\epsilon}^{h_i-\epsilon}J_{i,j}(t_n-u,u)\end{align}

by considering the following scenarios: (1) $\sum_{i=1}^{M} N_i(t_n-h_i)$ counts how many vertices should finish their POWs at $t_n$ and become free tips at time $t_{n+1}$ ; (2) if a free tip was selected as a parent by a vertex whose POW will be finished at time $t_n$ , then it might be the case that this free tip ceases to be a tip due to the completion of the POW directed to it. We use $\sum_{i=1}^M F_i(t_n-h_i)$ to consider this situation; (3) some of the Type $i>1$ pending tips that are anticipated to finish POW after $t_{n}$ may have been selected by Type $j < i$ arrivals at time $t_n-h_j$ , thus changing to Type j pending tips and hence are validated at time $t_n$ . Such consideration corresponds to the first double summation in the second line of (2.8); (4) in scenarios (2) and (3), some of the Type i pending tips created at time $t_n - h_i$ may have been changed to Type $j < i$ at some time before $t_n-h_i$ , and thus have completed their POW prior to time $t_n$ . These tips should be removed from the count of the reduction of $L(t_n)$ and are taken into account through the last term in (2.8). Scenario (4) demonstrates how random POW durations complicate the model.

Next, we consider $W_i(t_{n}, u)$ , the number of Type i pending tips with RLT u at time $t_n$ . The value $W_i\big(t_{n+1}, h_i-\epsilon\big)$ consists of: (1) the free tips selected by Type i arrivals at $t_{n}$ and hence become Type i pending tips at time $t_{n+1}$ ; (2) the pending tips that just change to Type i at time $t_{n+1}$ . Hence we define

(2.9) \begin{align}W_i(t_{n+1}, h_i-\epsilon) \,:\!=\,F_i(t_n)+\sum_{j>i}\sum_{u=h_i}^{h_j-\epsilon} J_{j,i}(t_n,u).\end{align}

The Type i pending tips with RLT u will have a lower RLT at the next step but some of them will change to Type j for some $j < i$ . Hence we define

(2.10) \begin{align}W_i(t_{n+1}, u - \epsilon) \,:\!=\, W_i(t_{n}, u) - \sum_{h_j\leq u}J_{i,j}(t_n, u)\quad\text{for }{u\leq h_i-\epsilon}.\end{align}

If a pending tip has RLT equal to zero, it will no longer be a tip at the next step and hence ${W}_i(t_n, u)=0$ for $u<0$ . Furthermore, if the RLT of a Type i pending tip is less than $h_k$ then it cannot change to a Type k pending tip at any later time. In particular, if the RLT of a Type i pending tip is less than $h_1$ then it will always be Type i until it is no longer a tip. Therefore, we define $W_i(t_n + s, u-s) = W_i(t_n, u)$ for all $u<h_1$ , $s<u$ .

The assumption of deterministic arrival and its generalization to binomial arrival lead to tractable update rules as described by (2.7)–(2.10), where each right-hand side of the update rules has finite terms. In contrast, Poisson arrivals introduce an unbounded number of possible arrivals in the update rules and continuous-time dynamics, both of which complicate the analysis.

Since a tip is either a pending tip or a free tip, we have $W(t_n)=L(t_n)-F(t_n)$ , and hence knowing (2.7)–(2.10) is enough. However, for the purpose of studying the fluid limit that will be introduced in Section 2.5, we introduce an alternative equation to calculate the number of pending tips, i.e. $W(t_n)$ . Note that a pending tip must be a free tip at a prior time. This implies that a pending tip at $t_n$ must correspond to a free tip that is selected as a parent by a Type i arrival in the time range $[t_n-h_i,\ t_n-\epsilon]$ for some i. Among the free tips that are selected as parents by any Type i arrivals in $[t_n-h_i,\ t_n-\epsilon]$ for some i, some of them might be selected again and get attached before or at time $t_n$ . So we need to subtract the number of pending tips that have anticipated attachment time after $t_n$ but changed to pending tips with lower Type and got attached before $t_n$ . These quantities that need to be subtracted are calculated by using $J_{i,j}$ . To illustrate, we first give an example using $M=2$ , i.e. there are two possible values of POW duration.

For the special case of $M=2$ , the number of pending tips is calculated as

\begin{align*}W(t_n) = \sum_{i=1}^2\sum_{s=\epsilon}^{h_i}F_i(t_n-s)-\sum_{s=h_1+\epsilon}^{h_2-\epsilon}\sum_{u=s}^{h_2-\epsilon} J_{2,1}(t_n-s,u)\!,\end{align*}

where the range of $J_{2,1}(t_n-s,u)$ should satisfy the following conditions: (1) we subtract $J_{2,1}(t_n-s,u)$ because the Type 2 pending tips were anticipated to finish POW at time $\geq t_n$ but changed to Type 1 and hence finished before $t_n$ . So $t_n-s+u\geq t_n$ and $t_n-s+h_1<t_n$ ; (2) since the greatest possible RLT is $h_2-\epsilon$ , we have $u\leq h_2-\epsilon$ ; (3) when $u=h_2-\epsilon$ , by (1) we need $t_n-s+h_2-\epsilon\geq t_n$ . Hence $s\leq h_2-\epsilon$ .

For $M>1$ , repeating the argument by generalizing the pair (2,1) to (i, j) yields

(2.11) \begin{align}W(t_n) = \sum_{i=1}^M\sum_{s=\epsilon}^{h_i}F_i(t_n-s)-\sum_{i=2}^M\sum_{j=1}^{i-1}\sum_{s=h_j+\epsilon}^{h_i-\epsilon}\sum_{u=s}^{h_i-\epsilon} J_{i,j}(t_n-s,u) .\end{align}

2.5. The fluid limit

We are interested in the model considering the random variables $L(t_n)$ , $F(t_n)$ , $W_i(t_n,u)$ in the limit where $\lambda, N, \epsilon^{-1} \rightarrow \infty$ . In particular, we will first make a reasonable conjecture for the fluid limit based on the expected values and then establish that the fluid limit is close to the random process. For the random variables, we focus on the leading-order term in their expected values and assume that the remainder terms are negligible in the limit $\lambda, N, \epsilon^{-1} \rightarrow \infty$ . Using the distributions from Sections 2.2 and 2.3, we obtain the following leading-order terms:

(2.12) \begin{align}N_i(t_n)& = N p_i + \cdots,\nonumber \\F_i(t_n) &= 2 \, N \, p_i \, \frac{F(t_n)}{L(t_n)} + \cdots,\nonumber \\J_{i,j}(t_n, u) &= 2 \, N \, p_j \, \frac{W_{i}(t_n, u)}{L(t_n)} + \cdots, \quad\quad \text{for } i > j,\ h_j\leq u.\end{align}

Note that the 2 multiplier appears in (2.12) and later derivations because two parents are selected for each vertex. If k parents are selected for each vertex instead, 2 will be replaced by k.

We expect that there are deterministic functions $l(t),f(t),w_i(t,u)$ such that $l(t_n)\approx L(t_n)/\lambda$ , $f(t_n) \approx F(t_n)/\lambda$ , and $w_i(t_n, u) \approx W_i(t_n, u)/N$ . By the update rules in (2.7)–(2.11), taking expected values on the right-hand side and only keeping the leading-order term in each expected value yields

(2.13) \begin{align}&f(t_{n+1}) - f(t_n) = \epsilon - 2 \, \epsilon \, \frac{f(t_n)}{l(t_n)} \nonumber,\\&l(t_{n+1}) - l(t_n) = \epsilon - 2 \, \epsilon \sum_{i=1}^M p_i\frac{f(t_n-h_i)}{l(t_n-h_i)} - 2 \, \epsilon \sum_{j < i}\sum_{u=h_j+\epsilon}^{h_i-\epsilon}\epsilon p_j\frac{w_i(t_n-h_i,u)}{l(t_n-h_i)} \nonumber \\&\quad\quad\quad\quad\quad\quad\quad\quad + 2 \, \epsilon \sum_{j < i}\sum_{u=h_j+\epsilon}^{h_i-\epsilon} \epsilon p_j\frac{w_i(t_n-u,u)}{l(t_n-u)}\nonumber,\\&w_i(t_{n+1}, u-\epsilon) - w_i(t_n, s) = - 2 \, \epsilon \,\sum_{h_j\leq u} p_j\frac{w_i(t_n,u)}{l(t_n)} \nonumber,\\&w_i(t_{n+1},h_i-\epsilon) = 2 \, p_i \, \frac{f(t_n)}{l(t_n)}+\sum_{j>i}\sum_{u= h_i}^{h_j-\epsilon}2p_i\frac{\epsilon w_j(t_n,u)}{l(t_n)}\nonumber,\\&w(t_n) =\sum_{i=1}^{M}\sum_{s=\epsilon}^{h_i}2\epsilon p_i\frac{f(t_n-s)}{l(t_n-s)}-2\sum_{i>j}\sum_{s=h_j+\epsilon}^{h_i-\epsilon}\sum_{u=s}^{h_i-\epsilon} \epsilon^2p_j\frac{w_i(t_n-s,u)}{l(t_n-s)}.\end{align}

We now formally take the limit $\lambda, N, \epsilon^{-1} \rightarrow \infty$ . The asymptotic equations (2.13) lead to the following delayed partial differential equations (PDEs), and we call the corresponding solution to the PDEs the fluid limit of the random DAG model:

(2.14) \begin{align}\frac{\mathrm{d}f}{\mathrm{d}t} & =1-2\frac{f(t)}{l(t)},\end{align}
(2.15) \begin{align}\frac{\mathrm{d}l}{\mathrm{d}t} & =1-2\sum_{i=1}^M p_i\frac{f(t-h_i)}{l(t-h_i)}-2\sum_{j < i}\int_{h_j}^{h_i}p_j\frac{w_i(t-h_i,u)}{l(t-h_i)}\,\mathrm{d}u\nonumber\\&\quad +2\sum_{j < i}\int_{h_j}^{h_i}p_j\frac{w_i(t-u,u)}{l(t-u)}\,\mathrm{d}u,\end{align}

(2.16) \begin{align}&\left(\frac{\partial}{\partial t}-\frac{\partial}{\partial u}\right)w_i(t,u)=-2\sum_{h_j\leq u} p_j\frac{w_i(t,u)}{l(t)},\\[-13pt]\nonumber\end{align}
(2.17)
(2.18)

As mentioned in Section 2.2, the assumption of deterministic arrivals can be generalized to binomial arrivals. This generalization can be carried out by multiplying the expected values in (2.12) by $1-p_0$ , similar to how $p_i$ appears in the calculations. The ansatz for the fluid limit and the proof of its approximation quality in Section 5 then proceed similarly.

In order to show the existence and uniqueness of the solution to the PDEs, we will first define the initial condition in Section 2.6.

2.6. Initial conditions

Both the random process of $\{F(t_n),L(t_n), W_{i}(t_n,u)\colon u\in[0,\ h_i], \epsilon \text{ {divides} } u\}$ defined in (2.7)–(2.10) and their corresponding PDEs that describe the fluid limit should be supplemented with an initial condition. The initial condition of the random process, which can be found by observing the update rules, is defined as follows:

\begin{align*}&F(t_j)=x_j,\quad\quad j=-2h_M\epsilon^{-1},\ldots,0,\\&L(t_j)=y_j,\quad\quad j=-2h_M\epsilon^{-1},\ldots,0,\\&W_i(t_0,k\epsilon)=\omega_k, \quad\quad k=1,2,\ldots, h_i\epsilon^{-1}-1,\\&N_i(t_j)=a_{i,j}, \quad\quad j=-h_M\epsilon^{-1},-h_M\epsilon^{-1}+1,\ldots,-1,\\&F_i(t_j)=b_{i,j}, \quad\quad j=-h_M\epsilon^{-1},-h_M\epsilon^{-1}+1,\ldots,-1,\\&J_{i,r}(t_j,k\epsilon)=c_{i,r,j,k}, \quad\quad j=-h_M\epsilon^{-1},-h_i\epsilon^{-1}+1,\ldots,-1,\\&\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad k=h_i\epsilon^{-1}-1,h_i\epsilon^{-1}-2,\ldots,h_r\epsilon^{-1}.\end{align*}

Fixing these values is sufficient to generate the future of the model using (2.7)–(2.10).

Definition 2.1. An initial condition for (2.7)–(2.10) is a proper initial condition $\underline{\mathrm{with respect to}\ \theta_1}$ if the requirements stated in equations (5.2), (5.3), (5.7), (5.8), (5.11), (5.12), and (5.14) hold for given $\theta_1$ .

The requirements assumed in (5.2) , (5.3), 5.7, (5.8), (5.11), (5.12), and (5.14), which are stated in the lemmas where they are needed, are reasonable assumptions of initial conditions which require that some quantities in the initial condition are close to their expected values when they are generated. The validity of these conditions is stated in Proposition 2.1.

Proposition 2.1. Given any $\theta_1>0$ , there exist $\lambda^*$ and $N^*$ such that for $\lambda>\lambda^*$ and $N>N^*$ , the collection $\mathbb{P}i(\theta_1)$ of proper initial conditions with respect to $\theta_1$ is nonempty.

The proof of Proposition 2.1 is equivalent to the proof of the parallel statements that arise within the intervals $t\geq0$ instead of $t\leq 0$ . For illustration, see Lemmas 5.2 and 5.3, where Lemma 5.2 assumes the initial condition specified by (5.2) and Lemma 5.3 validates this assumption. The core idea is that the initial conditions are also randomly generated using the described dynamics, and therefore the properties that are established in $t\geq0$ can be applied to $t\leq0$ as well.

For the initial condition of the PDEs corresponding to the fluid limit, we first let $f(t_j)=F(t_j)/\lambda, l(t_j)=L(t_j)/\lambda$ for $t_j=-2h_M,\ldots,0$ . One can then use interpolations such as the Lagrange polynomial or cubic spline to construct f(t), l(t) for $t\in[{-}2h_M,\,0]$ so that both functions are continuously differentiable. By following (2.16) and (2.17), we can generate $w_i(t,u)$ for $t\in[{-}h_M,\,0]$ as part of the initial condition for the fluid limit.

Using the constructed initial condition for the fluid limit, we will use a similar approach as in [Reference King14] based on a method called the method of steps [Reference Driver5] to show the existence and uniqueness of the fluid limit. We first consider $f,l,w_i$ in the interval $t\in[0,\ h_1]$ . Note that l(t) for $t\in[0,\ h_1]$ can be directly computed using (2.15) because the right-hand side only considers initial conditions. Together with (2.14), the solution for f(t) within $t\in[0,\ h_1]$ can be calculated. Then l(t), f(t) within $t\in[0,\ h_1]$ along with the initial conditions can then be used to solve $w_i(t,h_i)$ within $t\in[0,\ h_1]$ using (2.16) and (2.17). Using the initial conditions and the solutions $f,l,w_i$ within the interval $t\in[0,\ h_1]$ , repeating the same procedure leads to the solution in $t\in[h_1,\ 2h_1]$ , and so on iteratively.

3. Main Results

We now state the main result of this paper, which establishes that the scaled random process is close to its fluid limit under suitable conditions.

Theorem 3.1. Let [0, T] denote a fixed time interval with T being the time horizon. Let $\delta,\gamma, \xi,m>0$ . Define

(3.1) \begin{align}g(t)=\sup_{0 < s\leq t}\left\{\bigg|\frac{F(s)}{\lambda}-f(s)\bigg|+\bigg|\frac{L(s)}{\lambda}-l(s)\bigg|+\sum_{i=1}^M \sup_{u\leq h_i} \bigg|\frac{W_i(s,u)}{N}-w_i(s,u)\bigg| \right\}\!,\end{align}

and $n_T=\epsilon^{-1}T$ (assuming it is an integer for simplicity). Then there exist $\alpha,\beta>0$ independent of $\lambda,N,\epsilon$ such that

(3.2) \begin{align}\mathbb{P}(g(T)>\delta)\leq& \beta\epsilon^{-1}\exp\left\{{-}\beta \frac{\theta^2}{T\epsilon} \right\}+\beta T\epsilon^{-1}\exp\left\{{-}\beta\lambda\right\}\nonumber\\&+\beta \exp\big\{{-}\beta\theta^2N\big\}+\beta T\frac{1}{\epsilon^3N\theta^2}+\beta T\frac{1}{\epsilon^3N}\exp\big\{{-}\beta\theta^2\epsilon^2N\big\}\end{align}

if the following conditions hold: (1) the initial condition for the random process as in Section 2.6 is a proper initial condition with respect to $\theta=(\delta e^{-\alpha T})/(3\alpha)$ ; (2) the fluid limit within the interval [0,T] and its initial condition satisfy (3.3)–(3.5) given by

(3.3) \begin{align}\bigg|\frac{\partial}{\partial u}\frac{w_i(t-u,u)}{l(t-u)}\bigg|&\leq \gamma \quad \text{for } t\in[0,\ T],\ u\in(0,h_M),\end{align}
(3.4) \begin{align}\frac{w_i(t,u)}{l(t)}&\leq \xi \quad \text{for } t\geq [{-}h_M,T], \ u\in[0,\ h_i],\end{align}
(3.5) \begin{align}l(t)&\geq m \quad \text{for all } t\in[{-}h_M,\ T).\end{align}

While the complete proof of Theorem 3.1 will be provided in Section 5, we will also provide an introduction to the main idea of the proof in Section 4.

Equation (3.2) provides a probability bound for understanding the difference g(T) between the random process and its fluid limit. It implies that the difference goes to zero in probability provided that $\epsilon^{-1},\lambda,N,\epsilon^3N\rightarrow \infty$ . Since $\lambda=N/\epsilon$ , (3.2) implies that it suffices for N to go to infinity faster than $\epsilon^{-3}$ . In addition, the difference g(T) goes to zero in probability for any finite time horizon T, but may diverge if T also goes to infinity. Our parametrization with $\epsilon \rightarrow 0$ and fixed $\{h_i\}_{i=1,\ldots,M}$ makes this result interpretable in terms of real-world timescales. In contrast, with $\epsilon=1$ and $T, \{h_i\}_{i=1,\ldots,M} \rightarrow \infty$ , time is measured in units of inter-arrival intervals, making it difficult to interpret what constitutes a ‘fixed time horizon’ in natural time units.

By Theorem 3.1, one can predict the evolution of the random process either by using the fluid limit, which is the solution to a set of delayed partial differential equations, or by using a simulation result of the random process. Notice that for simulation, instead of actually performing the POW, which contributes to almost all the computational power consumed in the system, it suffices to generate a random duration of POW. Figure 4 provides an example.

Figure 4. Simulations with $\lambda=400, N=20, \epsilon=0.05$ where the vertical axis displays values for $L(t)/\lambda,F(t)/\lambda$ and the horizontal axis represents time $t_n$ . Here, M is assumed to be 2, corresponding to Proposition 3.1 The panel (a) corresponds to the simulation with $p_1=0.8$ and $p_2=0.2$ while the panel (b) corresponds to the simulation with $p_1=0.3$ and $p_2=0.7$ . Both simulations assume $h_1=3$ and $h_2=5$ . Multiple simulations are performed with the same parameters and, we can see that the scaled random processes behave almost deterministically with minor deviation.

However, studying the equilibrium point of the fluid limit can provide useful insight into the random process. By assuming $f(t)=f$ , $l(t)=l$ , $w(t)=l-f$ , and $w_i(t,u)=w_i(u)$ , we can solve the fluid limit equations and derive the equilibrium point. Proposition 3.1 provides a result when M, the number of possible POW durations, is 2.

Proposition 3.1. When $M=2$ , the equilibrium point of (2.14)–(2.17) is given by

\begin{align*} w_1(t,u)&=p_1,f(t)=w(t)=\frac{l(t)}{2},\\[4pt] w_2(t,u)&=p_2\exp{\left\{{-}\frac{2p_1}{l}(h_2-u)\right\}}\quad\text{for } u\in[h_2-h_1,\ h_2],\\[4pt] \left(1 - \frac{p_2}{p_1}\right) l &= 2 h_1 - \frac{p_2}{p_1} l \exp\left\{{-} \frac{2 p_1}{l} (h_2 - h_1)\right\}\!. \end{align*}

Proposition 3.1 and Figure 4 demonstrate that the parameters for POW durations including $p_i$ and $h_i$ determine the values of f(t), w(t), l(t) in stable state where the property that $f(t)=w(t)=l(t)/2$ remains fixed. As demonstrated by Figure 4, f(t), l(t), w(t) in a stable state have lower values when $\sum_{i=1}^Mh_ip_i$ is smaller. This is because tips are attached by future vertices at a higher rate when the expected value of POW duration $\sum_{i=1}^Mh_ip_i$ decreases, and therefore fewer tips will be waiting to be attached.

For the equilibrium point when $M>2$ , $f(t)=w(t)=l(t)/2$ holds by (2.14), and $w_i(t,u)$ first decays exponentially and then remains constant as u changes because of (2.16). It is worth noting that when k parents are selected for each arriving vertex instead of 2, (2.12) implies that $f(t)=l(t)/k$ in the stable state. In addition, the stable state value l in Proposition 3.1 coincides with the finding of [Reference Penzkofer and Saa20], which studies the stable state of tips assuming the arrival of vertices follows a Poisson process.

3.1. Discussion

While Theorem 3.1 assumes that if POW durations are independent, identically distributed with bounded support on M finite values, where M can be any finite number, the result can be extended to a distribution with unbounded support. One can construct a sequence of distributions with $M\rightarrow \infty$ so that the sequence converges to the unbounded distribution for POW durations. Each distribution with finite M can be seen as a truncated version of the unbounded distribution. By showing that the random process with bounded POW durations can approximate the process with unbounded POW durations within the fixed time horizon [0, T], Theorem 3.1 extends to unbounded distributions.

There are multiple future directions to extend this topic. For example, [Reference King14] has investigated how fast the fluid limit converges to a stable state when the POW has a fixed duration. Whether the convergence rate remains the same when POW is random is an interesting question. In addition, adversarial vertices are an important topic in DL systems, particularly in relation to security. This has been studied in [Reference Sankagiri, Gandlur and Hajek22] for blockchain systems but has not yet been explored in the context of IOTA and other similar protocols. For example, [Reference Dembo3, Reference Göbel, Keeler, Krzesinski and Taylor11] investigate the security of blockchain under adversarial attack by including the consideration of communication delay. Because of the differences between the attachment rules for blockchain and DAG-based protocols, many results from blockchain systems cannot be directly applied to protocols with DAG representation. Therefore, conducting deeper investigations into IOTA and similar systems by considering adversarial attack and communication delay provides valuable insights into the performance differences between DAG-based protocols and blockchain.

4. Main Idea of Proof

We first introduce the main idea of the proof for Theorem 3.1 which has also been used in [Reference King14]. Consider a stochastic process $\big\{X(t_n)\big\}_n$ that depends on a parameter $\lambda$ with $t_n=n/\lambda$ and $X(t_n)\rightarrow \infty$ as $\lambda\rightarrow \infty$ . The goal is to illustrate the idea of bounding $\big|X(t_n)/\lambda-x(t_n)\big|$ where x(t) represents the solution to the differential equation as the fluid limit. We construct the initial condition of the fluid limit such that $x(t_0)=X(t_0)/\lambda$ . We first use a telescoping sum to decompose the difference $\big|X(t_n)/\lambda-x(t_n)\big|$ :

\begin{align*}\bigg|\frac{1}{\lambda} X(t_n)-x(t_n)\bigg|\leq\Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1} \left[ X(t_{j+1})-X(t_{j})-\lambda\big(x(t_{j+1})-x(t_{j})\big) \right]\Bigg|.\end{align*}

Denote $\Delta(X,j)\,:\!=\, X(t_{j+1})-X(t_{j})$ . Let $\mathcal{F}_n$ denote the $\sigma$ -algebra generated by all $X(t_k)$ up to time $t_n$ . We then have

(4.1) \begin{align}&\bigg|\frac{1}{\lambda} X(t_n)-x(t_n)\bigg|\leq \Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1} \bigg[ \Delta(X,j)-\lambda\int_{t_j}^{t_{j+1}}x'(s)\,\mathrm{d}s\bigg] \Bigg|\\&\leq \Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1} \bigg[ \Delta(X,j)-\mathbb{E}\big[\Delta(X,j)|\mathcal{F}_{j}\big]+\mathbb{E}\big[\Delta(X,j)|\mathcal{F}_{j}\big]-\lambda\int_{t_j}^{t_{j+1}}x'(s)\,\mathrm{d}s\bigg] \Bigg|\nonumber\\&\leq \Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1} \Delta(X,j)-\mathbb{E}\big[\Delta(X,j)|\mathcal{F}_{j}\big]\Bigg|+\Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1}\bigg[ \mathbb{E}\big[\Delta(X,j)|\mathcal{F}_{j}\big]-\lambda\int_{t_j}^{t_{j+1}}x'(s)\,\mathrm{d}s\bigg] \Bigg|\nonumber.\end{align}

The difference $\big|X(t_n)/\lambda-x(t_n)\big|$ has been separated into two terms, where the first absolute value represents the fluctuations and the second term represents the differences between expected values and their fluid limit. We will refer to such a technique as the separation technique. We then apply probability inequalities such as Lemma 5.1 to get that for any $\theta>0$ ,

(4.2) \begin{align}\mathbb{P} \Bigg(\Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1} \Delta(X,j)-\mathbb{E}\big[\Delta(X,j)|\mathcal{F}_{j}\big]\Bigg|\geq \theta \Bigg)\leq o(\lambda^{-1}).\end{align}

For the second part in (4.1), with the construction of the fluid limit we can derive that

(4.3) \begin{align}\Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1} \mathbb{E}\big[\Delta(X,j)|\mathcal{F}_{j}\big]-\lambda\int_{t_j}^{t_{j+1}}x'(s)\,\mathrm{d}s \Bigg|\leq o(\lambda^{-1})\sum_{j=0}^{n-1} g(t_j)+o(\lambda^{-1}),\end{align}

where $g(t)=\sup_{0\leq s\leq t}\big|\lambda^{-1}X(t)-x(t)\big|$ . In this paper, the inequality (4.3) holds because the fluid limit is derived by dropping lower-order terms in the conditional expectation and taking the limit, where the things we dropped are bounded by the term $o(\lambda^{-1})$ and the difference between the remaining terms and the fluid limit will be bounded by $o(\lambda^{-1})\sum_{j=0}^{n-1}g(t_j)$ .

For T denoting the time horizon, generalizing $\big|\lambda^{-1}X(t_n)-x(t_n)\big|$ to $\big|\lambda^{-1}X(t)-x(t)\big|$ yields, on some event that has probability going to one,

\begin{align*}g(T)\leq \theta+o(\lambda^{-1})\sum_{j=0}^{\lfloor T\lambda^{-1}\rfloor-1} g(t_j)+o(\lambda^{-1}).\end{align*}

We then use discrete Gronwall’s lemma with a suitable choice of $\theta$ to get the desired upper bound of $\mathbb{P}\left(g(T)>\delta\right)$ .

5. Proof of Theorem 3.1

Throughout the discussion, let T be the time horizon and, without loss of generality, denote the integer $n_T=T\epsilon^{-1}$ . Also, we define $\sigma_n$ to be the $\sigma$ -algebra generated by $N_i(t_k),F_i(t_k)$ for $k< n$ and $i\leq M$ , as well as $J_{i,j}(t_k,u)$ for $k< n,j < i\leq M$ , and $u\in[h_j,h_i)$ , which corresponds to the parents selection and POW duration generated for all arrivals that arrive before $t_n$ . Note that by the evolution equations (2.7)–(2.10), the values of $F(t_n)$ , $L(t_n)$ , and $W_i(t_n,u)$ are $\sigma_{n}$ -measurable.

To prove Theorem 3.1, we will first analyze the difference between the scaled number of free tips $F(t_n)/\lambda$ and its fluid limit $f(t_n)$ , followed by generalizing the differences at discrete times to the continuous time interval $s\in [0,T]$ . This will be done in Section 5.1. We will repeat this process for the other terms that appear in (3.1), where the work in Section 5.2 corresponds to the terms for the number of tips and that in Section 5.3 corresponds to the numbers of pending tips. Finally, we will put them together and use Gronwall’s lemma to establish the theorem in Section 5.4.

We first introduce an important lemma that will be used repeatedly throughout the proof to provide probability bounds like Lemmas 5.2 and 5.4.

Lemma 5.1. Let $\{\sum_{i=1}^{n}Y_i\}_{n\in\mathbb{N}}$ be a martingale and let $|Y_i|\leq 4N$ , then by the Azuma–Hoeffding inequality [Reference Williams26],

\begin{align*}\mathbb{P}\Bigg\{\sup_{0\leq n\leq n_T}\Bigg|\sum_{j=1}^{n}Y_j\Bigg|\geq \theta\Bigg\}\leq 2\exp\left\{{-}\frac{\theta^2}{2(n_T)(16N^2)}\right\}\\\implies\mathbb{P}\Bigg\{\sup_{0\leq n\leq n_T}\Bigg|\frac{1}{\lambda}\sum_{j=1}^{n}Y_j\Bigg|\geq \theta\Bigg\}\leq 2\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}\!.\end{align*}

5.1. Difference in number of free tips

In this section, we establish the upper bound of $\big|\lambda^{-1}F(t_n)-f(t_n)\big|$ , which will be stated in (5.5). In order to do that, we will follow the idea introduced through (4.1)–(4.3). From now on, we assume that $n>0$ :

(5.1) \begin{align}\frac{1}{\lambda}F(t_n)-f(t_n) =&\frac{1}{\lambda}\sum_{j=0}^{n-1}F(t_{j+1})-F(t_j)-\lambda(f(t_{j+1})-f(t_j))\nonumber\\=&\frac{1}{\lambda}\sum_{j=0}^{n-1}\left(\sum_{i=1}^{M} N_i(t_j-h_i)-\sum_{i=1}^{M}F_i(t_j)-\lambda\int_{t_j}^{t_{j+1}} 1-2\frac{f(s)}{l(s)}\,\mathrm{d}s\right)\nonumber\\=&\frac{1}{\lambda}\sum_{j=0}^{n-1}\left(-N+\sum_{i=1}^{M} N_i(t_j-h_i)\right)+\sum_{i=1}^{M}\frac{1}{\lambda}\sum_{j=0}^{n-1}\big(\mathbb{E}\big[F_i(t_j)|\sigma_j\big]-F_i(t_j)\big)\nonumber\\&+\sum_{i=1}^{M}\frac{1}{\lambda}\sum_{j=0}^{n-1}\left(\lambda\int_{t_j}^{t_{j+1}} 2p_i\frac{f(s)}{l(s)}\,\mathrm{d}s-\mathbb{E}\big[F_i(t_j)|\sigma_j\big]\right)\!,\end{align}

which is the result following the idea in (4.1). We now establish Lemmas 5.2, 5.4, 5.5, and 5.6, which provide probability bounds on each summation in (5.1). All the lemmas from here on will be proved in Section 6.

Lemma 5.2. Let $D_n=\lambda^{-1}\sum_{j=0}^{n-1} (\!-N+\sum_{i=1}^M N_i(t_j-h_i))$ . Assume $\epsilon|h_i$ and let $n_i=h_i\epsilon^{-1}$ . If for the initial condition we have for all $i=1,2,\ldots,M$ , that

(5.2) \begin{align}\sup_{-n_i\leq k< 0}\Bigg|\frac{1}{\lambda}\sum_{j=-n_i}^{k} (N_i(t_j)-p_i N)\Bigg|<\theta_1 ,\end{align}

then

\begin{align*}\mathbb{P}(E_1)\,:\!=\,\mathbb{P}\Big(\sup_{0 < n\leq n_T}|D_n|\geq M\theta+M\theta_1\Big)\leq 2M\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}\!.\end{align*}

The validity of the assumption in (5.2) is given in Lemma 5.3, whose proof is analogous to the proof of Lemma 5.2. This similarity arises from the fact that both the random process after time $t=0$ and its initial condition are randomly generated by the same rules described in Section 2.3. Based on the illustration using Lemmas 5.2 and 5.3, the lemmas and proofs of other assumptions on the initial conditions will be omitted.

Lemma 5.3. Suppose the initial condition of the random process described in Section 2.6 is generated according to the distributions described in Section 2.3. Then

\begin{align*}\mathbb{P}\left(\bigcup_{i=1}^{M}\Bigg\{\sup_{-n_i\leq k<0}\Bigg|\frac{1}{\lambda}\sum_{j=-n_i}^{k} \left(N_i(t_j)-p_iN\right)\Bigg|\geq\theta_1\Bigg\}\right)\leq 2M\exp\left\{{-}\frac{\theta_1^2}{32h_M\epsilon}\right\}\!,\end{align*}

which vanishes for $\epsilon\rightarrow 0$ .

Lemma 5.4. For any i, we have

\begin{align*}\mathbb{P}\big(E_{2,i}\big)&\,:\!=\,\mathbb{P}\Bigg\{\sup_{0< n\leq n_T}\Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1} \mathbb{E}\big[F_i(t_j)|\sigma_j\big]-F_i(t_j)\Bigg|\geq \theta \Bigg\}\leq \exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}\!,\\\mathbb{P}\big(E_2\big)&\,:\!=\,\mathbb{P}\Bigg(\bigcup_{i=1}^M E_{2,i}\Bigg)\leq 2M\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}\!.\end{align*}

Lemma 5.4 holds because $\mathbb{E}[F_i(t_j)|\sigma_j]-F_i(t_j)$ is a random variable that satisfies the requirements in Lemma 5.1.

Lemma 5.5. If for the initial condition we have

(5.3) \begin{align}\inf_{-h_M\epsilon^{-1}\leq n\leq \frac{3}{2}h_1\epsilon^{-1}}\sum_{j=n-2h_1\epsilon^{-1}}^{0} N_1(t_j)>1 ,\end{align}
\begin{align*}\text{then }\mathbb{P}\big(E_3\big)\,:\!=\,\mathbb{P}\Big(\inf_{-h_k\epsilon^{-1}\leq n\leq n_T}L_{\lambda}(t_n)\leq 1\Big)\leq T\epsilon^{-1} \exp\left\{{-}\frac{(1-p_1h_1/2)^2\lambda}{2h_1}\right\}\!.\end{align*}

Lemma 5.5 allows us to derive Lemma 5.6.

Lemma 5.6. Let

(5.4) \begin{align}H_{i,j+1}=-\mathbb{E}\big[F_i(t_j)|\sigma_j\big]+\lambda\int_{t_j}^{t_{j+1}} 2p_i\frac{f(s)}{l(s)}\,\mathrm{d}s.\end{align}

Then, on event $E_3^C$ , $\lambda^{-1}\sum_{j=0}^{n-1}|H_{i,j+1}|\leq 16\epsilon T+\sum_{j=0}^{n-1} 2p_i\epsilon g(t_{j+1})/m$ for all i.

Combining Lemmas 5.2, 5.4, 5.5, and 5.6 yields that, on the event $(\bigcup_{i=1}^3 E_i)^c$ ,

\begin{align*}\bigg|\frac{1}{\lambda}F(t_n)-f(t_n)\bigg|\leq 2M\theta+M\theta_1+16M\epsilon T+\sum_{j=0}^{n-1} \frac{2\epsilon g(t_{j+1})}{m}.\end{align*}

We can extend this inequality to $t\in[t_{n},\ t_{n+1})$ by defining that $F(t)=F(t_k)$ for $t\in[t_k,\ t_{k+1})$ :

(5.5) \begin{align}\bigg|\frac{1}{\lambda}F(t)-f(t)\bigg| &\leq\bigg|\frac{1}{\lambda}F(t_n)-f(t_n)\bigg|+|f(t_n)-f(t)|\nonumber\\[5pt]&\leq\bigg|\frac{1}{\lambda}F(t_n)-f(t_n)\bigg|+\bigg|\int_{t_n}^{t}1-2\frac{f(s)}{l(s)}\,\mathrm{d}s\bigg|\leq\bigg|\frac{1}{\lambda}F(t_n)-f(t_n)\bigg|+\epsilon+2\epsilon\nonumber\\[5pt]&\leq3\epsilon+2M\theta+M\theta_1+16M\epsilon T+\sum_{j=0}^{n-1} \frac{2\epsilon g(t_{j+1})}{m},\end{align}

where we use the property that $f(s)\leq l(s)$ .

5.2. Difference in number of tips

Next, by following a similar procedure as in Section 5.1, we establish (5.9) that provides an upper bound to $|L(t_n)/\lambda-l(t_n)|$ :

(5.6) \begin{align}\frac{1}{\lambda}L(t_n)-l(t_n) =&\frac{1}{\lambda}(L(t_n)-L(t_0))-(f(t_n)-f(t_0))\nonumber\\[5pt]=&\frac{1}{\lambda}\sum_{j=0}^{n-1}\bigg[L(t_{j+1})-L(t_j)-\lambda\int_{t_j}^{t_{j+1}}f'(t)\,\mathrm{d}t\bigg]\nonumber\\[5pt]=&\frac{1}{\lambda}\sum_{j=0}^{n-1}\left(\sum_{i=1}^{M} N_i(t_j-h_i)-N\right)\nonumber\\[5pt]&+\frac{1}{\lambda}\sum_{j=0}^{n-1}\left(-\sum_{i=1}^M F_i(t_j-h_i)+2\lambda\sum_{i=1}^M p_i\int_{t_j}^{t_{j+1}}\frac{f(t-h_i)}{l(t-h_i)}\,\mathrm{d}t\right)\nonumber\\[5pt]&+\frac{1}{\lambda}\sum_{j=0}^{n-1}\left[2\lambda\sum_{r < i}\!\int_{t_j}^{t_{j+1}}\!\int_{h_r}^{h_i}p_r\frac{w_i(t-h_i,u)}{l(t-h_i)}\,\mathrm{d}u\,\mathrm{d}t-\!\sum_{r < i}\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\!J_{i,r}(t_j-h_i,u)\right]\nonumber\\[5pt]&+\frac{1}{\lambda}\sum_{j=0}^{n-1}\left[2\lambda\sum_{r < i}\!\int_{t_j}^{t_{j+1}}\!\int_{h_r}^{h_i}p_r\frac{w_i(t-u,u)}{l(t-u)}\,\mathrm{d}u\,\mathrm{d}t-\sum_{r < i}\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\!J_{i,r}(t_j-u,u)\right],\end{align}

where the last equality is derived from (2.8) and (2.15). In (5.6), we have decomposed the difference $L(t_n)/\lambda-l(t_n)$ into four parts, and we will analyze each part on the right-hand side of (5.6) by subtracting and adding conditional expectations as mentioned in (4.1).

The first part in (5.6) is $\lambda^{-1}\sum_{j=0}^{n-1} (\sum_{i=1}^M N_i(t_j-h_i) -N)$ , which is already bounded using Lemma 5.1. The second part in (5.6) is

\begin{align*}\frac{1}{\lambda}&\sum_{j=0}^{n-1}\left[\left(2\lambda\sum_{i=1}^M p_i\int_{t_j}^{t_{j+1}}\frac{f(t-h_i)}{l(t-h_i)}\,\mathrm{d}t\right)-\sum_{i=1}^M F_i(t_j-h_i)\right]\\[6pt]& \quad =\frac{1}{\lambda}\sum_{i=1}^M \sum_{j=0}^{n-1}\left[\left(2 p_i\lambda\int_{t_j}^{t_{j+1}}\frac{f(t-h_i)}{l(t-h_i)}\,\mathrm{d}t\right)-F_i(t_j-h_i)\right.\\[6pt]& \qquad-\mathbb{E}\big[F_i(t_j-h_i)|\sigma_{j-n_i}\big] +\mathbb{E}\big[F_i(t_j-h_i)|\sigma_{j-n_i}\big]\Biggr],\end{align*}

whose summands are very similar to the objects in Lemmas 5.4 and 5.6 except that the time is $t-h_i$ instead of t, and hence some of the terms are initial conditions. Therefore, we first make some modifications based on Lemma 5.4 to establish Corollary 5.1, and then we directly apply Lemma 5.6 to obtain Corollary 5.2.

Corollary 5.1. By Lemma 5.4, if for any i we have

(5.7) \begin{align}\sup_{0 < n\leq n_i}\Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1} \mathbb{E}\big[F_i(t_j-h_i)|\sigma_j\big]-F_i(t_j-h_i)\Bigg|\leq \theta_1,\end{align}

then

\begin{align*}\mathbb{P}\big(E_{4,i}\big)&\,:\!=\,\mathbb{P}\left\{\sup_{0< n\leq n_T}\Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1} \mathbb{E}\big[F_i(t_j-h_i)|\sigma_{j-n_i}\big]-F_i(t_j-h_i)\Bigg|\geq \theta_1+\theta \right\}\\[5pt]&\leq 2\exp\left\{{-}\frac{\theta^2}{32 T\epsilon}\right\}\\[5pt]\mathbb{P}(E_4)&\,:\!=\,\mathbb{P}\left(\bigcup_{i=1}^M E_{4,i}\right)\leq 2M \exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}\!.\end{align*}

Note that the eligibility of 5.7 as an assumption for the initial conditions follows the same idea as shown in the proof of Lemma 5.2.

Corollary 5.2. By the proof of Lemma 5.6, on the event $E_3^c$ ,

\begin{align*}\frac{1}{\lambda} \sum_{j=0}^{n-1}\Bigg|\left(2 p_i\lambda\int_{t_j}^{t_{j+1}}\frac{f(t-h_i)}{l(t-h_i)}dt\right)-\mathbb{E}\big[F_i(t_j-h_i)|\sigma_{j-n_i}\big]\Bigg|\leq 16\epsilon T+\sum_{j=0}^{n-1} \frac{2p_i\epsilon g(t_{j+1})}{m}.\end{align*}

The third part in the right-hand side of (5.6) is

\begin{align*}\frac{1}{\lambda}\sum_{j=0}^{n-1}&\Bigg[2\lambda\sum_{r < i}\int_{t_j}^{t_{j+1}}\int_{h_r}^{h_i}p_r\frac{w_i(t-h_i,u)}{l(t-h_i)}\,\mathrm{d}u\,\mathrm{d}t-\sum_{r < i}\sum_{u=h_r+\epsilon}^{h_i-\epsilon}J_{i,r}(t_j-h_i,u)\Bigg]\\&=\sum_{r < i}\frac{1}{\lambda}\sum_{j=0}^{n-1}\Bigg[2\lambda\int_{t_j}^{t_{j+1}}\int_{h_r}^{h_i}p_r\frac{w_i(t-h_i,u)}{l(t-h_i)}\,\mathrm{d}u\,\mathrm{d}t-\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\mathbb{E}\big[J_{i,r}(t_j-h_i,u)|\sigma_{j-n_i}\big]\\&+\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\mathbb{E}\big[J_{i,r}(t_j-h_i,u)|\sigma_{j-n_i}\big]-\sum_{u=h_r+\epsilon}^{h_i-\epsilon}J_{i,r}(t_j-h_i,u)\Bigg],\end{align*}

which will be taken care of by Lemmas 5.7 and 5.8 with $h(u)=n_i$ .

Lemma 5.7. Given any pair $r < i$ , let h(u) be either $n_i\,:\!=\,h_i/\epsilon$ (so the function is constant with respect to u) or $u\epsilon^{-1}$ . If for the initial condition we have

(5.8) \begin{align}\sup_{0 < n\leq n_T}\Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1}\sum_{u=h_r+\epsilon}^{h_i-\epsilon}(\mathbb{E}[J_{i,r}(t_{j-h(u)},u)|\sigma_{j-h(u)}]-J_{i,r}(t_{j-h(u)},u))I_{\{j-h(u)\leq 0\}}\Bigg|<\theta_1 ,\end{align}

and we let $E_{5,i,r}(h(u)\epsilon)\,:\!=\,$

\begin{align*}\Bigg\{\sup_{0 < n\leq n_T}\Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1}\Bigg(\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\mathbb{E}[J_{i,r}\big(t_{j-h(u)},u\big)|\sigma_{j-h(u)}]-\sum_{u=h_r+\epsilon}^{h_i-\epsilon}J_{i,r}\big(t_{j-h(u)},u\big)\Bigg)\Bigg|\geq\theta_1+\theta\Bigg\}\!,\end{align*}

then

\begin{align*}&\mathbb{P}(E_{5,i,r}(h(u)\epsilon))\leq 2T\epsilon^{-1}\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}\!,\\&\mathbb{P}(E_{5,1})\,:\!=\,\mathbb{P}\left(\bigcup_{r < i}E_{5,i,r}(h_i)\right)\leq \frac{M^2}{2}\cdot2T\epsilon^{-1}\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}\!,\\&\mathbb{P}(E_{5,2})\,:\!=\,\mathbb{P}\left(\bigcup_{r < i}E_{5,i,r}(u)\right)\leq \frac{M^2}{2}\cdot2T\epsilon^{-1}\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}\!,\\&\mathbb{P}(E_5)\,:\!=\,\mathbb{P}\big(E_{5,1}\cup E_{5,2}\big)\leq M^2T\epsilon^{-1}\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}\!.\end{align*}

Lemma 5.8. Using the same setup for h(u) from Lemma 5.7, we get

\begin{align*}\Bigg| \frac{1}{\lambda} \sum_{j=0}^{n-1} & \Bigg[2\lambda\int_{t_j}^{t_{j+1}}\int_{h_r}^{h_i} p_r \frac{w_i(t-h(u)\epsilon,u)}{l(t-h(u)\epsilon)} \,\mathrm{d}u \,\mathrm{d}t-\sum_{u=h_r+\epsilon}^{h_i-\epsilon} \mathbb{E}\big[J_{i,r}(t_{j-h(u)},u)|\sigma_{j-h(u)}\big]\Bigg]\Bigg| \\{}& \leq 16\epsilon T+2Tp_r\epsilon(\xi+(h_i-h_r)^2\gamma)+6p_r(h_i-h_r)\epsilon \sum_{j=0}^{n-1}\frac{g(t_{j+1})}{m}.\end{align*}

For the fourth part of (5.6), we apply Lemmas 5.7 and 5.8 with $h(u)=u\epsilon^{-1}$ . Lemmas 5.2, 5.7, and 5.8 and Corollaries 5.1 and 5.2 cover all terms in (5.6). Putting them together, we get that, on the event $(E_{4}\cup E_{5})^c$ ,

\begin{align*}&\bigg|\frac{1}{\lambda}L(t_n)-l(t_n)\bigg|\leq (M^2+2M)(\theta_1+\theta)+16M\epsilon T+\sum_{j=0}^{n-1} \frac{2\epsilon g(t_{j+1})}{m}\\\leq& (M^2+2M)(\theta_1+\theta)+16(M^2+M)\epsilon T+4 M T\epsilon(\xi+\big(h_M - h_1\big)^2\gamma)\\&+\left[12M(h_M-h_1)\epsilon+2\epsilon\right]\sum_{j=0}^{n-1} \frac{ g(t_{j+1})}{m}.\end{align*}

We can extend this inequality to $t\in[t_n,\ t_{n+1})$ by setting $L(t)\,:\!=\,L(t_n)$ in this interval, yielding

(5.9) \begin{align}\bigg|\frac{L(t)}{\lambda}-l(t)\bigg|&\leq \bigg|\frac{L(t_n)}{\lambda}-l(t_n)\bigg|+\bigg|\int_{t_n}^t\frac{\mathrm{d}l}{\mathrm{d}s}\,\mathrm{d}s\bigg|\leq \bigg|\frac{L(t_n)}{\lambda}-l(t_n)\bigg|+\epsilon(1+2+4M(h_M-h_1)\xi)\nonumber\\&\leq 2M^2(\theta_1+\theta)+38M^2\epsilon T+4 M T\epsilon((h_M+1)\xi+(h_M)^2\gamma)\\\nonumber&\quad +\left[12 M h_M + 2\right]\epsilon\sum_{j=0}^{n-1} \frac{ g(t_{j+1})}{m}\\\nonumber&\leq 2M^2(\theta_1+\theta)+38M^2 T \epsilon +8 M T (h_M)^2(\xi+\gamma)\epsilon+13 M h_M \epsilon\sum_{j=0}^{n-1} \frac{ g(t_{j+1})}{m},\end{align}

where for simplicity of expression we assume $T\geq 1$ and $h_M>2$ .

5.3. Difference in number of pending tips

In this section, we will establish (5.20) as an upper bound of $|{W_i({\cdot})}/N-w_i({\cdot})|$ . As usual, we first rewrite the difference using a telescoping sum using (2.9)–(2.18):

\begin{align*}&\left[\frac{W_i(t_n,h_i-k\epsilon)}{N}-w_i(t_n,h_i-k\epsilon)\right]\nonumber\\[3pt]=& \frac{1}{N} \big\{W_i(t_n,h_i-k\epsilon)-W_i(t_n-(k-1)\epsilon,h_i-\epsilon) \nonumber\\[3pt]&-N \left[w_i(t_n,h_i-k\epsilon)- w_i(t_n-(k-1)\epsilon,h_i-\epsilon)\right]\big\}\nonumber\\[3pt]&+\frac{1}{N}\left[W_i(t_n-(k-1)\epsilon,h_i-\epsilon)-N w_i(t_n-(k-1)\epsilon,h_i-\epsilon)\right]\nonumber\\[3pt]=& \frac{1}{N} \sum_{j=1}^{k-1}\{W_i(t_{n-k+j+1},h_i- (j+1) \epsilon)-W_i (t_{n-k+j},h_i-j \epsilon)\nonumber\\[3pt]&-N[w_i(t_{n-k+j+1},h_i- (j+1) \epsilon)-w_i(t_{n-k+j},h_i-j \epsilon)]\}\\&+ \frac{1}{N}\left[F_i(t_{n-k})+\sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}J_{j,i}(t_{n-k},u)-N\left(2p_i\frac{f(t_{n-k})}{l(t_{n-k})}+\sum_{j>i} \int_{h_i}^{h_j}2p_i\frac{w_j(t_{n-k},u)}{l(t_{n-k})}\,\mathrm{d}u\right)\right.\nonumber\\[3pt]&\left.+2N \int_{0}^{\epsilon} \sum_{h_j\leq u} p_j\frac{w_i(t_{n-k}+u,h_i-u)}{l(t_{n-k}+u)}\,\mathrm{d}u\right]\nonumber\end{align*}

(5.10) \begin{align}=& \frac{1}{N} \sum_{j=1}^{k-1} \sum_{h_r\leq h_i-j\epsilon}\left\{{-}J_{i,r}(t_{n-k+j},h_i-j\epsilon)+N \int_{0}^{\epsilon}2 p_r\frac{w_i(t_{n-k+j}+u,h_i-j\epsilon-u)}{l(t_{n-k+j}+u)}\,\mathrm{d}u\right\}\nonumber\\[3pt]&+ \frac{1}{N}\left(F_i(t_{n-k})-2p_i N\frac{f(t_{n-k})}{l(t_{n-k})}\right)+ \frac{1}{N}\left(2N\int_{0}^{\epsilon} \sum_{h_j\leq u} p_j\frac{w_i(t_{n-k}+u,h_i-u)}{l(t_{n-k}+u)}\,\mathrm{d}u\right)\nonumber\\[3pt]&+\frac{1}{N}\left(\sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}J_{j,i}(t_{n-k},u)-N\sum_{j>i} \int_{h_i}^{h_j}2p_i\frac{w_j(t_{n-k},u)}{l(t_{n-k})}\,\mathrm{d}u\right)\!.\end{align}

In the last equality of (5.10), we have separated the difference into four terms, each of which is scaled by N. Using the separation technique to add and subtract conditional expectations, we will establish Lemmas 5.9 and 5.10 for the first scaled term and establish Lemmas 5.11 and 5.12 for the second scaled term. The fourth term will be covered by Lemma 5.15, which is derived from Lemmas 5.13 and 5.14. Last, we will establish an upper bound for the third scaled term in (5.15). Combining these results leads to the upper bound for (5.10) as given in (5.16). With (5.16), we are then able to derive an upper bound on the difference between the number of pending tips and its fluid limit as given in (5.20). Let

\begin{align*} G_{i,n,k,j}\,:\!=\,\sum_{h_r\leq h_i-j\epsilon}\big\{{-}J_{i,r}(t_{n-k+j},h_i-j\epsilon)+\mathbb{E}[J_{i,r}(t_{n-k+j},h_i-j\epsilon)|\sigma_{n-k+j}]\big\}\!.\end{align*}

Lemma 5.9. If for the initial condition we have

(5.11) \begin{align}&\sup_{i\leq M}\sup_{0 < n\leq n_T}\sup_{1\leq k\leq n_i}\Bigg|\frac{1}{N}\sum_{j=1}^{k-1} G_{i,n,k,j}\mathbf{1}_{\{n-k+j < 0\}}\Bigg|\leq\theta_1,\\[5pt]{then}\ \mathbb{P}\big(E_6\big)&\,:\!=\,\mathbb{P}\left(\sup_{i\leq M}\sup_{0 < n\leq n_T}\sup_{1\leq k\leq n_i}\Bigg| \frac{1}{N} \sum_{j=1}^{k-1} G_{i,n,k,j}\Bigg|\geq \theta_1+\theta\right)\nonumber\\[5pt]&\leq M T h_M^2 \frac{\epsilon^{-3}}{N}\left(512\frac{h_M^2}{\theta^2}+\exp\left\{{-}\frac{\theta^2 \epsilon^2 N}{8h^2_M}\right\}\right)\!.\nonumber\end{align}

Lemma 5.10.

\begin{align*}&\Bigg| \frac{1}{N} \sum_{j=1}^{k-1} \sum_{h_r\leq h_i-j\epsilon}\Bigg(-\mathbb{E}[J_{i,r}(t_{n-k+j},h_i-j\epsilon)|\sigma_{n-k+j}]\\&\quad\quad\quad\quad\quad\quad\quad\quad+ \int_{0}^{\epsilon}2 Np_r\frac{w_i(t_{n-k+j}+u,h_i-j\epsilon-u)}{l(t_{n-k+j}+u)}\,\mathrm{d}u\Bigg)\Bigg|\\\leq&16 T M \epsilon+2p_r N \epsilon^2 \gamma +6\epsilon\sum_{j=1}^{n-1} \frac{g(t_{j})}{m}.\end{align*}

Lemma 5.11. If for the initial condition we have

(5.12) \begin{align}\sup_{i\leq M} \sup_{n\in[{-}h_M\epsilon^{-1},0]} \bigg| \frac{1}{N} (F_i(t_{n})-\mathbb{E}[F_i(t_{n})|\sigma_{n}])\bigg|\leq \theta_1,\end{align}

then using the same proof as in Lemma 5.9 we get

\begin{align*}\mathbb{P}(F_i(t_n)-\mathbb{E}[F_i(t_n)|\sigma_{n}]> N\theta)\leq 256\epsilon\frac{1}{N\theta^2}+\exp{\left\{{-}\frac{\theta^2 N}{8}\right\}} \quad\text{for }n> 0,\end{align*}

and hence

\begin{align*}\mathbb{P}(E_7)\,:\!=\,\mathbb{P}\left(\sup_{i\leq M} \sup_{n\leq n_T} \sup_{k\leq h_M\epsilon^{-1}} \frac{1}{N} \left(F_i(t_{n-k})-\mathbb{E}[F_i(t_{n-k})|\sigma_{n-k}]\right)> \theta+\theta_1\right)\\\leq 256\frac{1}{N\theta^2}+\exp{\left\{{-}\frac{\theta^2 N}{8}\right\}}.\end{align*}

Lemma 5.12. By the proof of Lemma 5.6, we have

\begin{align*}\frac{1}{N}\bigg|\mathbb{E}\big[F_i(t_{n-k})|\sigma_{n-k}\big]-2p_i\frac{f(t_n-k)}{l(t_{n-k})}\bigg|\leq \frac{2p_i}{N} \frac{\big|f(t_{n-k})-F_\lambda(t_{n-k})\big|+ \big|(t_{n-k})-L_\lambda(t_{n-k})\big|}{m}.\end{align*}

We now establish Lemmas 5.13, 5.14, and 5.15 to cover the fourth scaled term in the last equality of (5.10). The fourth scaled term is

(5.13) \begin{align}\frac{1}{N}&\Bigg|\sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}J_{j,i}(t_{n-k},u)-N\sum_{j>i} \int_{h_i}^{h_j}2p_i\frac{w_j(t_{n-k},u)}{l(t_{n-k})}\,\mathrm{d}u\Bigg|\nonumber\\[5pt]\leq& \frac{1}{N}\Bigg|\sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}\left( J_{j,i}(t_{n-k},u)-\mathbb{E}\big[J_{j,i}(t_{n-k},u)|\sigma_{n-k}\big] \right)\Bigg|\nonumber\\[5pt] &+\frac{1}{N}\Bigg|\sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon} \mathbb{E}[J_{j,i}(t_{n-k},u)|\sigma_{n-k}]-N\sum_{j>i} \int_{h_i}^{h_j}2p_i\frac{w_j(t_{n-k},u)}{l(t_{n-k})}\,\mathrm{d}u\Bigg|.\end{align}

Lemma 5.13. If for the initial condition we have

(5.14) \begin{align}\sup_{n\in[{-}h_M\epsilon^{-1},\ 0]} \Bigg|\frac{1}{N}\sum_{j>i}\sum_{u=h_i}^{h_j-\epsilon} ( J_{j,i}(t_{n},u)-\mathbb{E}[J_{j,i}(t_{n},u)|\sigma_{n}])\Bigg| \leq \theta_1,\end{align}

then $\mathbb{P}\big(E_8\big)\,:\!=\,$

\begin{align*}&\mathbb{P}\left(\sup_{i}\sup_{n-k\leq n_T}\frac{1}{N}\Bigg|\sum_{j>i}\sum_{u=h_i}^{h_j-\epsilon} ( J_{j,i}(t_{n-k},u)-\mathbb{E}[J_{j,i}(t_{n-k},u)|\sigma_{n-k}])\Bigg|>\theta+\theta_1\right)\\&\leq 512 M^2 h_M\frac{1}{N\theta^2}+\exp\left\{{-}\frac{\theta^2N}{8}\right\}\!.\end{align*}

Next, we establish Lemma 5.14, which will be used in the proof of Lemma 5.15 and in the argument that extends (5.18)–(5.19).

Lemma 5.14. Consider $s_1,s_2\in[k\epsilon,\ (k+1)\epsilon)$ and $s_2>s_1$ . Then for $i=1,2,\ldots,M$ , $\bigg|w_i(t,s_2)-w_i(t,s_1)|\leq \alpha_{0,1}M(h_M)^M\exp\{2Mh_M/m\}\epsilon$ , where

\begin{align*}\alpha_{0,1}=(2+2Mh_M)\xi+\frac{6}{m}+\frac{4(3+4M h_M \xi)}{m}+2h_M\xi \frac{2(3+4 M h_M \xi)}{m}.\end{align*}

Applying Lemmas 5.13 and 5.14 on (5.13), we get the following Lemma 5.15 as an upper bound of (5.13).

Lemma 5.15. In the event of $E_8^c$ ,

\begin{align*}\frac{1}{N}&\Bigg|\sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}J_{j,i}(t_{n-k},u)-N\sum_{j>i} \int_{h_i}^{h_j}2p_i\frac{w_j(t_{n-k},u)}{l(t_{n-k})}\,\mathrm{d}u\Bigg|\\\leq& \theta+\theta_1+32Mh_M\frac{\epsilon^{2}}{N}+2\alpha_{0,1}M^2(h_M)^{M+1}\exp\left\{\frac{2Mh_M}{m}\right\}\epsilon\\&+2p_i\epsilon \sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}\left(\frac{ \big|W_j(t_{n-k},u)/N- w_j(t_{n-k},u)\big|}{m}\right)+\frac{4p_i}{m}Mh_M\big|L_{\lambda}(t_{n-k})-l(t_{n-k})\big|.\end{align*}

For the third term in last equality of (5.10), recall that $|w_i(t,u)/l(t)|<\xi$ as stated in the conditions of Theorem 3.1, so we have

(5.15) \begin{align}\Bigg|2\int_{0}^{\epsilon} \sum_{h_j\leq u} p_j\frac{w_i(t_{n-k}+u,h_i-u)}{l(t_{n-k}+u)}\,\mathrm{d}u\Bigg|&\leq2 \int_{0}^{\epsilon} \sum_{h_j\leq u} p_j \bigg|\frac{w_i(t_{n-k}+u,h_i-u)}{l(t_{n-k}+u)}\bigg|\,\mathrm{d}u\nonumber\\&\leq 2\int_{0}^{\epsilon} \sum_{h_j\leq u} p_j \xi \,\mathrm{d}u\leq 2\epsilon \xi .\end{align}

By applying Lemmas 5.35.3, and 5.3 as well as (5.15)–(5.10), we conclude that there exists a constant $\alpha_1$ independent of $\lambda,\epsilon$ such that

(5.16) \begin{align}\bigg|&\frac{W_i(t_n,h_i-k\epsilon)}{N}-w_i(t_n,h_i-k\epsilon)\bigg|\nonumber\\\leq& 3\theta_1+3\theta +16 T M \epsilon+2p_r N \epsilon^2 \gamma +2\epsilon \xi+ 32Mh_M\frac{\epsilon^{2}}{N}\nonumber\\&+2\alpha_{0,1}M^2(h_M)^{M+1}\exp\left\{\frac{2Mh_M}{m}\right\}\epsilon +6\epsilon\sum_{j=1}^{n-1} \frac{g(t_{j})}{m}\nonumber\\&+2 \frac{p_i}{N}\frac{\big|f(t_{n-k})-F_\lambda(t_{n-k})\big|+ \big|l(t_{n-k})-L_\lambda(t_{n-k})\big|}{m}\nonumber\\&+\frac{4p_i}{m}Mh_M\big|L_{\lambda}(t_{n-k})-l(t_{n-k})\big|+2p_i\epsilon \sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}\left(\frac{ \big|W_j(t_{n-k},u)/N- w_j(t_{n-k},u)\big|}{m}\right)\nonumber\\\leq&3\theta_1+3\theta+\alpha_1\epsilon+6\epsilon\sum_{j=1}^{n-1} \frac{g(t_{j})}{m}+2 \frac{1}{N}\frac{\big|f(t_{n-k})-F_\lambda(t_{n-k})\big|+ \big|l(t_{n-k})-L_\lambda(t_{n-k})\big|}{m}\nonumber\\&+\frac{4}{m}Mh_M\big|L_{\lambda}(t_{n-k})-l(t_{n-k})\big|+2p_i\epsilon \sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}\left(\frac{ \big|W_j(t_{n-k},u)/N- w_j(t_{n-k},u)\big|}{m}\right)\nonumber\\=\!:\,&y(n,k)+2p_i\epsilon \sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}\left(\frac{ \big|W_j(t_{n-k},u)/N- w_j(t_{n-k},u)\big|}{m}\right) ,\end{align}

where

(5.17) \begin{align}y(n,k)\,:\!&=\,3\theta_1+3\theta+\alpha_1\epsilon+6\epsilon\sum_{j=1}^{n-1} \frac{g(t_{j})}{m} +\frac{4}{m}Mh_M\big|L_{\lambda}(t_{n-k})-l(t_{n-k})\big|\nonumber\\&\quad +2 \frac{\big|f(t_{n-k})-F_\lambda(t_{n-k})\big|+ \big|l(t_{n-k})-L_\lambda(t_{n-k})\big|}{mN}.\end{align}

Equation (5.16) will be bounded above using induction as follows:

Base case: $i=M$ . In such a case there is nothing in the term $\sum_{j>i}$ , and hence

\begin{align*} \bigg|\frac{W_i(t_n,h_i-k\epsilon)}{N}-w_i(t_n,h_i-k\epsilon)\bigg|\leq y(n,k).\end{align*}

Inductive case: $i<M$ . First, consider $i=M-1$ and use the base case to get

\begin{align*}\bigg|&\frac{W_i(t_n,h_i-k\epsilon)}{N}-w_i(t_n,h_i-k\epsilon)\bigg|\\&\leq y(n,k)+2p_i\epsilon \sum_{u=h_i}^{h_M-\epsilon}\left(\frac{ \big|W_M(t_{n-k},u)/N- w_M(t_{n-k},u)\big|}{m}\right)\\&\leq \left(2\frac{h_M}{m}+1\right)y(n,k).\end{align*}

Iterating on $i\leq M$ , we have that there exists an $\alpha_2$ independent of $\lambda,\epsilon$ such that for all i,

(5.18) \begin{align} \bigg|\frac{W_i(t_n,h_i-k\epsilon)}{N}-w_i(t_n,h_i-k\epsilon)\bigg|\leq \alpha_2 y(n,k). \end{align}

Now we extend the inequality to $t\in [t_n,\ t_{n+1})$ and $s\in((k-1)\epsilon,k\epsilon]$ :

(5.19) \begin{align}\bigg|&\frac{W_i(t,s)}{N}-w_i(t,s) \bigg|\nonumber\leq \bigg|\frac{W_2(t_n,k\epsilon)}{N}-w_2(t_n,k\epsilon) \bigg|+ |w_2(t,s)-w_2(t_n,k\epsilon)|\nonumber\\[6pt]\leq& \bigg|\frac{W_2(t_n,k\epsilon)}{N}-w_2(t_n,k\epsilon) \bigg|+|w_2(t,k\epsilon-(t-t_n))-w_2(t_n,k\epsilon)|\nonumber\\[6pt]&+|w_2(t,k\epsilon-(t-t_n))-w_2(t,s)|\nonumber\\[6pt]\leq&\bigg|\frac{W_2(t_n,k\epsilon)}{N}-w_2(t_n,k\epsilon) \bigg|+\int_{0}^{t-t_n}\Bigg|-2\sum_{h_j\leq u}p_j\frac{w_i(t_n+u,k\epsilon-u)}{l(t_n+u)}\Bigg|\,\mathrm{d}u\nonumber\\[6pt]&+\alpha_{0,1}M(h_M)^M\exp\left\{\frac{2Mh_M}{m}\right\}\epsilon \nonumber\\[6pt]\leq& \alpha_2 y(n,k)+2\xi\epsilon+\alpha_{0,1}M(h_M)^M\exp\left\{\frac{2Mh_M}{m}\right\}\epsilon,\end{align}

where $\alpha_{0,1}$ is defined in Lemma 5.14.

Recall (5.17). Then there exist $\alpha_3$ independent of $\lambda, \epsilon$ such that in the event $(E_6\cup E_7\cup E_8)^c$ ,

(5.20) \begin{align}&\bigg|\frac{W_i(t,s)}{N}-w_i(t,s) \bigg|\nonumber\\&\leq \alpha_3\left(\theta_1+\theta+\epsilon+\epsilon\sum_{j=1}^{n-1} \frac{g(t_{j})}{m}+ \big|f(t_{n-k})-F_\lambda(t_{n-k})\big| +\big|L_{\lambda}(t_{n-k})-l(t_{n-k})\big|\right)\!.\end{align}

5.4. Applying Gronwall’s inequality

To summarize, we have (5.5), (5.9), and (5.20) hold on the event $(\bigcup_{i=1}^8 E_i)^c$ . Recall that

\begin{align*}g(t)=\sup_{0 < s\leq t}\left\{\bigg|\frac{F(s)}{\lambda}-f(s)\bigg|+\bigg|\frac{L(s)}{\lambda}-l(s)\bigg|+\sum_{i=1}^M \sup_{u\leq h_i} \bigg|\frac{W_i(s,u)}{N}-w_i(s,u) \bigg| \right\}\!.\end{align*}

Since g is non-decreasing, (5.5), (5.9), and (5.20) imply

\begin{align*}g(t_n)\leq& \sup_{t\leq t_n} \big|F_\lambda(t)-f(t)\big|+\sup_{t\leq t_n} \big|L_\lambda(t)-l(t)\big|+M\sup_{t\leq t_n}\sup_{u\leq h_i} \bigg|\frac{W_i(t,u)}{N}-w_i(s,u)\bigg|\\\leq& (2M+2M^2+\alpha_3)(\theta_1+\theta)+(16MT+38M^2 T+8 M T (h_M)^2(\xi+\gamma)+\alpha_3)\ \epsilon\\&+(2+13 M h_M+\alpha_3)\epsilon\! \sum_{j=0}^{n-1} \frac{ g(t_{j+1})}{m}+\alpha_3(|f(t_{n-1})-F_\lambda(t_{n-1})|+|L_{\lambda}(t_{n-1})-l(t_{n-1})|)\\\leq& \big[2M+2M^2+\alpha_3(1+2M+2M^2)\big](\theta_1+\theta)\\&+\big[16MT+38M^2 T+8 M T h_M^2(\xi+\gamma)+\alpha_3(1+54M^2 T+8 M T h_M^2(\xi+\gamma))\big]\ \epsilon\\&+\left[2+13 M h_M+\alpha_3\big(3+13 M h_M\big)\right]\epsilon \sum_{j=0}^{n-1} \frac{ g(t_{j})}{m}+(2+13 M h_M+\alpha_3)\ \epsilon \frac{ g(t_n)}{m}.\end{align*}

Hence, there exists a constant $\alpha$ independent of $\lambda,\epsilon$ such that

\begin{align*}&g(t_n)\leq \frac{1}{2}\alpha\Bigg(\theta_1+\theta+\epsilon+\epsilon \sum_{j=0}^{n-1} g(t_{j})\Bigg)+\big(2+13 M h_M+\alpha_3\big)\epsilon \frac{ g(t_n)}{m}.\end{align*}

If we assume $\epsilon$ small enough such that $\big(2+13 M h_M+\alpha_3\big)\epsilon <\frac{1}{2}$ , then we get $g(t_n)\leq \alpha(\theta_1+\theta+\epsilon+\epsilon \sum_{j=0}^{n-1} g(t_{j}))$ . By Gronwall’s lemma,

\begin{align*}g(t_n) &\leq \alpha \big(\theta_1+\theta+\epsilon\big)\exp{\left\{\sum_{j=0}^{n-1} \alpha \epsilon\right\}}\leq \alpha \big(\theta_1+\theta+\epsilon\big)e^{\alpha T},\\g(T)&\leq \alpha (\theta_1+\theta+\epsilon)e^{\alpha T}.\end{align*}

Given $\delta>0$ , choose $\theta=\theta_1=\delta e^{-\alpha T}/(3\alpha)$ and $\epsilon_0=\theta$ . Then for $\epsilon\leq \epsilon_0$ , we have $\alpha (\theta_1+\theta+\epsilon)e^{\alpha T}\leq \delta$ . Therefore, on the event $\big(\bigcup_{i=1}^8 E_i\big)^c$ , $g(T)\leq \delta$ . Hence

\begin{align*}\mathbb{P}\big(g(T)>\delta\big)&\leq \mathbb{P}\left(\bigcup_{i=1}^8 E_i\right) \leq \sum_{i=1}^8 \mathbb{P}\big(E_i\big)\\\leq& 2M\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\} +2M\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\} +T\epsilon^{-1} \exp\left\{{-}\frac{(1-p_1h_1/2)^2\lambda}{2h_1}\right\}\\& +2M \exp\left\{{-}\frac{\theta^2}{32 T \epsilon}\right\} +M^2T\epsilon^{-1}\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}+ 256\frac{1}{N\theta^2}+\exp{\left\{{-}\frac{\theta^2 N}{8}\right\}}\\& +M T h_M^2 \frac{\epsilon^{-3}}{N}\left(512\frac{h_M^2}{\theta^2}+\exp\left\{{-}\frac{\theta^2 \epsilon^2 N}{8h^2_M}\right\}\right)+ 512 M^2 h_M\frac{1}{N\theta^2}+\exp\left\{{-}\frac{\theta^2N}{8}\right\}\!.\end{align*}

Combining similar terms shows that there exists positive constant $\beta$ such that

\begin{align*}\mathbb{P}\big(g(T)>\delta\big)\leq& \beta\epsilon^{-1}\exp\left\{{-}\beta \frac{\theta^2}{T\epsilon} \right\}+\beta T\epsilon^{-1}e^{-\beta\lambda}\\&+\beta \exp\big\{{-}\beta\theta^2N\big\}+\beta T\frac{1}{\epsilon^3N\theta^2}+\beta T\frac{1}{\epsilon^3N}\exp\big\{{-}\beta\theta^2\epsilon^2N\big\}\!.\end{align*}

Hence as $\lambda,\epsilon^{-1},N,\epsilon^3N\rightarrow \infty$ the probability goes to 0.

6. Proof of Lemmas

Proof of Lemma 5.2. In order to prove the lemma, we divide the calculation of $D_n$ into M parts and derive a probability bound for each part. Recall the definition of $D_n$ , which can be expressed as follows:

\begin{align*}D_n\,:\!=\,\frac{1}{\lambda}\sum_{j=0}^{n-1} \left(-N+\sum_{i=1}^M N_i(t_j-h_i)\right)=\sum_{i=1}^M\left\{\frac{1}{\lambda}\sum_{j=0}^{n-1} (N_i(t_j-h_i)-p_i N)\right\}\!.\end{align*}

Note that $N_i(t_j)$ for $t_j\leq0$ is consider to be the initial condition, and recall that $n_i=h_i/\epsilon$ ; we then split each of the M parts as

(6.1) \begin{align}\sum_{j=0}^{n-1} \frac{(N_i(t_j-h_i)-p_i N)}{\lambda}=\sum_{j=-n_i}^{-1\land (n-n_i-1)} \frac{(N_i(t_j)-p_i N)}{\lambda}+\sum_{j=0}^{n-n_i-1} \frac{(N_i(t_j)-p_i N)}{\lambda},\nonumber\\\quad\end{align}

where the first summation on the right-hand side of (6.1) is regarded as the initial condition. Since $\mathbb{E}\big[N_i(t_j)\big]=p_iN$ for $t_j>0$ and $N_i(t_j)$ is binomial distributed, Lemma 5.1 leads to

(6.2) \begin{align}\mathbb{P}\left(\sup_{n_i< n\leq n_T} \Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-n_i-1} (N_i(t_j)-p_i N)\Bigg|\geq \theta\right)\leq 2\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}\!.\end{align}

Recall the assumption stated in (5.2) which says for a given $\theta_1$ ,

(6.3) \begin{align}\sup_{-n_i\leq k<0}\Bigg|\frac{1}{\lambda}\sum_{j=-n_i}^{k} (N_i(t_j)-p_iN)\Bigg|<\theta_1 \quad\text{for all }i=1,\ldots,M.\end{align}

With (6.1)–(6.3), we get

\begin{align*}\mathbb{P}&\left(\sup_{n_i\leq n\leq n_T} \Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1} \left(N_i(t_j-h_i)-p_i N\right)\Bigg|\geq \theta+\theta_1\right)\leq 2\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}\\&\implies\mathbb{P}\Big(\sup_{0 < n\leq n_T}|D_n|\geq M\theta+M\theta_1\Big)\\&\leq \sum_{i=1}^M \mathbb{P}\left(\sup_{n_i\leq n\leq n_T} \Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1} \left(N_i(t_j-h_i)-p_i N\right)\Bigg|\geq \theta+\theta_1\right)\leq 2M\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}\!,\end{align*}

and the lemma follows.

Proof of Lemma 5.3. The proof follows the same idea as for (6.2). By the assumption of Lemma 5.3, the quantities $N_i(t_j)$ for $i=1,\ldots,M$ and $j=-n_i,\ldots,-1$ in the initial condition follow the same distribution as in (2.1). Hence, applying the same derivation as for (6.2) using Lemma 5.1, we obtain

\begin{align*}\mathbb{P}\left(\sup_{-n_i\leq k<0}\Bigg|\frac{1}{\lambda}\sum_{j=-n_i}^{k} \left(N_i(t_j)-p_iN\right)\Bigg| \geq \theta_1\right) \leq 2\exp\left\{{-}\frac{\theta_1^2}{32h_M\epsilon}\right\} \quad\text{for }i=1,\ldots,M.\end{align*}

Therefore, the initial condition of interest is violated with probability

\begin{align*}\mathbb{P}&\left(\bigcup_{i=1}^{M}\left\{\sup_{-n_i\leq k<0}\Bigg|\frac{1}{\lambda}\sum_{j=-n_i}^{k} (N_i(t_j)-p_iN)\Bigg|\geq\theta_1\right\}\right)\\ &\leq\sum_{i=1}^M\mathbb{P}\left(\sup_{-n_i\leq k<0}\Bigg|\frac{1}{\lambda}\sum_{j=-n_i}^{k} (N_i(t_j)-p_iN)\Bigg| \geq \theta_1\right) \leq 2M\exp\left\{{-}\frac{\theta_1^2}{32h_M\epsilon}\right\}\!,\end{align*}

which approaches zero as $\epsilon\rightarrow 0$ . This implies that when the parameters $\lambda,N,\epsilon$ approach their limit, the desired initial condition holds with probability approaching one.

In (6.1), we decompose the term into a part with the initial condition and a part involving the variables generated after $t=0$ , and both parts follow the same rule when generated. Therefore, it is sufficient to derive the probability bound for the latter part by assuming that the initial condition follows a similar behavior. The same idea can be applied to other assumptions on the initial condition, and hence we will omit the lemmas and proofs on the validity of these assumptions. Since only finitely many such assumptions are required, the probability that any of them is violated also vanishes and hence Proposition 2.1 follows.

Proof of Lemma 5.5. By the construction of the process, a vertex that finishes its POW between the time $t_{n-h_1\epsilon^{-1}}$ and $t_{n-1}$ will still be a tip at time $t_n$ , i.e.

\begin{align*}L\big(t_{n}\big)&\geq\sum_{j=n-h_1\epsilon^{-1}}^{n-1}\sum_{i=1}^{M} N_i(t_j-h_i),\\L_{\lambda}(t_{n})&\geq\frac{1}{\lambda}\sum_{j=n-h_1\epsilon^{-1}}^{n-1} N_1(t_j-h_1)\,=\!:\,\frac{1}{\lambda}\Sigma(n,-)+\frac{1}{\lambda}\Sigma(n,+)\!,\end{align*}

where $L_\lambda(t_n)\,:\!=\,L(t_n)/\lambda$ , and $\Sigma(n,-)$ includes the summands that are considered as initial conditions while $\Sigma(n,+)$ includes the summands that are generated after $t=0$ .

Recall that $N_1(t_k)$ is a binomial random variable that counts the number of arrivals having POW duration as $h_1$ . Hence, with $\mathbb{P}(X_i=1)=p_1$ we can rewrite $\Sigma(n,+)=\sum_{j=1}^{r} X_i$ for some $r \leq c_1\lambda$ (since some $N_1$ are in initial conditions) and the value of r depends on $t_n$ . By the Azuma–Hoeffding inequality,

\begin{align*}\mathbb{P}\Bigg(\frac{1}{\lambda}\sum_{j=1}^{r} \big(X_i-p_1\big)\leq -\theta\Bigg)&=\mathbb{P}\Bigg(\sum_{j=1}^{r} \big(X_i-p_1\big)\leq -\theta\lambda\Bigg)\\&\leq\exp\left\{{-}\frac{\theta^2\lambda^2}{2r}\right\}\leq\exp\left\{{-}\frac{\theta^2\lambda^2}{2h_1\lambda}\right\}=\exp\left\{{-}\frac{\theta^2\lambda}{2h_1}\right\}\!.\end{align*}

Case 1: $r\geq \frac{1}{2}h_1\lambda$ (i.e. $n>\frac{3}{2}h_1\epsilon^{-1}$ )

\begin{align*}\mathbb{P}\big({L_\lambda(t_n)\leq 1}\big)&\leq\mathbb{P}\left({\frac{1}{\lambda}\Sigma(n,+)\leq 1}\right)=\mathbb{P}\Bigg(\frac{1}{\lambda}\sum_{j=1}^r(X_i-p_1)\leq 1-\frac{p_1r}{\lambda}\Bigg)\\&\leq\exp\left\{{-}\frac{(1-p_1r/\lambda)^2\lambda}{2h_1}\right\}\leq\exp\left\{{-}\frac{(1-p_1h_1/2)^2\lambda}{2h_1}\right\}\!,\end{align*}

where, without loss of generality, we assume $h_1p_1/2>1$ since we can change the time unit, and therefore $(1-p_1r/\lambda)^2\geq (1-p_1h_1)^2$ . Then

\begin{align*}\mathbb{P}\Big(\inf_{h_1\epsilon^{-1}/2\leq n\leq n_T}L_\lambda(t_n)\leq 1\Big)&\leq \sum_{n=h_1\epsilon^{-1}/2}^{n_T}\mathbb{P}(L_\lambda(t_n)\leq 1)\leq\sum_{n=h_1\epsilon^{-1}/2}^{n_T}\mathbb{P}(L_\lambda(t_n)\leq 1)\\&\leq T\epsilon^{-1}\exp\left\{{-}\frac{(1-p_1h_1/2)^2\lambda}{2h_1}\right\}\!.\end{align*}

Case 2: $r< \frac{1}{2}h_1\lambda$ (i.e. $n\leq \frac{3}{2}h_1\epsilon^{-1}$ ). Focus on $\Sigma(n,-)$ with the number of $X_i$ larger than $\frac{1}{2}h_1\lambda$ . By the same idea as in Case 1, if we consider the initial condition is randomly generated following the evolution rule, it is reasonable to assume

\begin{align*}\inf_{-h_M\epsilon^{-1}\leq n\leq \frac{3}{2}h_1\epsilon^{-1}}\sum_{j=n-2h_1\epsilon^{-1}}^{0} N_1(t_j)>1,\end{align*}

which implies

\begin{align*}\mathbb{P}\Big(\inf_{-h_M\epsilon^{-1}\leq n\leq n_T}L_{\lambda}(t_n)\leq 1\Big)\leq T\epsilon^{-1} \exp\left\{{-}\frac{(1-p_1h_1/2)^2\lambda}{2h_1}\right\}\!.\end{align*}

Proof of Lemma 5.6. Recall (2.5). We rewrite it as

\begin{align*}F_i(t_n) = \sum_{j=1}^{F(t_n)-F_1(t_n)-\cdots-F_{i-1}(t_n)} X_j\,\colon \,\, X_j\in \{0,1\}, \,\, \mathbb{P}(X_j=0)=\left(1 - \frac{1}{L(t_n)}\right)^{2 N_i(t_n)}.\end{align*}

Since $\sum_{r=1}^{i}F_r(t_n)$ are the number of free tips that are selected at time $t_n$ by any arrival with Type no larger than i, then we also have

\begin{align*}\sum_{r=1}^{i}F_r(t_n) = \sum_{j=1}^{F(t_n)} X_j\,\colon \,\, X_j\in \{0,1\}, \quad \mathbb{P}(X_j=0)=\left(1 - \frac{1}{L(t_n)}\right)^{2 \sum_{r=1}^{i}N_r(t_n)},\end{align*}

and hence

(6.4) \begin{align}&\mathbb{E}\left[\sum_{r=1}^{i}F_r(t_j)\middle|\sigma_{j}\right]=\mathbb{E}\left[F(t_j)\left[1-\left(1-\frac{1}{L(t_j)}\right)^{2\sum_{r=1}^i N_r(t_j)}\right]\middle|\sigma_{j}\right]\nonumber\\&=\mathbb{E}\left[F(t_j)\left[\frac{2\sum_{r=1}^i N_r(t_j)}{L(t_j)}-\sum_{d=2}^{2\sum_{r=1}^i N_r(t_j)}\binom{2\sum_{r=1}^i N_r(t_j)}{d}\left(-\frac{1}{L(t_j)}\right)^d\right]\middle|\sigma_{j}\right]\nonumber\\&=\frac{2\sum_{r=1}^i p_r N F(t_j)}{L(t_j)}-F(t_j)\mathbb{E}\left[\sum_{d=2}^{2\sum_{r=1}^i N_r(t_j)}\binom{2\sum_{r=1}^i N_r(t_j)}{d}\left(-\frac{1}{L(t_j)}\right)^d\middle|\sigma_{j}\right]. \end{align}

Let Y be a random variable such that $0 \leq Y\leq N$ , and denote $F_\lambda(t)=F(t)/\lambda$ and $L_\lambda=L(\lambda)$ . Then

(6.5) \begin{align}&F(t_j)\Bigg|\mathbb{E}\left[\sum_{d=2}^{2Y}\binom{2Y}{d}\left(-\frac{1}{L(t_j)}\right)^d\Big|\sigma_{j}\right]\Bigg|\leq F(t_j)\Bigg|\mathbb{E}\left[\sum_{d=2}^{2Y}\binom{2Y}{d}\left(\frac{1}{L(t_j)}\right)^d\Big|\sigma_{j}\right]\Bigg|\nonumber\\&\leq F(t_j)\Bigg|\mathbb{E}\left[\sum_{d=2}^{2N}\binom{2N}{d}\left(\frac{1}{L(t_j)}\right)^d\Big|\sigma_{j}\right]\Bigg|\leq F(t_j)\Bigg|\sum_{d=2}^{2N}\binom{2N}{d}\frac{1}{L(t_j)^d}\Bigg|\nonumber\\&\leq F(t_j)\Bigg|\sum_{d=2}^{2N}(2N)^d\frac{1}{L(t_j)^d}\Bigg|\leq \frac{F(t_j)}{L(t_j)}\Bigg|\sum_{d=2}^{2N}\frac{(2N)^d}{\lambda^{d-1}}\frac{1}{L_\lambda(t_j)^{d-1}}\Bigg|,\end{align}

where

(6.6) \begin{align}\frac{F(t_j)}{L(t_j)}\Bigg|\sum_{d=2}^{2N}\frac{(2N)^d}{\lambda^{d-1}}\frac{1}{L_\lambda(t_j)^{d-1}}\Bigg|\leq\frac{4\epsilon^2\lambda}{1-2\epsilon}\leq 8\epsilon^2\lambda\end{align}

because we have restricted our analysis in the event $E_3^C$ (recall that it says $L_\lambda>1$ ) and we make an assumption that $\epsilon<\frac{1}{4}$ for simplicity. Note that we can replace F by $W_{i}\leq 2N$ in (6.6) to get the bound $16\epsilon^3\lambda$ , which we will use later. Equations (6.4)–(6.6) then lead to

(6.7) \begin{align}\mathbb{E}\big[F_i(t_j)|\sigma_j\big]=\mathbb{E}\left[\sum_{r=1}^{i}F_r(t_j)-\sum_{r=1}^{i-1}F_r(t_j)\Big|\sigma_{j}\right]=\frac{2p_iNF(t_j)}{L(t_j)}+error,\end{align}

where $|error|<16\epsilon^2\lambda$ . In contrast,

(6.8) \begin{align}\bigg|\frac{f(s)}{l(s)}-\frac{F(s)}{L(s)}\bigg|&=\bigg|\frac{f(s)-F_\lambda(s)}{l(s)}+\frac{F_\lambda(s)}{L_\lambda(s)}\cdot\frac{L_\lambda(s)-l(s)}{l(s)}\bigg|\nonumber\\\leq&\frac{|f(s)-F_\lambda(s)|+|L_\lambda(s)-l(s)|}{l(s)}\leq\frac{|f(s)-F_\lambda(s)|+|L_\lambda(s)-l(s)|}{m}\leq \frac{g(s)}{m},\end{align}

where $0 < m\leq l(s)$ is the lower bound assumption for l(s). For future reference, if $s\leq 0$ , the above equation is 0 by construction of the initial condition of the fluid limit.

Putting (5.4), (6.7), and (6.8) together we get

\begin{align*}|H_{i,j+1}|&=\bigg|\lambda\int_{t_j}^{t_{j+1}} 2p_i\frac{f(s)}{l(s)}\,\mathrm{d}s-\frac{2p_iNF(t_j)}{L(t_j)}-error\bigg|\\&=\bigg|\lambda\int_{t_j}^{t_{j+1}}\left( 2p_1\frac{f(s)}{l(s)}-2p_1\frac{F(t_j)}{L(t_j)}\right)\,\mathrm{d}s-error\bigg|\leq \frac{2p_iNg(t_{j+1})}{m}+16\epsilon^2\lambda\\&\implies\frac{1}{\lambda}\sum_{j=0}^{n-1}|H_{i,j+1}|\leq 16\epsilon T+\sum_{j=0}^{n-1} \frac{2p_i\epsilon g(t_{j+1})}{m}.\end{align*}

Proof of Lemma 5.7. First, consider the following with n and i fixed

\begin{align*}\frac{1}{\lambda}&\sum_{j=0}^{n-1}\left(\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\mathbb{E}\big[J_{i,r}\big(t_{j-h(u)},u\big)|\sigma_{j-h(u)}\big]-\sum_{u=h_r+\epsilon}^{h_i-\epsilon}J_{i,r}\big(t_{j-h(u)},u\big)\right)\\=\!:\,&\frac{1}{\lambda}\sum_{j=0}^{n-1}\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\left(\Delta(j,u)\right)\!,\end{align*}

where $\Delta(j,u)\,:\!=\,\mathbb{E}[J_{i,r}(t_{j-h(u)},u)|\sigma_{j-h(u)}]-J_{i,r}(t_{j-h(u)},u)$ . Despite the potential terms that are the initial conditions which are bounded by $\theta_1$ , the rest are the sum of random variables with expected value of 0. Each term $\Delta(j,u)$ is measurable by $\sigma_{j-h(u)+1}$ but not by $\sigma_{j-h(u)}$ ; so we can group the terms $\Delta(j,u)$ by the first time they are measurable. Then each grouped sum is a $Y_s$ in Lemma 5.1 since the total number of jumps possible at each time is 2N, then these are just a sum of $Y_s$ with s at most $n_T\,:\!=\,T/\epsilon$ . Then applying Lemma 5.1,

\begin{align*}&\mathbb{P}\left(\bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1}\left(\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\mathbb{E}\big[J_{i,r}\big(t_{j-h(u)},u\big)|\sigma_{j-h(u)}\big]-\sum_{u=h_r+\epsilon}^{h_i-\epsilon}J_{i,r}\big(t_{j-h(u)},u\big)\right)\bigg|\geq \theta_1+\theta\right)\\&\leq 2\exp\left\{{-}\frac{\theta^2}{32\epsilon}\right\}\!.\end{align*}

Then using the union bound we get

\begin{align*}&\mathbb{P}\left(\sup_{0 < n\leq n_T}\bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1}\left(\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\mathbb{E}\big[J_{i,r}(t_{j-h(u)},u)|\sigma_{j-h(u)}\big]-\sum_{u=h_r+\epsilon}^{h_i-\epsilon}J_{i,r}(t_{j-h(u)},u)\right)\bigg|\geq\theta_1+\theta\right)\\&\leq 2T\epsilon^{-1}\exp\left\{{-}\frac{\theta^2}{32T\epsilon}\right\}\!.\end{align*}

Proof of Lemma 5.8. Following the same idea as used in the proof of Lemma 5.6,

\begin{align*}\mathbb{E}\bigg[\sum_{r\leq k}J_{i,r}(t_{j-h(u)},u)|\sigma_{j-h(u)}\bigg]&= W_{i}(t_{j-h(u)},u)\mathbb{E}\left[1-\left(1-\frac{1}{L\big(t_{j-h(u)}\big)}\right)^{2\sum_{r\leq k}N_r}|\sigma_{j-h(u)}\right]\\&=\frac{2\sum_{r\leq k} p_r N W_{i}(t_{j-h(u)},u)}{L(t_{j-h(u)})}+error,\end{align*}

for $h_k\leq u$ , where $|error|\leq 16\epsilon^3\lambda$ . Hence

\begin{align*}\mathbb{E}&\big[J_{i,r}(t_{j-h(u)},u)|\sigma_{j-h(u)}\big]\\=&\,\mathbb{E}\bigg[\sum_{r\leq k}J_{i,r}(t_{j-h(u)},u)|\sigma_{j-h(u)}\bigg]-\mathbb{E}\bigg[\sum_{r\leq k-1}J_{i,r}(t_{j-h(u)},u)|\sigma_{j-h(u)}\bigg]\\=\,&\frac{2 p_r N W_{i}(t_{j-h(u)},u)}{L(t_{j-h(u)})}+error^*,\end{align*}

where $|error^*|\leq 32\epsilon^3\lambda$ . Therefore,

(6.9) \begin{align}\Bigg|&\frac{1}{\lambda}\sum_{j=0}^{n-1}\Bigg[2\lambda\int_{t_j}^{t_{j+1}}\int_{h_r}^{h_i}p_r\frac{w_i(t-h(u)\epsilon,u)}{l(t-h(u)\epsilon)}dudt-\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\mathbb{E}\big[J_{i,r}(t_{j-h(u)},u)|\sigma_{j-h(u)}\big]\Bigg]\Bigg|\nonumber\\\leq&\Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1}\Bigg[2\lambda\int_{t_j}^{t_{j+1}}\int_{h_r}^{h_i}p_r\frac{w_i(t-h(u)\epsilon,u)}{l(t-h(u)\epsilon)}dudt-\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\frac{2p_r N W_{i}(t_{j-h(u)},u)}{L(t_{j-h(u)})}-error^*\Bigg]\Bigg|\nonumber\\\leq&\Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1}\Bigg[2\lambda p_r\int_{t_j}^{t_{j+1}}\left(\int_{h_r}^{h_i}\frac{w_i(t-h(v)\epsilon,v)}{l(t-h(v)\epsilon)}dv-\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\frac{ W_{i}(t-h(u)\epsilon,u)}{L(t-h(u)\epsilon)}\right) dt\Bigg]\Bigg|\nonumber\\&+32\epsilon T(h_i-h_r),\end{align}

where $h(u)\epsilon$ is either $h_i$ or u. As usual we will combine the integration and summation in the last equation. First, note that

(6.10) \begin{align}\bigg|\frac{w_i(t,s)}{l(t)}-\frac{W_i(t,s)/N}{L(t)/\lambda}\bigg|&\leq\bigg|\frac{w_i(t,s)-W_i(t,s)/N}{l(t)}+\frac{W_i(t,s)/N}{L_\lambda(t)}\cdot\frac{L_{\lambda}(t)-l(t)}{l(t)}\bigg|\nonumber\\&\leq\frac{g(t)}{m}+2\frac{g(t)}{m}\leq3\frac{g(t)}{m}.\end{align}

In the case where $h(u)\epsilon=h_i$ , the following can be obtained using (6.10):

\begin{align*}&\Bigg|\int_{h_r}^{h_i}\frac{w_i(t-h_i,v)}{l(t-h_i)}\,\mathrm{d}v-\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\frac{ W_{i}(t-h_i,u)}{L(t-h_i)}\Bigg|\nonumber\\&\leq \Bigg|\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\left(\int_{u-\epsilon}^u \frac{w_i(t-h_i,v)}{l(t-h_i)}\,\mathrm{d}v-\frac{ W_{i}(t-h_i,u)}{L(t-h_i)}\right)\Bigg|+\Bigg| \int_{h_i-\epsilon}^{h_i}\frac{w_i(t-h_i,v)}{l(t-h_i)}\,\mathrm{d}v\Bigg|\nonumber\end{align*}
(6.11) \begin{align}&\leq \Bigg|\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\int_{u-\epsilon}^u\left(\frac{w_i(t-h_i,v)}{l(t-h_i)}-\frac{ W_{i}(t-h_i,v)/N}{L(t-h_i)/\lambda}\right)\,\mathrm{d}v\Bigg|+\xi\epsilon\nonumber\\&\leq\xi\epsilon+\sum_{u=h_r+\epsilon}^{h_i-\epsilon}3\epsilon \frac{g(t)}{m}\leq\xi\epsilon+3(h_i-h_r)\frac{g(t)}{m},\end{align}

where $\xi$ is the upper bound for $\frac{w_i(t,u)}{l(t)}$ for all i, t, s. Similarly, in the case where $h(u)\epsilon=u$ ,

(6.12) \begin{align}\Bigg|&\int_{h_r}^{h_i}\frac{w_i(t-v,v)}{l(t-v)}\,\mathrm{d}v-\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\frac{ W_{i}(t-u,u)}{L(t-u)}\Bigg|\nonumber\\[4pt]\leq&\Bigg|\int_{h_r}^{h_i}\frac{w_i(t-v,v)}{l(t-v)}\,\mathrm{d}v-\epsilon\sum_{u=h_r+\epsilon}^{h_i}\frac{w_i(t-u,u)}{l(t-u)}\Bigg|\nonumber\\[4pt]&+\Bigg|\epsilon\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\frac{w_i(t-u,u)}{l(t-u)}-\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\frac{ W_{i}(t-u,u)}{L(t-u)}\Bigg|+\bigg|\epsilon\frac{w_i(t-h_i,h_i)}{l(t-h_i)}\bigg|\nonumber\\[4pt]\leq&\gamma(h_i-h_r)^2\epsilon+3\epsilon \sum_{u=h_r+\epsilon}^{h_i-\epsilon}\frac{g(t)}{m}+\xi\epsilon=\gamma(h_i-h_r)^2\epsilon+3(h_i-h_r)\frac{g(t)}{m}+\xi\epsilon,\end{align}

where $\gamma$ is specified by (3.3). Then by putting (6.9), (6.11), and (6.12) together, we get

\begin{align*}&\Bigg|\frac{1}{\lambda}\sum_{j=0}^{n-1}\Bigg[2\lambda\int_{t_j}^{t_{j+1}}\int_{h_r}^{h_i}p_r\frac{w_i(t-h(u)\epsilon,u)}{l(t-h(u)\epsilon)}\,\mathrm{d}u\,\mathrm{d}t-\sum_{u=h_r+\epsilon}^{h_i-\epsilon}\mathbb{E}\big[J_{i,r}\big(t_{j-h(u)},u\big)|\sigma_{j-h(u)}\big]\Bigg]\Bigg|\\&\leq32\epsilon T+2Tp_r\epsilon\big(\xi+(h_i-h_r)^2\gamma\big)+6p_r(h_i-h_r)\epsilon \sum_{j=0}^{n-1}\frac{g(t_{j+1})}{m}.\end{align*}

Proof of Lemma 5.9. Consider N balls randomly dropped into L boxes with equal probability, where each ball has a probability $1-p$ of disappearing. Let X denote the number of boxes that satisfy the following requirements: (a) they boxes are the first W boxes, and (b) each of the boxes has at least one ball. Let Y denote the number of balls that drop into the first W boxes, with W and L constant. Then with $p=p_r$ , W corresponds to $W_i(t,u)$ , L corresponds to L(t), and X corresponds to $J_{i,r}$ , the number of pending tips that change to Type r. We then have $X\leq Y$ , $\mathbb{E}[Y]=pNW/L$ , and $\mathbb{E}[X]=pNW/L+error$ , where $|error|\leq 32\epsilon^3\lambda$ is deduced from the conditional expectation of $J_{i,r}$ (see proof of Lemma 5.8; a similar result of bounded error can be generalized to any bounded W and will be used later). Then

(6.13) \begin{align}-32\epsilon^3\lambda&\leq \mathbb{E}[X-Y],\end{align}
(6.14) \begin{align}\kern-28pt X-Y&\leq 0,\end{align}

which, with the Markov inequality, we get that for any $\delta>0$ , $\mathbb{P}(X-Y\leq -\delta)< 32\epsilon^3\lambda\delta^{-1}$ . This combined with Hoeffding’s inequality for Y leads to

\begin{align*}\mathbb{P}\left(|X-\mathbb{E}[Y]|>\frac{N\theta}{2}\right)&\leq \mathbb{P}\left(X-Y< -\frac{N\theta}{4}\right)+\mathbb{P}\bigg(|Y-\mathbb{E}[Y]|>\frac{N\theta}{4}\bigg)\\&\leq 32\epsilon^3\lambda\frac{16}{N^2\theta^2}+\exp\left\{{-}\frac{\theta^2N^2}{8N}\right\}\leq 512\epsilon^2\frac{1}{N\theta^2}+\exp\left\{{-}\frac{\theta^2N}{8}\right\}\!.\end{align*}

Assume that $\epsilon$ is small enough that $\mathbb{E}[Y-X]/N< \theta/2$ , then we get

(6.15) \begin{align}\mathbb{P}(|X-\mathbb{E}[X]|>N\theta)&\leq \mathbb{P}\left(|X-\mathbb{E}[Y]|>\frac{N\theta}{2}\right)\leq 512\epsilon^2\frac{1}{N\theta^2}+\exp\left\{{-}\frac{\theta^2N}{8}\right\}\!.\end{align}

Recall that

\begin{align*} G_{i,n,k,j}\,:\!=\,\sum_{h_r\leq h_i-j\epsilon}-J_{i,r}\big(t_{n-k+j},h_i-j\epsilon\big)+\sum_{h_r\leq h_i-j\epsilon}\mathbb{E}\big[J_{i,r}\big(t_{n-k+j},h_i-j\epsilon\big)|\sigma_{n-k+j}\big],\end{align*}

where the first summation counts the number of pending tips that changes Type, and the second summation counts the corresponding expected value. Let $p=\sum_{h_r\leq h_i-j\epsilon}p_r$ ; this implies that each arriving vertex has probability p of having POW duration less or equal than $h_i-j\epsilon$ , and X represents the number of pending tips with RLT $h_i-j\epsilon$ that changes Type. Therefore, we can use (6.15) to get

\begin{align*}&\mathbb{P}\left(\bigg|\frac{1}{N}G_{i,n,k,j}\mathbf{1}_{\{n-k+j\geq 0\}}\bigg|\geq \epsilon\theta\right)\leq 512\frac{1}{N\theta^2}+\exp\left\{{-}\frac{\theta^2 \epsilon^2 N}{8}\right\}\!,\end{align*}

where $\mathbf{1}_{\{\}}$ is the indicator function. Using the fact that $k\leq h_M\epsilon^{-1}$ , we get

\begin{align*}&\mathbb{P}\Bigg(\Bigg|\frac{1}{N}\sum_{j=1}^{k-1} G_{i,n,k,j}\mathbf{1}_{\{n-k+j\geq 0\}}\Bigg|\geq\theta\Bigg)\leq\mathbb{P}\Bigg(\Bigg|\frac{1}{N}\sum_{j=1}^{k-1} G_{i,n,k,j}\mathbf{1}_{\{n-k+j\geq 0\}}\Bigg|\geq\frac{k-1}{h_M\epsilon^{-1}}\theta\Bigg)\\&\leq \sum_{j=1}^{k-1} \mathbb{P}\left(\bigg|\frac{1}{N} G_{i,n,k,j}\mathbf{1}_{\{n-k+j\geq 0\}}\bigg|\geq \frac{\epsilon\theta}{h_M}\right)\leq \frac{h_M}{\epsilon}\left(512\frac{h_M^2}{N\theta^2}+\exp\left\{{-}\frac{\theta^2 \epsilon^2 N}{8h^2_M}\right\}\right)\!.\end{align*}

Therefore, with the assumption that (5.11) is true, we have

\begin{align*}\mathbb{P}&\Bigg(\Bigg|\frac{1}{N}\sum_{j=1}^{k-1} G_{i,n,k,j}\Bigg|\geq\theta+\theta_1\Bigg)\leq h_M\epsilon^{-1}\left(512\frac{h_M^2}{N\theta^2}+\exp\left\{{-}\frac{\theta^2 \epsilon^2 N}{8h^2_M}\right\}\right)\\&\implies\mathbb{P}\left(\sup_{i\leq M}\sup_{0 < n\leq n_T}\sup_{1\leq k\leq n_i}\Bigg| \frac{1}{N} \sum_{j=1}^{k-1} G_{i,n,k,j}\Bigg|\geq \theta_1+\theta\right)\\&\leq M T h_M^2 \frac{\epsilon^{-3}}{N}\left(512\frac{h_M^2}{\theta^2}+\exp\left\{{-}\frac{\theta^2 \epsilon^2 N}{8h^2_M}\right\}\right)\!.\end{align*}

Proof of Lemma 5.10. We will use statements derived in the proof for Lemma 5.8. We have

\begin{align*}\mathbb{E}\big[J_{i,r}\big(t_{n-k+j},h_i-j\epsilon\big)|\sigma_{n-k+j}\big]&=\frac{2p_r N W_{i}(t_{n-k+j},h_i-j\epsilon)}{L(t_{n-k+j})}+error,\end{align*}

where $|error|\leq 16\epsilon^3\lambda$ . Hence

\begin{align*}&\Bigg|\frac{1}{N}\sum_{j=1}^{k-1} \left(\sum_{h_r\leq h_i-j\epsilon}-\mathbb{E}\big[J_{i,r}\big(t_{n-k+j},h_i-j\epsilon\big)|\sigma_{n-k+j}\big]\right.\nonumber\\&\quad\left.+N \int_{0}^{\epsilon}2 p_r\frac{w_i\big(t_{n-k+j}+u,h_i-j\epsilon-u\big)}{l(t_{n-k+j}+u)}\,\mathrm{d}u\right)\Bigg|\end{align*}
\begin{align*}\leq& \Bigg| \frac{1}{N} \sum_{j=1}^{k-1} M\left\{{-}\frac{2p_r N W_{i}\big(t_{n-k+j},h_i-j\epsilon\big)}{L(t_{n-k+j})}+error+2 p_r N\epsilon \frac{w_i\big(t_{n-k+j},h_i-j\epsilon\big)}{l(t_{n-k+j})}\right\}\Bigg|\\&+\Bigg|\frac{1}{N}\sum_{j=1}^{k-1} M\left\{{-}2 p_r N\epsilon \frac{w_i\big(t_{n-k+j},h_i-j\epsilon\big)}{l(t_{n-k+j})}+N \int_{0}^{\epsilon}2 p_r\frac{w_i\big(t_{n-k+j}+u,h_i-j\epsilon-u\big)}{l(t_{n-k+j}+u)}\,\mathrm{d}u\right\}\Bigg|\\\leq& 16 T M \epsilon+\sum_{j=1}^{k-1}6\epsilon \frac{g(t_{n-k+j})}{m}+2p_r N \epsilon^2 \gamma\leq 16 T M \epsilon+2p_r N \epsilon^2 \gamma +6\epsilon\sum_{j=1}^{n-1} \frac{g(t_{j})}{m}.\end{align*}

Proof of Lemma 5.13. The term $\sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}J_{j,i}(t_{n-k},u)$ describes the number of pending tips which have Type higher than i and RLT longer than $h_i$ that are selected by Type i arrivals. Hence, by repeating the proof of Lemma 5.9 with suitable choice of W as shown in the following, we get

\begin{align*}&W=\sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}W_{j}(t_{n-k},u)\leq M h_M\epsilon^{-1} (2N)=2Mh_M\lambda\\&\implies\mathbb{E}[X]=\frac{NW}{L}+error \quad\quad\quad\text{where }|error|\leq32Mh_M\epsilon^2\lambda.\end{align*}

Hence with the assumption of (5.14),

\begin{align*}&\mathbb{P}\left(\Bigg|\sum_{j>i}\sum_{u=h_i}^{h_j-\epsilon} \left( J_{j,i}(t_{n-k},u)-\mathbb{E}\big[J_{j,i}(t_{n-k},u)|\sigma_{n-k}\big] \right)\Bigg|>N(\theta+\theta_1)\right)\\&\leq512 M h_M\epsilon\frac{1}{N\theta^2}+\exp\left\{{-}\frac{\theta^2N}{8}\right\}\!.\end{align*}

Therefore,

\begin{align*}&\mathbb{P}\left(\sup_{i}\sup_{n-k\leq n_T}\frac{1}{N}\Bigg|\sum_{j>i}\sum_{u=h_i}^{h_j-\epsilon} \left( J_{j,i}(t_{n-k},u)-\mathbb{E}\big[J_{j,i}(t_{n-k},u)|\sigma_{n-k}\big] \right)\Bigg|>\theta+\theta_1\right)\\&\leq 512 M^2 h_M \frac{1}{N\theta^2}+\exp\left\{{-}\frac{\theta^2N}{8}\right\}\!.\end{align*}

Proof of Lemma 5.14. It is easier for proof to replace $s_j$ by $h_i-s_j$ , that is, we focus on

\begin{align*}w_i(t,h_i-s_1)&= w_i(t-s_1,h_i)+\int_{0}^{s_1} \sum_{h_j\leq h_i-u} -2p_j\frac{w_i(t-s_1+u,h_i-u)}{l(t-s_1+u)}\,\mathrm{d}u\end{align*}

and

\begin{align*}&w_i(t,h_i-s_2)= w_i(t-s_2,h_i)+\int_{0}^{s_2} \sum_{h_j\leq h_i-u} -2p_j\frac{w_i(t-s_2+u,h_i-u)}{l(t-s_2+u)}\,\mathrm{d}u\\&=w_i(t-s_2,h_i)+\int_{s_1-s_2}^{s_1} \sum_{h_j\leq h_i-(u-s_1+s_2)}-2p_j\frac{w_i(t-s_1+u,h_i-(u-s_1+s_2))}{l(t-s_1+u)} \,\mathrm{d}u.\end{align*}

Hence we have the following:

(6.16) \begin{align}&|w_i(t,h_i-s_2)-w_i(t,h_i-s_1)|\nonumber\\\leq& |w_i(t-s_2,h_i)- w_i(t-s_1,h_i)|+\bigg|\int_{s_1-s_2}^0-2\sum_{h_j} p_j\frac{w_i(t-s_1+u,h_i-(u-s_1+s_2))}{l(t-s_1+u)}\,\mathrm{d}u\bigg|\nonumber\\&+\int_{0}^{s_1}2\sum_{h_j\leq h_i-u}p_j \bigg|\frac{w_i(t-s_1+u,h_i-u)-w_i(t-s_1+u,h_i-(u-s_1+s_2))}{l(t-s_1+u)}\bigg|\,\mathrm{d}u\nonumber\\&+2\int_{0}^{s_1}\bigg|\sum_{h_j\leq h_i-(u-s_1+s_2)} \frac{p_j w_i(t-s_1+u,h_i-(u-s_1+s_2))}{l(t-s_1+u)} \nonumber\\&\quad\quad\quad\quad\quad\quad\quad-\sum_{h_j\leq h_i-u} \frac{p_j w_i(t-s_1+u,h_i-(u-s_1+s_2))}{l(t-s_1+u)}\bigg|\,\mathrm{d}u .\end{align}

For the first term on the right-hand side of (6.16),

(6.17) \begin{align}|&w_i(t-s_2,h_i)- w_i(t-s_1,h_i)|\nonumber\\&\leq\bigg|2p_i\frac{f(t-s_2)}{l(t-s_2)}-2p_i\frac{f(t-s_1)}{l(t-s_1)}\bigg|+\sum_{j>i}\int_{ h_i}^{h_j}2p_i\bigg|\frac{w_j(t-s_1,u)}{l(t-s_1)}-\frac{w_j(t-s_2,u)}{l(t-s_2)}\bigg| du\nonumber\\&\leq2\left[\frac{f(t-s_2)-f(t-s_1)}{l(t-s_2)}+\frac{f(t-s_1)}{l(t-s_1)}\frac{l(t-s_1)-l(t-s_2)}{l(t-s_2)}\right]+\cdots\nonumber\\&\leq \frac{6\epsilon}{m}+\frac{2(3+4M h_M \xi)\epsilon}{m}+\sum_{j>i}\int_{ h_i}^{h_j}2p_i\bigg|\frac{w_j(t-s_1,u)}{l(t-s_1)}-\frac{w_j(t-s_2,u)}{l(t-s_2)}\bigg| \,\mathrm{d}u\nonumber\\&\leq\frac{6\epsilon}{m}+\frac{2(3+4M h_M \xi)\epsilon}{m}+2h_M\xi \frac{2(3+4M h_M\xi)\epsilon}{m}\nonumber\\&\quad+\sum_{j>i}\int_{ h_i}^{h_j}2p_i|w_j(t-s_1,u)-w_j(t-s_2,u)|\,\mathrm{d}u,\end{align}

where we use the facts $f(t)<l(t)$ , $s_2-s_1<\epsilon$ , $l(t)>m$ , and $\xi>w_j$ . For the second term on the right-hand side of (6.16),

(6.18) \begin{align}\Bigg|\int_{s_1-s_2}^0-2 \sum_{h_j}p_j \frac{w_i(t-s_1+u,h_i-(u-s_1+s_2))}{l(t-s_1+u)}\,\mathrm{d}u\Bigg|\leq\int_{0}^{\epsilon}2\xi\leq 2\xi\epsilon,\end{align}

where $\xi$ is an upper bound for $\frac{w_2(s+u,u)}{l(s+u)}$ for all s, u. For the last term on the right-hand side of (6.16), because $|s_1-s_2|<\epsilon$ , we have

\begin{align*}&2\int_{0}^{s_1}\Bigg|\sum_{h_j\leq h_i-(u-s_1+s_2)} p_j\frac{w_2(t-s_1+u,h_i-(u-s_1+s_2))}{l(t-s_1+u)}\nonumber\\&\quad\quad\quad\quad\quad-\sum_{h_j\leq h_i-u} p_j\frac{w_2(t-s_1+u,h_i-(u-s_1+s_2))}{l(t-s_1+u)}\Bigg|\,\mathrm{d}u\end{align*}
(6.19) \begin{align}&\leq 2\int_{0}^{s_1}\Bigg|\sum_{h_i-u < h_j\leq h_i-(u-s_1+s_2)} p_j\frac{w_2(t-s_1+u,h_i-(u-s_1+s_2))}{l(t-s_1+u)}\Bigg|\,\mathrm{d}u\nonumber\\&\leq 2\sum_{j < i}\int_{h_i-h_j+s_2-s_i}^{h_i-h_j}\bigg|p_j\frac{w_2(t-s_1+u,h_i-(u-s_1+s_2))}{l(t-s_1+u)}\bigg|\,\mathrm{d}u\leq 2 \xi \epsilon .\end{align}

All that left is the third term in (6.16). Define

(6.20) \begin{align}y_i(u)=|w_i(t-s_1+u,h_i-u)-w_i(t-s_1+u,h_i-(u-s_1+s_2))|.\end{align}

Then we have $\big|w_i(t,h_i-s_2)-w_i(t,h_i-s_1)\big|=y_i(s_1)$ , and substituting (6.17)–(6.20) into (6.16) we get

\begin{align*}y_i(s_1)\leq& 4\xi\epsilon+\frac{6\epsilon}{m}+\frac{2(3+4M h_M\xi)\epsilon}{m}+2h_M\xi \frac{2(3+4M h_M \xi)\epsilon}{m}\\&+\sum_{j>i}\int_{ h_i}^{h_j}2p_i|w_j(t-s_1,u)-w_j(t-s_2,u)|\,\mathrm{d}u+\frac{2}{m}\int_{0}^{s_1}y_i(u)\,\mathrm{d}u.\end{align*}

And since

\begin{align*}\big|w_j&(t-s_1,u)-w_j(t-s_2,u)\big|\\\leq&\big|w_j(t-s_1,u)-w_j(t-s_1,u-s_2+s_1)\big|\nonumber\\&+2\bigg|\int_{0}^{s_1-s_2}\sum_{h_r\leq u-s_2+s_1+s}p_j\frac{w_j(t-s_1+s,u-s_2+s_1+s)}{l(t-s_2+s)}\,\mathrm{d}u\bigg|\nonumber\\\leq&|w_j(t-s_1,u)-w_j(t-s_1,u-s_2+s_1)|+2\xi\epsilon,\end{align*}

we get

(6.21) \begin{align}&y_i(s_1)\leq 4\xi\epsilon+\frac{6\epsilon}{m}+\frac{2(3+4M h_M \xi)\epsilon}{m}+2h_M\xi \frac{2(3+4M h_M \xi)\epsilon}{m}\nonumber\\&+2Mh_M\xi\epsilon+\sum_{j>i}\int_{ h_i}^{h_j}2p_i\big|w_j(t-s_1,u)-w_j(t-s_1,u-s_2+s_1)\big|\,\mathrm{d}u+\frac{2}{m}\int_{0}^{s_1}y_i(u)\,\mathrm{d}u.\end{align}

We now use induction on i on (6.21).

Case 1: $i=M$ . There is no term in the summation $\sum_{j>i}$ included in (6.21). Thus,

\begin{align*}y_i\big(s_1\big)\leq&\ (2+2Mh_M)\xi\epsilon+\frac{6\epsilon}{m}+\frac{2(3+4 M h_M \xi)\epsilon}{m}+\frac{2(3+4M h_M \xi)\epsilon}{m}\\&+2h_M\xi \frac{2(3+4 M h_M \xi)\epsilon}{m}+\frac{2}{m}\int_{0}^{s_1}y(u)\,\mathrm{d}u\\=\!:\,&\ \alpha_{1,0}\epsilon+\frac{2}{m}\int_{0}^{s_1}y_i(u)\,\mathrm{d}u.\end{align*}

Applying Gronwall’s inequality we get

\begin{align*}y_M(s_i)=\big|w_M(t,h_M-s_2)-w_M(t,h_M-s_1)\big|\leq\alpha_{1,0}\epsilon \exp\left\{\int_{0}^{s_1}\frac{2}{m}du\right\}\leq\alpha_{1,0}\epsilon \exp\left\{\frac{2h_M}{m}\right\}\!.\end{align*}

Case 2: $i=M-1$ . Using the result for Case 1, we get

\begin{align*}y_i(s_1)&\leq \alpha_{1,0}\epsilon+\sum_{j>i}\int_{ h_i}^{h_j}2p_i\big|w_j(t-s_1,u)-w_j(t-s_2,u)\big|\,\mathrm{d}u+\frac{2}{m}\int_{0}^{s_1}y_i(u)\,\mathrm{d}u\\&\leq \alpha_{1,0}\epsilon+\int_{ h_i}^{h_M}2\big|w_M(t-s_1,u)-w_M(t-s_2,u)\big|\,\mathrm{d}u +\frac{2}{m}\int_{0}^{s_1}y_i(u)\,\mathrm{d}u\\&\leq \alpha_{1,0}\epsilon+2h_M\alpha_{1,0} \exp\left\{\frac{2h_M}{m}\right\}\epsilon+\frac{2}{m}\int_{0}^{s_1}y_i(u)\,\mathrm{d}u\\&\leq(3h_M)\alpha_{1,0} \exp\left\{\frac{2h_M}{m}\right\}\epsilon+\frac{2}{m}\int_{0}^{s_1}y_i(u)\,\mathrm{d}u.\end{align*}

Again applying Gronwall’s inequality, we get

\begin{align*}\big|w_{M-1}(t,s_2)-w_{M-1}(t,s_1)\big|\leq(2h_M+1)\alpha_{1,0} \exp\left\{\frac{4h_M}{m}\right\}\epsilon.\end{align*}

Repeating and iterating on i, for all $i\leq M$ we get

\begin{align*}\big|w_i(t,s_2)-w_i(t,s_1)\big|\leq \alpha_{0,1}M(h_M)^M\exp\left\{\frac{2Mh_M}{m}\right\}\epsilon.\end{align*}

Proof of Lemma 5.15. Using Lemmas 5.13 and 5.14, we get

\begin{align*}& \frac{1}{N}\Bigg|\sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}J_{j,i}(t_{n-k},u)-N\sum_{j>i} \int_{h_i}^{h_j}2p_i\frac{w_j(t_{n-k},u)}{l(t_{n-k})}\,\mathrm{d}u\Bigg|\nonumber\\&\leq \theta+\theta_1+ \frac{1}{N}\Bigg|\sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}\mathbb{E}\big[J_{j,i}(t_{n-k},u)|\sigma_{n-k}\big]-N\sum_{j>i} \int_{h_i}^{h_j}2p_i\frac{w_j(t_{n-k},u)}{l(t_{n-k})}\,\mathrm{d}u\Bigg|\nonumber\end{align*}
\begin{align*}&\leq \theta+\theta_1+ \frac{1}{N}\Bigg|\sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}\frac{2p_i N W_j(t_{n-k},u)}{L(t_{n-k})}-N\sum_{j>i} \int_{h_i}^{h_j}2p_i \frac{w_j(t_{n-k},u)}{l(t_{n-k})}\,\mathrm{d}u\Bigg|\nonumber\\&\quad+ \frac{1}{N}\big(M\cdot h_M\epsilon^{-1}\cdot 32\epsilon^3\big)\nonumber\\&\leq \theta+\theta_1+\frac{1}{N}\big(M\cdot h_M\epsilon^{-1}\cdot 32\epsilon^3\big)+2\alpha_{0,1}M^2(h_M)^{M+1}\exp\left\{\frac{2Mh_M}{m}\right\}\epsilon\nonumber\\& \quad +\frac{1}{N}\Bigg|2p_i N \sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}\left(\frac{ W_j(t_{n-k},u)}{L(t_{n-k})}-\epsilon\frac{ w_j(t_{n-k},u)}{l(t_{n-k})}\right)\Bigg|\\&\leq\theta+\theta_1+32Mh_M\frac{\epsilon^{2}}{N}+2\alpha_{0,1}M^2(h_M)^{M+1}\exp\left\{\frac{2Mh_M}{m}\right\}\epsilon\end{align*}
\begin{align*}&\quad+2p_i \epsilon \sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}\left(\frac{ |W_j(t_{n-k},u)/N- w_j(t_{n-k},u)|+2|L_{\lambda}(t_{n-k})-l(t_{n-k})|}{m}\right)\\&\leq\theta+\theta_1+32Mh_M\frac{\epsilon^{2}}{N}+2\alpha_{0,1}M^2(h_M)^{M+1}\exp\left\{\frac{2Mh_M}{m}\right\}\epsilon\\&\quad+2p_i\epsilon \sum_{j>i} \sum_{u=h_i}^{h_j-\epsilon}\left(\frac{ \big|W_j(t_{n-k},u)/N- w_j(t_{n-k},u)\big|}{m}\right)+\frac{4p_i}{m}Mh_M\big|L_{\lambda}(t_{n-k})-l(t_{n-k})\big|.\end{align*}

Funding information

There are no funding bodies to thank relating to the creation of this article.

Competing interests

There were no competing interests to declare which arose during the preparation or publication process of this article.

References

Ameen, T., Sankagiri, S. and Hajek, B. (2022). Blockchain security when messages are lost. In Soares, J. M., Song, D. and Vukolić, M. (eds.), Proceedings of the 2022 ACM Workshop on Developments in Consensus, Association for Computing Machinery, New York, USA, pp. 114.10.1145/3560829.3563557CrossRefGoogle Scholar
Back, A. (2002). Hashcash - a denial of service counter-measure. http://www.hashcash.org/papers/hashcash.pdf.Google Scholar
Dembo, A. ET AL. (2020). Everything is a race and Nakamoto always wins. In Ligatti, J. and Ou, X. (eds.), Proceedings of the 2020 ACM SIGSAC Conference on Computer and Communications Security, Association for Computing Machinery, New York, USA, pp. 859878.10.1145/3372297.3417290CrossRefGoogle Scholar
Dey, P. S. and Gopalan, A. (2025). A framework for blockchain architecture design. arXiv:2202.05080 Google Scholar
Driver, R. D. (2012). Ordinary and Delay Differential Equations. Springer Verlag, New York.Google Scholar
Feng, J., King, C. and Duffy, K. R. (2025). Almost sure one-endedness of a random graph model of distributed ledgers. Stochastic Syst. 15(3), 252272.10.1287/stsy.2023.0045CrossRefGoogle Scholar
Ferraro, P., King, C. and Shorten, R. (2018). Distributed ledger technology for smart cities, the sharing economy, and social compliance. IEEE Access 6, 6272862746.10.1109/ACCESS.2018.2876766CrossRefGoogle Scholar
Ferraro, P., King, C. and Shorten, R. (2020). On the stability of unverified transactions in a DAG-based distributed ledger. IEEE Trans. Autom. Control 9, 37723783.10.1109/TAC.2019.2950873CrossRefGoogle Scholar
Frolkova, M. and Mandjes, M. (2019). A bitcoin-inspired infinite-server model with a random fluid limit. Stochastic Models 35(1), 132.10.1080/15326349.2018.1559739CrossRefGoogle Scholar
Gazi, P., Ren, L. and Russell, A. (2022). Practical settlement bounds for proof-of-work blockchains. In Yin, H. and Stavrou, A. (eds.), Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, Association for Computing Machinery, New York, USA, pp. 12171230.10.1145/3548606.3559368CrossRefGoogle Scholar
Göbel, J., Keeler, H. P., Krzesinski, A. E. and Taylor, P. G. (2016). Fraud risk assessment within blockchain transactions. Perform. Eval. 104, 2341.10.1016/j.peva.2016.07.001CrossRefGoogle Scholar
Goffard, P. O. (2019). Fraud risk assessment within blockchain transactions. Adv. Appl. Probab. 51(2), 443467.10.1017/apr.2019.18CrossRefGoogle Scholar
Gopalan, A., Sankararaman, A., Walid, A. and Vishwanath, S. (2020). Stability and scalability of blockchain systems. Proc. ACM Meas. Anal. Comput. Syst. 4(2), 135.10.1145/3392153CrossRefGoogle Scholar
King, C. (2021). The fluid limit of a random graph model for a shared ledger. Adv. Appl. Probab. 53(1), 81106.10.1017/apr.2020.48CrossRefGoogle Scholar
Lewenberg, Y., Sompolinsky, Y. and Zohar, A. (2015). Inclusive block chain protocols. In Böhme, R. and Okamoto, T. (eds.), International Conference on Financial Cryptography and Data Security, Springer Berlin, Heidelberg, pp. 528547.10.1007/978-3-662-47854-7_33CrossRefGoogle Scholar
Müller, S., Amigo, I., Reiffers-Masson, A. and Ruano-Rincón, S. (2023). Stability of local tip pool sizes. arXiv:2302.01625 Google Scholar
Müller, S. ET AL. (2022). Tangle 2.0 leaderless Nakamoto consensus on the heaviest DAG. IEEE Access 10, 105807105842.10.1109/ACCESS.2022.3211422CrossRefGoogle Scholar
Nakamoto, S. (2008). Bitcoin: a peer-to-peer electronic cash system. https://bitcoin.org/bitcoin.pdf.Google Scholar
Pahlajani, S., Kshirsagar, A. and Pachghare, V. (2019). Survey on private blockchain consensus algorithms. In 1st International Conference on Innovations in Information and Communication Technology, IEEE, Piscataway, USA, pp. 16.10.1109/ICIICT1.2019.8741353CrossRefGoogle Scholar
Penzkofer, A., Saa, O. and Dziubałtowska, D. (2022). Impact of delay classes on the data structure in IOTA. In Garcia-Alfaro, J., Muñoz-Tapia, J. L., Navarro-Arribas, G., and Soriano, M. (eds.), Data Privacy Management, Cryptocurrencies and Blockchain Technology, Springer Cham, pp. 289300.10.1007/978-3-030-93944-1_19CrossRefGoogle Scholar
Popov, S. (2018). The tangle.Google Scholar
Sankagiri, S., Gandlur, S. and Hajek, B. (2023). The longest-chain protocol under random delays. Stochastic Syst. 13(4), 457478.10.1287/stsy.2022.0031CrossRefGoogle Scholar
Sankagiri, S., Wang, X., Kannan, S. and Viswanath, P. (2021). Stability of local tip pool sizes. arXiv:2010.13711 Google Scholar
Sompolinsky, Y., Lewenberg, Y. and Zohar, A. (2016). Spectre: a fast and scalable cryptocurrency protocol. Cryptology ePrint Archive. https://eprint.iacr.org/2016/1159.Google Scholar
Sompolinsky, Y., Wyborski, S. and Zohar, A. (2021). PHANTOM GHOSTDAG: a scalable generalization of Nakamoto consensus: September 2, 2021. In Baldimtsi, F. and Roughgarden, T. (eds.), Proceedings of the 3rd ACM Conference on Advances in Financial Technologies, Association for Computing Machinery, New York, USA, pp. 5771.10.1145/3479722.3480990CrossRefGoogle Scholar
Williams, D. (1991). Probability with Martingales. Cambridge University Press, Cambridge. 10.1017/CBO9780511813658CrossRefGoogle Scholar
Zhang, S. and Lee, J. H. (2020). Analysis of the main consensus protocols of blockchain. ICT Express 6(2), 9397.10.1016/j.icte.2019.08.001CrossRefGoogle Scholar
Zheng, Z. ET AL. (2017). An overview of blockchain technology: architecture, consensus, and future trends. In Gannon, D., Prasad, S. K. and Goul, M. (eds.), IEEE International Congress on Big Data, IEEE, Piscataway, USA, pp. 557564.10.1109/BigDataCongress.2017.85CrossRefGoogle Scholar
Figure 0

Figure 1. Example of a DAG. A solid directed edge implies that the associated POW has been completed and the data has been accepted into the ledger. For example, vertex 3 has selected 1 as its parent and finished its POW. A dashed vertex with outgoing dashed edge implies the POW has started but not yet been finished. For example, vertex 5 has selected 2 and 3 as its parents but its POW has not yet been finished. We generally refer to the solid vertices that have no solid edge pointing toward them as tips, which means the vertices have been accepted to the distributed ledger but have not yet been attached by any other vertices. For example, vertices 2 and 3 are tips because there is no solid edge pointing toward them, while vertices 0 and 1 are not tips because they have been attached by other solid vertices. Also, vertices 4 and 5 are not considered to be tips because they have not been accepted to the system, as indicated by the fact that their POW have not been finished

Figure 1

Figure 2. A demonstration of a DAG modeling IOTA at some time t. Each time there are $N=2$ arriving vertices, for example, $C_1,C_2$ arrived at the same time while $D_1,D_2$ arrived simultaneously $\epsilon$ time after $C_1,C_2$ arrived. A solid vertex with outgoing solid edge(s) implies that its corresponding POW has been completed and the vertex has been accepted into the system. For example, vertex $C_1$ has selected $A_1$ as its parent and finished its POW, hence it has been accepted into the system. A dashed vertex with outgoing dashed edge(s) implies the POW has not yet been finished. For example, vertex $D_2$ has selected $B_1$ and $A_2$ as its parents but the POW corresponding to $D_2$ has not yet been finished. The dashed vertices are the ones that have not been accepted into the system because their POWs are still in process. A dashed vertex will be included in the system once its corresponding POW is finished, then the dashed vertex and the dashed edge(s) originating from it will become solid

Figure 2

Figure 3. Example of the random variables $F_i(t_k)$. In this example, $N=3$ which means there are 3 arrivals at each time. At some time $t_k$ the vertices $D_1,D_2,D_3$ arrive and they have POW durations $h_1,h_2,h_3$, respectively. The graph on the left represents the graph at $t_k$ when $D_1,D_2,D_3$ arrive while the right graph represents the graph at $t_{k+1}$. At time $t_k$, the vertices $A_1,B_1,C_1,C_2$ are free tips since their corresponding POWs have been finished but they have not yet been selected as parents. The free tip $A_1$ is selected as a parent by a Type 1 arrival $D_1$. The free tip $B_1$ is selected by a Type 1 arrival $D_1$ and a Type 2 arrival $D_2$. The free tip $C_2$ is selected by a Type 2 arrival $D_2$ and a Type 3 arrival $D_3$. Then $F_1(t_k)=2$ because both $A_1,B_1$ are selected by a Type 1 arrival. Because $B_1$ is already included in $F_1(t_k)$, we have $F_2(t_k)=1$ although both $B_1,C_2$ are selected by a Type 2 arrival. Finally, $F_3(t_k)=0$ because $C_2$ is the only free tip that is selected by a Type 3 arrival but $C_2$ has been counted in $F_2(t_k)$. The selection will be reflected at $t_{k+1}$ and hence $A_1,B_1,C_2$ become pending tips at $t_{k+1}$

Figure 3

Figure 4. Simulations with $\lambda=400, N=20, \epsilon=0.05$ where the vertical axis displays values for $L(t)/\lambda,F(t)/\lambda$ and the horizontal axis represents time $t_n$. Here, M is assumed to be 2, corresponding to Proposition 3.1 The panel (a) corresponds to the simulation with $p_1=0.8$ and $p_2=0.2$ while the panel (b) corresponds to the simulation with $p_1=0.3$ and $p_2=0.7$. Both simulations assume $h_1=3$ and $h_2=5$. Multiple simulations are performed with the same parameters and, we can see that the scaled random processes behave almost deterministically with minor deviation.