To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
We study an open discrete-time queueing network. We assume data is generated at nodes of the network as a discrete-time Bernoulli process. All nodes in the network maintain a queue and relay data, which is to be finally collected by a designated sink. We prove that the resulting multidimensional Markov chain representing the queue size of nodes has two behavior regimes depending on the value of the rate of data generation. In particular, we show that there is a nontrivial critical value of the data rate below which the chain is ergodic and converges to a stationary distribution and above which it is non-ergodic, i.e., the queues at the nodes grow in an unbounded manner. We show that the rate of convergence to stationarity is geometric in the subcritical regime.
In this study, we analysed the relationship between meteorological factors and the number of patients with coronavirus disease 2019 (COVID-19). The study period was from 12 April 2020 to 13 October 2020, and daily meteorological data and the daily number of patients with COVID-19 in each state of the United States were collected. Based on the number of COVID-19 patients in each state of the United States, we selected four states (California, Florida, New York, Texas) for analysis. One-way analysis of variance ( ANOVA), scatter plot analysis, correlation analysis and distributed lag nonlinear model (DLNM) analysis were used to analyse the relationship between meteorological factors and the number of patients with COVID-19. We found that the significant influencing factors of the number of COVID-19 cases differed among the four states. Specifically, the number of COVID-19 confirmed cases in California and New York was negatively correlated with AWMD (P < 0.01) and positively correlated with AQI, PM2.5 and TAVG (P < 0.01) but not significantly correlated with other factors. Florida was significantly correlated with TAVG (positive) (P < 0.01) but not significantly correlated with other factors. The number of COVID-19 cases in Texas was only significantly negatively associated with AWND (P < 0.01). The influence of temperature and PM2.5 on the spread of COVID-19 is not obvious. This study shows that when the wind speed was 2 m/s, it had a significant positive correlation with COVID-19 cases. The impact of meteorological factors on COVID-19 may be very complicated. It is necessary to further explore the relationship between meteorological factors and COVID-19. By exploring the influence of meteorological factors on COVID-19, we can help people to establish a more accurate early warning system.
This paper outlines frameworks to use for reserving validation and gives the reader an overview of current techniques being employed. In the experience of the authors, many companies lack an embedded reserve validation framework and reserve validation can appear piecemeal and unstructured. The paper outlines a case study demonstrating how successful machine learning techniques will become and then goes on to discuss the implications of machine learning to the future of reserving departments, processes, data and validation techniques. Reserving validation can take many forms, from simple checks to full independent reviews to add value to the reserving process, enhance governance and increase confidence in and reliability in results. This paper discusses covers common weaknesses and their solutions and suggestions of a framework in which to apply validation tools. The impacts of the COVID-19 pandemic on reserving validation is also covered as are early warning indicators and the topic of IFRS 17 from the standpoint of reserving validation. The paper looks at the future for reserving validation and discusses the data challenges that need overcoming on the path to embedded reserving process validation.
While many of the prevalent stochastic mortality models provide adequate short- to medium-term forecasts, only few provide biologically plausible descriptions of mortality on longer horizons and are sufficiently stable to be of practical use in smaller populations. Among the very first to address the issue of modelling adult mortality in small populations was the SAINT model, which has been used for pricing, reserving and longevity risk management by the Danish Labour Market Supplementary Pension Fund (ATP) for more than a decade. The lessons learned have broadened our understanding of desirable model properties from the practitioner’s point of view and have led to a revision of model components to address accuracy, stability, flexibility, explainability and credibility concerns. This paper serves as an update to the original version published 10 years ago and presents the SAINT model with its modifications and the rationale behind them. The main improvement is the generalization of frailty models from deterministic structures to a flexible class of stochastic models. We show by example how the SAINT framework is used for modelling mortality at ATP and make comparisons to the Lee-Carter model.
This paper considers a generalized panel data transformation model with fixed effects where the structural function is assumed to be additive. In our model, no parametric assumptions are imposed on the transformation function, the structural function, or the distribution of the idiosyncratic error term. The model is widely applicable and includes many popular panel data models as special cases. We propose a kernel-based nonparametric estimator for the structural function. The estimator has a closed-form solution and is easy to implement. We study the asymptotic properties of our estimator and show that it is asymptotically normally distributed. The Monte Carlo simulations demonstrate that our new estimator performs well in finite samples.
AI has had many summers and winters. Proponents have overpromised, and there has been hype and disappointment. In recent years, however, we have watched with awe, surprise, and hope at the successes: Better than human capabilities of image-recognition; winning at Go; useful chatbots that seem to understand your needs; recommendation algorithms harvesting the wisdom of crowds. And with this success comes the spectre of danger. Machine behaviours that embed the worst of human prejudice and biases; techniques trying to exploit human weaknesses to skew elections or prompt self-harming behaviours. Are we seeing a perfect storm of social media, sensor technologies, new algorithms and edge computing? With this backdrop: is AI coming of age?
By gathering 680 publicly available Sequence Read Archives from isolates of Mycobacterium tuberculosis complex (MTBC) including 190 belonging to the lineage 2 Beijing, and using an in-house bioinformatical pipeline, the TB-Annotator, that analyses more than 50 000 characters, we describe herein a new L2 sublineage from 20 isolates found in the Tochigi province, (Japan), that we designate as asia ancestral 5 (AAnc5). These isolates harbour a number of specific criteria (42 SNPs) and their intra-cluster pairwise distance suggests historical and not epidemiological transmission. These isolates harbour a mutation in rpoC, and do not fulfil, any of the modern Beijing lineage criteria, nor any of the other ancestral Beijing lineages described so far. Asia ancestral 5 isolates do not possess mutT2 58 and ogt 12 characteristics of modern Beijing, but possess ancestral Beijing SNPs characteristics. By looking into the literature, we found a reference isolate ID381, described in Kobe and Osaka belonging to the ‘G3’ group, sharing 36 out of the 42 specific SNPs found in AAnc5. We also assessed the intermediate position of the asia ancestral 4 (AAnc4) sublineage recently described in Thailand and propose an improved classification of the L2 that now includes AAnc4 and AAnc5. By increasing the recruitment into TB-Annotator to around 3000 genomes (including 642 belonging to L2), we confirmed our results and discovered additional historical ancestral L2 branches that remain to be investigated in more detail. We also present, in addition, some anthropological and historical data from Chinese and Japan history of tuberculosis, as well as from Korea, that could support our results on L2 evolution. This study shows that the reconstruction of the early history of tuberculosis in Asia is likely to reveal complex patterns since its emergence.
Data driven analysis is proven to create a competitive advantage to business. Governments and nonprofit organizations also turn to Big Data to harness its benefits and use it for social good. Among different types of data sources, location data collected from mobile networks is especially valuable for its representativeness, real-time observation, and versatility. There is a distinction between mobile positioning data (MPD) generated by the exchanges between mobile devices and the core network; versus over-the-top or system-level location data collecting individual GPS location. MPD is composed of all mobile network events regardless of the mobile phone brand, operating system, app usage, frequency bands or mobile generation; it is uniform and ubiquitous. Getting the best out of MPD relies on the knowledge of how to create an advanced algorithm for homogeneously processing this massive, complex data into insightful indicators. Anonymized and aggregated MPD enables the testing of multiple combinations with other data sources, fully abiding by GDPR, to arrive at innovative solutions. These unique insights can help tackle societal challenges (the state of mobile data for social good June 2017 GSMA, UN Global pulse). It can help to establish accurate statistics about population movements, density, location, social patterns, finances, and ambient environmental conditions. This article demonstrates how MPD has been used to help combat Covid-19 in Europe, the Middle East, and Africa. Furthermore, depending on the future direction, MPD and data analysis can serve powering economic development as well as working toward the Sustainable Development Goals, whilst respecting data privacy.
We introduce the generalized join the shortest queue model with retrials and two infinite capacity orbit queues. Three independent Poisson streams of jobs, namely a smart, and two dedicated streams, flow into a single-server system, which can hold at most one job. Arriving jobs that find the server occupied are routed to the orbits as follows: Blocked jobs from the smart stream are routed to the shortest orbit queue, and in case of a tie, they choose an orbit randomly. Blocked jobs from the dedicated streams are routed directly to their orbits. Orbiting jobs retry to connect with the server at different retrial rates, i.e., heterogeneous orbit queues. Applications of such a system are found in the modeling of wireless cooperative networks. We are interested in the asymptotic behavior of the stationary distribution of this model, provided that the system is stable. More precisely, we investigate the conditions under which the tail asymptotic of the minimum orbit queue length is exactly geometric. Moreover, we apply a heuristic asymptotic approach to obtain approximations of the steady-state joint orbit queue-length distribution. Useful numerical examples are presented and shown that the results obtained through the asymptotic analysis and the heuristic approach are agreed.
We study weighted ensemble, an interacting particle method for sampling distributions of Markov chains that has been used in computational chemistry since the 1990s. Many important applications of weighted ensemble require the computation of long time averages. We establish the consistency of weighted ensemble in this setting by proving an ergodic theorem for time averages. As part of the proof, we derive explicit variance formulas that could be useful for optimizing the method.
Latouche and Nguyen (2015b) constructed a sequence of stochastic fluid processes and showed that it converges weakly to a Markov-modulated Brownian motion (MMBM). Here, we construct a different sequence of stochastic fluid processes and show that it converges strongly to an MMBM. To the best of our knowledge, this is the first result on strong convergence to a Markov-modulated Brownian motion. Besides implying weak convergence, such a strong approximation constitutes a powerful tool for developing deep results for sophisticated models. Additionally, we prove that the rate of this almost sure convergence is $o(n^{-1/2} \log n)$. When reduced to the special case of standard Brownian motion, our convergence rate is an improvement over that obtained by a different approximation in Gorostiza and Griego (1980), which is $o(n^{-1/2}(\log n)^{5/2})$.
This study quantified CO2 emissions from tropical peat swamp soils in Brunei Darussalam. At each site, soil was collected from areas of intact and degraded peat and CO2 flux, and total organic content were measured ex situ. Soil organic content (~20–99%) was not significantly different between intact and degraded forest samples. CO2 flux was higher for intact forest samples than degraded forest samples (~1.0 vs. ~0.6 μmol CO2 m−2 s−1, respectively) but did not differ among forest locations. From our laboratory experiments, we estimated a potential emissions of ~10–20 t CO2 ha−1 y−1 which is in the lower range of values reported for other tropical peat swamps. However, our results are likely affected by unmeasured variation in root respiration and the lability of resident carbon. Overall, these findings provide experimental evidence to support that clearance of tropical peat swamp forests can increase CO2 emissions due to faster rates of decomposition.
For a non-negative separable random field Z(t), $t\in \mathbb{R}^d$, satisfying some mild assumptions, we show that $ H_Z^\delta =\lim_{{T} \to \infty} ({1}/{T^d}) \mathbb{E}\{{\sup_{ t\in [0,T]^d \cap \delta \mathbb{Z}^d } Z(t) }\} <\infty$ for $\delta \ge 0$, where $0 \mathbb{Z}^d\,:\!=\,\mathbb{R}^d$, and prove that $H_Z^0$ can be approximated by $H_Z^\delta$ if $\delta$ tends to 0. These results extend the classical findings for Pickands constants $H_{Z}^\delta$, defined for $Z(t)= \exp( \sqrt{ 2} B_\alpha (t)- \lvert {t} \rvert^{2\alpha})$, $t\in \mathbb{R}$, with $B_\alpha$ a standard fractional Brownian motion with Hurst parameter $\alpha \in (0,1]$. The continuity of $H_{Z}^\delta$ at $\delta=0$ is additionally shown for two particular extensions of Pickands constants.
A method for the construction of Stein-type covariance identities for a nonnegative continuous random variable is proposed, using a probabilistic analogue of the mean value theorem and weighted distributions. A generalized covariance identity is obtained, and applications focused on actuarial and financial science are provided. Some characterization results for gamma and Pareto distributions are also given. Identities for risk measures which have a covariance representation are obtained; these measures are connected with the Bonferroni, De Vergottini, Gini, and Wang indices. Moreover, under some assumptions, an identity for the variance of a function of a random variable is derived, and its performance is discussed with respect to well-known upper and lower bounds.
Discover a comprehensive set of tools and techniques for analyzing the impact of uncertainty on large-scale engineered systems. Providing accessible yet rigorous coverage, it showcases the theory through detailed case studies drawn from electric power application problems, including the impact of integration of renewable-based power generation in bulk power systems, the impact of corrupted measurement and communication devices in microgrid closed-loop controls, and the impact of components failures on the reliability of power supply systems. The case studies also serve as a guide on how to tackle similar problems that appear in other engineering application domains, including automotive and aerospace engineering. This is essential reading for academic researchers and graduate students in power systems engineering, and dynamic systems and control engineering.
Multisystem inflammatory syndrome in adults (MIS-A) is a hyperinflammatory illness related to severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) infection. The characteristics of patients with this syndrome and the frequency with which it occurs among patients hospitalised after SARS-CoV-2 infection are unclear. Using the Centers for Disease Control and Prevention case definition for MIS-A, we created ICD-10-CM code and laboratory criteria to identify potential MIS-A patients in the Premier Healthcare Database Special COVID-19 Release, a database containing patient-level information on hospital discharges across the United States. Modified MIS-A criteria were applied to hospitalisations with discharge from March to December 2020. The proportion of hospitalisations meeting electronic health record criteria for MIS-A and descriptive statistics for patients in the potential MIS-A cohort were calculated. Of 34 515 SARS-CoV-2-related hospitalisations with complete clinical and laboratory data, 53 met modified criteria for MIS-A (0.15%). The median age was 62 years (IQR 52–74). Most patients met the severe cardiac illness criterion through either myocarditis (66.0%) or new-onset heart failure (35.8%). A total of 79.2% of patients required ICU admission, while 43.4% of patients in the cohort died. MIS-A appears to be a rare but severe outcome of SARS-CoV-2 infection. Additional studies are needed to investigate how this syndrome differs from severe coronavirus disease 2019 (COVID-19) in adults.
We consider a variant of a classical coverage process, the Boolean model in $\mathbb{R}^d$. Previous efforts have focused on convergence of the unoccupied region containing the origin to a well-studied limit C. We study the intersection of sets centered at points of a Poisson point process confined to the unit ball. Using a coupling between the intersection model and the original Boolean model, we show that the scaled intersection converges weakly to the same limit C. Along the way, we present some tools for studying statistics of a class of intersection models.
In our data-driven society, personal data affecting individuals as data subjects are increasingly being collected and processed by sizeable and international companies. While data protection laws and privacy technologies attempt to limit the impact of data breaches and privacy scandals, they rely on individuals having a detailed understanding of the available recourse, resulting in the responsibilization of data protection. Existing data stewardship frameworks incorporate data-protection-by-design principles but may not include data subjects in the data protection process itself, relying on supplementary legal doctrines to better enforce data protection regulations. To better protect individual autonomy over personal data, this paper proposes a data protection-focused data commons to encourage co-creation of data protection solutions and rebalance power between data subjects and data controllers. We conduct interviews with commons experts to identify the institutional barriers to creating a commons and challenges of incorporating data protection principles into a commons, encouraging participatory innovation in data governance. We find that working with stakeholders of different backgrounds can support a commons’ implementation by openly recognizing data protection limitations in laws, technologies, and policies when applied independently. We propose requirements for deploying a data protection-focused data commons by applying our findings and data protection principles such as purpose limitation and exercising data subject rights to the Institutional Analysis and Development (IAD) framework. Finally, we map the IAD framework into a commons checklist for policy-makers to accommodate co-creation and participation for all stakeholders, balancing the data protection of data subjects with opportunities for seeking value from personal data.
We investigate properties of random mappings whose core is composed of derangements as opposed to permutations. Such mappings arise as the natural framework for studying the Screaming Toes game described, for example, by Peter Cameron. This mapping differs from the classical case primarily in the behaviour of the small components, and a number of explicit results are provided to illustrate these differences.