To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
An empirical average will converge, in some sense, to the corresponding expectation. This famous result, called the Law of Large Numbers, can be anticipated based on the concentration inequalities introduced in the previous chapter, but some appropriate notions of convergence for random variables need to be defined in order to make a rigorous statement. Beyond mere convergence, the fluctuations of an empirical average around the associated expectation can be characterized by the Central Limit Theorem, and are known to be Gaussian in some asymptotic sense. The chapter also discusses the limit of extremes such as the maximum of a sample.
This study assessed neonatal visual maturity in infants with congenital heart disease (CHD) and its predictive value for neurodevelopmental outcomes. Neonates with CHD underwent a standardized visual assessment before and after cardiopulmonary bypass surgery. Visual maturity was rated as normal versus abnormal by means of normative reference data. Twelve-month neurodevelopment was assessed with the Bayley-III. Twenty-five healthy controls served as the reference group. Neonatal visual assessment was performed in five neonates with CHD preoperatively and in 24 postoperatively. Only postoperative assessments were considered for further analysis. Median [IQR] age at assessment was 27.0 [21.5, 42.0] days of life in postoperative neonates with CHD and 24.0 [15.0, 32.0] in controls. Visual performance was within reference values in 87.5% in postoperative CHD versus 90.5% in healthy controls (p = 1.0). Visual maturity was not predictive of neurodevelopment at 12 months. These results demonstrate the limited feasibility and predictive value of neonatal visual assessments in CHD.
We consider a collection of statistically identical two-state continuous time Markov chains (channels). A controller continuously selects a channel with the view of maximizing infinite horizon average reward. A switching cost is paid upon channel changes. We consider two cases: full observation (all channels observed simultaneously) and partial observation (only the current channel observed). We analyze the difference in performance between these cases for various policies. For the partial observation case with two channels or an infinite number of channels, we explicitly characterize an optimal threshold for two sensible policies which we name “call-gapping” and “cool-off.” Our results present a qualitative view on the interaction of the number of channels, the available information, and the switching costs.
Stochastic processes model experiments whose outcomes are collections of variables organized in some fashion. We focus here on Markov processes, which include random walks (think of the fortune of a person gambling on black/red at the roulette over time) and branching processes (think of the behavior of a population of an asexual species where each individual gives birth to a number of otherwise identical offsprings according to a given probability distribution) .
In this chapter we consider distributions on the real line that have a discrete support. It is indeed common to count certain occurrences in an experiment, and the corresponding counts are invariably integer-valued. In fact, all the major distributions of this type are supported on the (non-negative) integers. We introduce the main ones here.
We consider an experiment resulting in two paired numerical variables. The general goal addressed in this chapter is that of quantifying the strength of association between these two variables. By association we mean dependence. Contrary to the previous chapter, here the two variables can be measurements of completely different kinds (e.g., height and weight). Several measures of association are introduced, and used to test for independence.
In some areas of mathematics, physics, and elsewhere, continuous objects and structures are often motivated, or even defined, as limits of discrete objects. For example, in mathematics, the real numbers are defined as the limit of sequences of rational numbers, and in physics, the laws of thermodynamics arise as the number of particles in a system tends to infinity (the so-called thermodynamic or macroscopic limit). Taking certain discrete distributions (discussed in the previous chapter) to their continuous limits, which is done by letting their support size increase to infinity in a controlled manner, gives rise to continuous distributions on the real line. We introduce and discuss such distributions in this chapter, including the normal (aka Gaussian) family of distributions, and in the process cover probability densities.
We consider in this chapter experiments where the variables of interest are paired. Importantly, we assume that these variables are directly comparable (in contrast with the following two chapters). Crossover trials are important examples of such experiments. The main question of interest here is that of exchangeability, which reduces to testing for symmetry when there are only two variables.
When a die (with 3 or more faces) is rolled, the result of each trial can take one of as many possible values. The same is true in the context of an urn experiment, when the balls in the urn are of multiple different colors. Such models are broadly applicable. Indeed, even `yes/no’ polls almost always include at least one other option like `not sure’ or `no opinion’. Another situation where discrete variables arise is when two or more coins are compared in terms of their chances of landing heads, or more generally, when two or more (otherwise identical) dice are compared in terms of their chances of landing on a particular face. In terms of urn experiments, the analog is a situation where balls are drawn from multiple urns. This sort of experiments can be used to model clinical trials where several treatments are compared and the outcome is dichotomous. When the coins are tossed together, or when the dice are rolled together, we might want to test for independence. We thus introduce some classical tests for comparing multiple discrete distributions and for testing for the independence of two or more discrete variables that are observed together.
A prototypical (although somewhat idealized) workflow in any scientific investigation starts with the design of the experiment to probe a question or hypothesis of interest. The experiment is modeled using several plausible mechanisms. The experiment is conducted and the data are collected. These data are finally analyzed to identify the most adequate mechanism, meaning the one among those considered that best explains the data. Although an experiment is supposed to be repeatable, this is not always possible, particularly if the system under study is chaotic or random in nature. When this is the case, the mechanisms above are expressed as probability distributions. We then talk about probabilistic modeling --- albeit with not one but several probability distributions. It is as if we contemplate several probability experiments, and the goal of statistical inference is to decide on the most plausible one in view of the collected data. We introduce core concepts such as estimators, confidence intervals, and tests.
The chapter focuses on discrete probability spaces, where probability calculations are combinatorial in nature. Urn models are presented as the quintessential discrete experiments.
Statistics is the science of data collection and data analysis. We provide, in this chapter, a brief introduction to principles and techniques for data collection, traditionally divided into survey sampling and experimental design --- each the subject of a rich literature. While most of this book is on mathematical theory, covering aspects of Probability Theory and Statistics, the collection of data is, by nature, much more practical, and often requires domain-specific knowledge. And careful data collection is of paramount importance. Indeed, data that were improperly collected can be completely useless and unsalvageable by any technique of analysis. And it is worth keeping in mind that the collection phase is typically much more expensive that the analysis phase that ensues (e.g., clinical trials, car crash tests, etc). Thus the collection of data should be carefully planned according to well-established protocols or with expert advice. We discuss the basics of data collection in this chapter.
This paper is concerned with the optimal number of redundant allocation to n-component coherent systems consisting of heterogeneous dependent components. We assume that the system is built up of L groups of different components, $L\geq 1$, where there are $n_i$ components in group i, and $\sum_{i=1}^{L}n_i=n$. The problem of interest is to allocate $v_i$ active redundant components to each component of type i, $i=1,\dots, L$. To get the optimal values of $v_i$ we propose two cost-based criteria. One of them is introduced based on the costs of renewing the failed components and the costs of refreshing the alive ones at the system failure time. The other criterion is proposed based on the costs of replacing the system at its failure time or at a predetermined time $\tau$, whichever occurs first. The expressions for the proposed functions are derived using the mixture representation of the system reliability function based on the notion of survival signature. We assume that a given copula function models the dependency structure between the components. In the particular case that the system is a series-parallel structure, we provide the formulas for the proposed cost-based functions. The results are discussed numerically for some specific coherent systems.
This chapter introduces Kolmogorov’s probability axioms and related terminology and concepts such as outcomes and events, sigma-algebras, probability distributions and their properties.
In this chapter we introduce and briefly discuss some properties of estimators and tests that make it possible to compare multiple methods addressing the same statistical problem. We discuss the notions of sufficiency and consistency, and various notions of optimality (including minimax optimality), both for estimators and for tests.