To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
This chapter presents the matrix deviation inequality, a uniform deviation bound for random matrices over general sets. Applications include two-sided bounds for random matrices, refined estimates for random projections, covariance estimation in low dimensions, and an extension of the Johnson–Lindenstrauss lemma to infinite sets. We prove two geometric results: the M* bound, which shows how random slicing shrinks high-dimensional sets, and the escape theorem, which shows how slicing can completely miss them. These tools are applied to a fundamental data science task – learning structured high-dimensional linear models. We extend the matrix deviation inequality to arbitrary norms and use it to strengthen the Chevet inequality and derive the Dvoretzky– Milman theorem, which states that random low-dimensional projections of high-dimensional sets appear nearly round. Exercises cover matrix and process-level deviation bounds, high-dimensional estimation techniques such as the Lasso for sparse regression, the Garnaev–Gluskin theorem on random slicing of the cross-polytope, and general-norm extensions of the Johnson–Lindenstrauss lemma.
This chapter introduces techniques for bounding random processes. We develop Gaussian interpolation to derive powerful comparison inequalities for Gaussian processes, including the Slepian, Sudakov–Fernique, and Gordon inequalities. We use this to get sharp bounds on the operator norm of Gaussian random matrices. We also prove the Sudakov lower bound using covering numbers. We introduce the concept of Gaussian width, which connects probabilistic and geometric perspectives, and apply it to analyze the size of random projections of high-dimensional sets. Exercises cover symmetrization and contraction inequalities for random processes, the Gordon min–max inequality, sharp bounds for Gaussian matrices, the nuclear norm, effective dimension, random projections, and matrix sketching.
This chapter introduces sub-Gaussian and sub-exponential distributions and develops basic concentration inequalities. We prove the Hoeffding, Chernoff, Bernstein, and Khintchine inequalities. Applications include robust mean estimation and analyzing degrees in random graphs. The exercises explore Mills ratio, small ball probabilities, Le Cam’s two-point method, the expander mixing lemma for random graphs, stochastic dominance, Orlicz norms, and the Bennett inequality.
Most of the material in this chapter is from basic analysis and probability courses. Key concepts and results are recalled here, including convexity, norms and inner products, random variables and random vectors, union bound, conditioning, basic inequalities (Jensen, Minkowski, Cauchy–Schwarz, Hölder, Markov, and Chebyshev), the integrated tail formula, the law of large numbers, the central limit theorem, normal and Poisson distributions, and handy bounds on the factorial.
This chapter presents some foundational methods for bounding random processes. We begin with the chaining technique and prove the Dudley inequality, which bounds a random process using covering numbers. Applications include Monte Carlo integration and uniform bounds for empirical processes. We then develop VC (Vapnik– Chervonenkis) theory, offering combinatorial insights into random processes and applying it to statistical learning. Building on chaining, we introduce generic chaining to obtain optimal two-sided bounds using Talagrand’s g2 functional. A key consequence is the Talagrand comparison inequality, a generalization of the Sudakov–Fernique inequality for sub-Gaussian processes. This is used to derive the Chevet inequality, a powerful tool for analyzing random bilinear forms over general sets. Exercises explore the Lipschitz law of large numbers in higher dimensions, one-bit quantization, and the small ball method for heavy-tailed random matrices.
This chapter begins with Maurey’s empirical method – a probabilistic approach to constructing economical convex combinations. We apply it to bound covering numbers and the volumes of polytopes, revealing their counterintuitive behavior in high dimensions. The exercises refine these bounds and culminate in the Carl–Pajor theorem on the volume of polytopes.
This chapter introduces several basic tools in high-dimensional probability: decoupling, concentration for quadratic forms (the Hanson–Wright inequality), symmetrization, and contraction. These techniques are illustrated through estimates of the operator norm of a random matrix. This is applied to matrix completion, where the goal is to recover a low-rank matrix from a random subset of its entries. Exercises explore variants of the Hanson–Wright inequality, mean estimation, concentration of the norm for anisotropic random vectors, distances to subspaces, graph cutting, the concept of type in normed spaces, non-Euclidean versions of the approximate Caratheodory theorem, and covariance estimation.
This chapter begins the study of random vectors in high dimensions, starting by showing their norm concentrates. We give a probabilistic proof of the Grothendieck inequality and apply it to semidefinite optimization. We explore a semidefinite relaxation for the maximum cut, presenting the Goemans–Williamson randomized approximation algorithm. We also give an alternative proof of the Grothendieck inequality with nearly the best known constant using the kernel trick, a method widely used in machine learning. The exercises explore invariant ensembles of random matrix theory, various versions of the Grothendieck inequality, semidefinite relaxations, and the notion of entropy.
This chapter develops a non-asymptotic theory of random matrices. It starts with a quick refresher on linear algebra, including the perturbation theory for matrices and featuring a short proof of the Davis–Kahan inequality. Three key concepts are introduced – nets, covering numbers, and packing numbers – and linked to volume and error-correcting codes. Bounds on the operator norm and singular values of random matrices are established. Three applications are given: community detection in networks, covariance estimation, and spectral clustering. Exercises explore the power method to compute the top singular value, the Schur bound on the operator norm, Hermitian dilation,Walsh matrices, the Wedin theorem on matrix perturbations, a semidefinite relaxation of the cut norm, the volume of high-dimensional balls, and Gaussian mixture models.
This chapter explores methods of concentration that do not rely on independence. We introduce the isoperimetric approach and discuss concentration inequalities across a variety of metric measure spaces – including the sphere, Gaussian space, discrete and continuous cubes, the symmetric group, Riemannian manifolds, and the Grassmannian. As an application, we derive the Johnson–Lindenstrauss lemma, a fundamental result in dimensionality reduction for high-dimensional data. We then develop matrix concentration inequalities, with an emphasis on the matrix Bernstein inequality, which extends the classical Bernstein inequality to random matrices. Applications include community detection in sparse networks and covariance estimation for heavy-tailed distributions. Exercises explore binary dimension reduction, matrix calculus, additional matrix concentration results, and matrix sketching.
'High-Dimensional Probability,' winner of the 2019 PROSE Award in Mathematics, offers an accessible and friendly introduction to key probabilistic methods for mathematical data scientists. Streamlined and updated, this second edition integrates theory, core tools, and modern applications. Concentration inequalities are central, including classical results like Hoeffding's and Chernoff's inequalities, and modern ones like the matrix Bernstein inequality. The book also develops methods based on stochastic processes – Slepian's, Sudakov's, and Dudley's inequalities, generic chaining, and VC-based bounds. Applications include covariance estimation, clustering, networks, semidefinite programming, coding, dimension reduction, matrix completion, and machine learning. New to this edition are 200 additional exercises, alongside extra hints to assist with self-study. Material on analysis, probability, and linear algebra has been reworked and expanded to help bridge the gap from a typical undergraduate background to a second course in probability.
Play of Chance and Purpose emphasizes learning probability, statistics, and stochasticity by developing intuition and fostering imagination as a pedagogical approach. This book is meant for undergraduate and graduate students of basic sciences, applied sciences, engineering, and social sciences as an introduction to fundamental as well as advanced topics. The text has evolved out of the author's experience of teaching courses on probability, statistics, and stochastic processes at both undergraduate and graduate levels in India and the United States. Readers will get an opportunity to work on several examples from real-life applications and pursue projects and case-study analyses as capstone exercises in each chapter. Many projects involve the development of visual simulations of complex stochastic processes. This will augment the learners' comprehension of the subject and consequently train them to apply their learnings to solve hitherto unseen problems in science and engineering.
Complementing the presentation of the Gaussian free field (GFF) with zero boundary conditions in Chapter 1, and on manifolds in Chapter 5, we devote this chapter to studying further variants of the field. The main example is the GFF in two dimensions with Neumann (or free) boundary conditions. We give a rigorous definition of this field as both a stochastic process indexed by suitable test functions and as a random distribution modulo constants. As in Chapter 1, we show the equivalence of these two viewpoints; however, in this case, further analytical arguments are required. We describe the covariance function of the field, prove key properties such as conformal covariance and the spatial Markov property, and discuss its associated Gaussian multiplicative chaos measures on the boundary of the domain where it is defined. We also cover the definition and properties of the whole plane Gaussian free field and the Gaussian free field with Dirichlet–Neumann boundary conditions, building on the construction of the Neumann GFF. We further prove that the whole plane GFF can be decomposed into a Dirichlet part and a Neumann part. Finally, we show that the total mass of the GMC associated to a Neumann GFF on the unit disc is almost surely finite.