To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
In this chapter, we present the notions of Markov decision problem, the T-stage evaluation and the discounted evaluation. We introduce and study contracting mappings, and use such mappings to show that the decision maker has a stationary discounted optimal strategy. We also define the concept of uniform optimality, and show that the decision maker has a stationary uniformly optimal strategy.
In this chapter, we define infinite orbits and show how to approximate such orbits. We prove that for every function that has no fixed point there is an approximate infinite orbit with unbounded variation, and we use this result to show that a certain class of quitting games admits undiscounted $\ep$-equilibria.
In this chapter, we extend the notion of discounted payoff to the model of stochastic games, and we define the concept of discounted equilibrium. We then prove that every two-player zero-sum stochastic game admits a discounted value, and that each player has a stationary discounted optimal strategy. The proof uses the same tools we employed in Chapter~\ref{section:mdp} to prove that in Markov decision problems the decision maker has a stationary discounted optimal strategy.
We finally prove that the discounted value is continuous in the parameters of the game, namely the payoff function, the transition function, and the discount factor.
In this chapter, we proveKakutani's Fixed Point Theorem, which is an extension of Brouwer's Fixed Point Theoremto correspondences (set-valued mappings). We then define the concept of $\lambda$-discounted equilibrium, and using Kakutani's Fixed Point Theorem we prove that every multiplayer stochastic game admits a stationary $\lambda$-discounted equilibrium, for every discount factor $\lambda \in (0,1]$.
Two simple mathematical models of advection and diffusion of hydrogen within the retina are discussed. The work is motivated by the hydrogen clearance technique, which is used to estimate blood flow in the retina. The first model assumes that the retina consists of three, well-mixed layers with different thickness, and the second is a two-dimensional model consisting of three regions that represent the layers in the retina. Diffusion between the layers and leakage through the outer edges are considered. Solutions to the governing equations are obtained by employing Fourier series and finite difference methods for the two models, respectively. The effect of important parameters on the hydrogen concentration is examined and discussed. The results contribute to understanding the dispersal of hydrogen in the retina and in particular the effect of flow in the vascular retina. It is shown that the predominant features of the process are captured by the simpler model.
Stochastic games have an element of chance: the state of the next round is determined probabilistically depending upon players' actions and the current state. Successful players need to balance the need for short-term payoffs while ensuring future opportunities remain high. The various techniques needed to analyze these often highly non-trivial games are a showcase of attractive mathematics, including methods from probability, differential equations, algebra, and combinatorics. This book presents a course on the theory of stochastic games going from the basics through to topics of modern research, focusing on conceptual clarity over complete generality. Each of its chapters introduces a new mathematical tool – including contracting mappings, semi-algebraic sets, infinite orbits, and Ramsey's theorem, among others – before discussing the game-theoretic results they can be used to obtain. The author assumes no more than a basic undergraduate curriculum and illustrates the theory with numerous examples and exercises, with solutions available online.
We develop a new analytical solution of a three-dimensional atmospheric pollutant dispersion. The main idea is to subdivide vertically the planetary boundary layer into sub-layers, where the wind speed and eddy diffusivity assume average values for each sub-layer. Basically, the model is assessed and validated using data obtained from the Copenhagen diffusion and Prairie Grass experiments. Our findings show that there is a good agreement between the predicted and observed crosswind-integrated concentrations. Moreover, the calculated statistical indices are within the range of acceptable model performance.
The paper deals with the construction of explicit Nordsieck second derivative general linear methods with s stages of order p with $p=s$ and high stage order $q=p$ with inherent Runge–Kutta or quadratic stability properties. Satisfying the order and stage order conditions together with inherent stability conditions leads to methods with some free parameters, which will be used to obtain methods with a large region of absolute stability. Examples of methods with r external stages and $p=q=s=r-1$ up to order five are given, and numerical experiments in a fixed stepsize environment are presented.
This chapter describes methods based on gradient information that achieve faster rates than basic algorithms such as those described in Chapter 3. These accelerated gradient methods, most notably the heavy-ball method and Nesterov’s optimal method, use the concept of momentum which means that each step combines information from recent gradient values but also earlier steps. These methods are described and analyzed using an analysis based on Lyapunov functions. The cases of convex and strongly convex functions are analyzed separately. We motivate these methods using continuous-time limits, which link gradient methods to dynamical systems described by differential equations. We mention also the conjugate gradient method, which was developed separately from the other method but which also makes use of momentum. Finally, we discuss the concept of lower bounds on algorithmic complexity, introducing a function on which no method based on gradients can attain convergence faster than a certain given rate.
Here, we describe methods for minimizing a smooth function over a closed convex set, using gradient information. We first state results that characterize optimality of points in a way that can be checked, and describe the vital operation of projection onto the feasible set. We next describe the projected gradient algorithm, which is in a sense the extension of the steepest-descent method to the constrained case, analyze its convergence, and describe several extensions. We next analyze the conditional-gradient method (also known as “Frank-Wolfe”) for the case in which the feasible set is compact and demonstrate sublinear convergence of this approach when the objective function is convex.
Here, we discuss concepts of duality for convex optimization problems, and algorithms that make use of these concepts. We define the Lagrangian function and its augmented Lagrangian counterpart. We use the Lagrangian to derive optimality conditions for constrained optimization problems in which the constraints are expressed as linear algebraic conditions. We introduce the dual problem, and discuss the concepts of weak and strong duality, and show the existence of positive duality gaps in certain settings. Next, we discuss the dual subgradient method, the augmented Lagrangian method, and the alternating direction method of multipliers (ADMM), which are useful for several types of data science problems.
In this introductory chapter, we outline the ways in which various problems in data analysis can be formulated as optimization problems. Specifically, we discuss least squares problems, problems in matrix optimization (particularly those involving low-rank matrices), linear and kernel support vector machines, binary and multiclass logistic regression, and deep learning. We also outline the scope of the remainder of the book.
We describe the stochastic gradient method, the fundamental algorithm for several important problems in data science, including deep learning. We give several example problems for which this method is suitable, then described its operation for the simple problem of computing a mean of a collection of values. We related it to a classical method, the Kaczmarz method for solving a system of linear equalities and inequalities. Next, we describe the key assumptions to be used in convergence analysis, then describe the convergence rates attainable by several variants of stochastic gradient under several scenarios. Finally, we discuss several aspects of practical implementation of stochastic gradient, including minibatching and acceleration.