To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Many of the preceding chapters involved optimization formulations: linear least squares, Procrustes, low-rank approximation, multidimensional scaling. All these have analytical solutions, like the pseudoinverse for minimum-norm least squares problems and the truncated singular value decomposition for low-rank approximation. But often we need iterative optimization algorithms, for example if no closed-form minimizer exists, or if the analytical solution requires too much computation and/or memory (e.g., singular value decomposition for large problems. To solve an optimization problem via an iterative method, we start with some initial guess and then the algorithm produces a sequence that hopefully converges to a minimizer. This chapter describes the basics of gradient-based iterative optimization algorithms, including preconditioned gradient descent (PGD) for the linear LS problem. PGD uses a fixed step size, whereas preconditioned steepest descent uses a line search to determine the step size. The chapter then considers gradient descent and accelerated versions for general smooth convex functions. It applies gradient descent to the machine learning application of binary classification via logistic regression. Finally, it summarizes stochastic gradient descent.
This chapter introduces matrix factorizations – somewhat like the reverse of matrix multiplication. It starts with the eigendecomposition of symmetric matrices, then generalizes to normal and asymmetric matrices. It introduces the basics of the singular value decomposition (SVD) of general matrices. It discusses a simple application of the SVD that uses the largest singular value of a matrix (the spectral norm), posed as an optimization problem, and then describes optimization problems related to eigenvalues and the smallest singular value. (The “real” SVD applications appear in subsequent chapters.) It discusses the special situations when one can relate the eigendecomposition and an SVD of a matrix, leading to the special class of positive (semi)definite matrices. Along the way there are quite a few small eigendecomposition and SVD examples.
An additional distant wall is known to highly alter the jetting scenarios of wall-proximal bubbles. Here, we combine high-speed photography and axisymmetric volume of fluid (VoF) simulations to quantitatively describe its role in enhancing the micro-jet dynamics within the directed jet regime (Zeng et al., J. Fluid Mech., vol. 896, 2020, A28). Upon a favourable agreement on the bubble and micro-jet dynamics, both experimental and simulation results indicate that the micro-jet velocity increases dramatically as $\eta$ decreases, where $\eta =H/R_{max}$ is the distance between two walls $H$ normalized by the maximum bubble radius $R_{max}$. The mechanism is related to the collapsing flow, which is constrained by the distant wall into a reverse stagnation-point flow that builds up pressure near the bubble's top surface and accelerates it into micro-jets. We further derive an equation expressing the micro-jet velocity $U_{jet}=87.94\gamma ^{0.5}(1+(1/3)(\eta -\lambda ^{1.2})^{-2})$, where ${\gamma =d/R_{max}}$ is the stand-off distance to the proximal wall with $d$ the distance between the initial bubble centre and the wall, $\lambda =R_{y,m}/R_{max}$ with $R_{y,m}$ the distance between the top surface and the proximal wall at the bubble's maximum expansion. Viscosity has a minimal impact on the jet velocity for small $\gamma$, where the pressure buildup is predominantly influenced by geometry.
This chapter reviews vectors and matrices, and basic properties like shape, orthogonality, determinant, eigenvalues, and trace. It also reviews operations like multiplication and transpose. These operations are used throughout the book and are pervasive in the literature. In short, arranging data into vectors and matrices allows one to apply powerful data analysis techniques over a wide spectrum of applications. Throughout, this chapter (and book) illustrates how the ideas are implemented in practice in Julia.
Quantum technologies exploit uniquely quantum behaviour of light and matter to achieve functionalities and performance in applications such as computing, navigation, communications, sensing and imaging that are not achievable by conventional, classical means. These “second-generation” technologies are under rapid development and commercialisation, with an accompanying need for assurance by test and characterisation.
We develop a mean-field model to examine the stability of a ‘quasi-2-D suspension’ of elongated particles embedded within a viscous membrane. This geometry represents several biological and synthetic settings, and we reveal mechanisms by which the anisotropic mobility of particles interacts with long-ranged viscous membrane hydrodynamics. We first show that a system of slender rod-like particles driven by a constant force is unstable to perturbations in concentration – much like sedimentation in analogous 3-D suspensions – so long as membrane viscous stresses dominate. However, increasing the contribution of viscous stresses from the surrounding 3-D fluid(s) suppresses such an instability. We then tie this result to the hydrodynamic disturbances generated by each particle in the plane of the membrane and show that enhancing subphase viscous contributions generates extensional fields that orient neighbouring particles in a manner that draws them apart. The balance of flux of particles aggregating versus separating then leads to a wave number selection in the mean-field model.
Many applications require solving a system of linear equations 𝑨𝒙 = 𝒚 for 𝒙 given 𝑨 and 𝒚. In practice, often there is no exact solution for 𝒙, so one seeks an approximate solution. This chapter focuses on least-squares formulations of this type of problem. It briefly reviews the 𝑨𝒙 = 𝒚 case and then motivates the more general 𝑨𝒙 ≈ 𝒚 cases. It then focuses on the over-determined case where 𝑨 is tall, emphasizing the insights offered by the SVD of 𝑨. It introduces the pseudoinverse, which is especially important for the under-determined case where 𝑨 is wide. It describes alternative approaches for the under-determined case such as Tikhonov regularization. It introduces frames, a generalization of unitary matrices. It uses the SVD analysis of this chapter to describe projection onto a subspace, completing the subspace-based classification ideas introduced in the previous chapter, and also introduces a least-squares approach to binary classifier design. It introduces recursive least-squares methods that are important for streaming data.
There are many applications of the low-rank signal-plus-noise model 𝒀 = 𝑿 + 𝒁 where 𝑿 is a low-rank matrix and 𝒁 is noise, such as denoising and dimensionality reduction. We are interested in the properties of the latent matrix 𝑿, such as its singular value decomposition (SVD), but all we are given is the noisy matrix 𝒀. It is important to understand how the SVD components of 𝒀 relate to those of 𝑿 in the presence of a random noise matrix 𝒁. The field of random matrix theory (RMT) provides insights into those relationships, and this chapter summarizes some key results from RMT that help explain how the noise in 𝒁 perturbs the SVD components, by analyzing limits as matrix dimensions increase. The perturbations considered include roundoff error, additive Gaussian noise, outliers, and missing data. This is the only chapter that requires familiarity with the distributions of continuous random variables, and it provides many pointers to the literature on this modern topic, along with several demos that illustrate remarkable agreement between the asymptotic predictions and the empirical performance even for modest matrix sizes.
Polymer chains in turbulent flows are generally modelled as dumbbells, i.e. two beads joined by a nonlinear spring. The dumbbell only maps a single spatial configuration, described by the polymer end-to-end vector, thus a multi-bead FENE (finitely extensible nonlinear elastic) chain seems a natural improvement for a more accurate characterisation of the polymer spatial conformation. At a large Weissenberg number, a comparison with the more accurate Kuhn chain reveals that the multi-bead FENE chain drastically overestimates the probability of folded configurations. Surprisingly, the dumbbell turns out to be the only meaningful bead-spring model to coarse-grain a polymer macromolecule in turbulent pipe flows.
This chapter focuses on artificial neural network models and methods. Although these methods have been studied for over 50 years, they have skyrocketed in popularity in recent years due to accelerated training methods, wider availability of large training sets, and the use of deeper networks that have significantly improved performance for many classification and regression problems. Previous chapters emphasized subspace models. Subspaces are very useful for many applications, but they cannot model all types of signals. For example, images of a single person’s face (in a given pose) under different lighting conditions lie in a subspace. However, a linear combination of face images from two different people will not look like a plausible face. Thus, all possible face images do not lie in a subspace. A manifold model is more plausible for images of faces (and handwritten digits) and other applications, and such models require more complicated algorithms. Entire books are devoted to neural network methods. This chapter introduces the key methods, focusing on the role of matrices and nonlinear operations. It illustrates the benefits of nonlinearity, and describes the classic perceptron model for neurons and the multilayer perceptron. It describes the basics of neural network training and reviews convolutional neural network models; such models are used widely in applications.
This chapter contains introductory material, including visual examples that motivate the rest of the book. It explains the book formatting, previews the notation, provides pointers for getting started with Julia, and briefly reviews fields and vector spaces.
This chapter discusses the important problem of matrix completion, where we know some, but not all, elements of a matrix and want to “complete” the matrix by filling in the missing entries. This problem is ill posed in general because one could assign arbitrary values to the missing entries, unless one assumes some model for the matrix elements. The most common model is that the matrix is low rank, an assumption that is reasonable in many applications. The chapter defines the problem and describes an alternating projection approach for noiseless data. It discusses algorithms for the practical case of missing and noisy data. It extends the methods to consider the effects of outliers with the robust principal component method, and applies this to video foreground/background separation. It describes nonnegative matrix factorization, including the case of missing data. A particularly famous application of low-rank matrix completion is the “Netflix problem”; this topic is also relevant to dynamic magnetic resonance image reconstruction, and numerous other applications with missing data (incomplete observations).
The linear stability of the Stuart vortices, which is a model of arrays of vortices often observed in the atmosphere and the oceans, in rotating stratified fluids is investigated by local and modal stability analysis. As in the case of the two-dimensional (2-D) Taylor–Green vortices, five types of instability appear in general: the pure-hyperbolic instability, the strato-hyperbolic instability, the rotational-hyperbolic instability, the centrifugal instability and the elliptic instability. The condition for each instability and the estimate of the growth rate derived by Hattori & Hirota (J. Fluid Mech., vol. 967, 2023, A32) are shown to also be useful for the Stuart vortices, which supports their applicability to general flows. The properties of each instability depend on stratification and rotation in a way similar to the case of the 2-D Taylor–Green vortices. For the Stuart vortices, however, the centrifugal instability and the elliptic instability become more dominant than the three hyperbolic instabilities in comparison to the 2-D Taylor–Green vortices; this is explained by the larger ratios of the maximum vorticity and the strain rate at the elliptic stagnation points to the strain rate at the hyperbolic stagnation points. Direct correspondence between the modal and local stability results is further established by comparing unstable modes to solutions to the local stability equations; this is useful for identifying the types of modes since the mechanism of instability is readily known in the local stability analysis. This helps us to discover the modes of the ring-type elliptic instability, which have been predicted only theoretically.
Insects flip their wings around each stroke reversal and may enhance lift in the early stage of a half-stroke. The possible lift-enhancing mechanism of this rapid wing rotation and its strong connection with wake vortices are still underexplored, especially when unsteady leading-edge vortex (LEV) behaviours occur. Here, we numerically studied the lift generation and underlying vorticity dynamics during the rapid rotation of a low aspect ratio flapping wing at a Reynolds number (${\textit {Re}}$) of 1500. Our findings prove that when the outboard LEV breaks down, an advanced rotation can still enhance the lift in the early stage of a half-stroke, which originates from an interaction with the breakdown vortex in the outboard region. This interaction, named the breakdown-vortex jet mechanism, results in a jet and thus a higher pressure on the upwind surface, including a stronger wingtip suction force on the leeward surface. Although the stable LEV within the mid-span retains its growth and location during an advanced rotation, it can be detrimental to lift enhancement as it moves underneath the wing. Therefore, for a flapping wing at ${\textit {Re}}\sim 10^3$, the interactions with stable and breakdown leading-edge vortices lead to the single-vortex suction and breakdown-vortex jet mechanisms, respectively. In other words, the contribution of wing–wake interaction depends on the spanwise location. The current work also implies the importance of wing kinematics to this wing–wake interaction in flapping wings, and provides an alternative perspective for understanding this complex flow phenomenon at ${\textit {Re}}\sim 10^3$.
Previous chapters considered the Euclidean norm, the spectral norm, and the Frobenius norm. These three norms are particularly important, but there are many other important norms for applications. This chapter discusses vector norms, matrix norms, and operator norms, and uses these norms to analyze the convergence of sequences. It revisits the Moore–Penrose pseudoinverse from a norm-minimizing perspective. It applies norms to the orthogonal Procrustes problem and its extensions.