1. Introduction
Geometric measure theory explores the relationship between the geometry of subsets of
$\operatorname{\mathbb{T}}^d$, and analytic properties of Borel measures supported on those subsets. From the perspective of harmonic analysis, it is interesting to explore what geometric information can be gathered from the Fourier analytic properties of these measures, in particular, their Fourier decay. Some results show that the support of measures with Fourier decay must contain certain arrangements of points, such as arithmetic progressions [Reference Chan, Łaba and Pramanik1, Reference Henriot, Łaba and Pramanik6, Reference Łaba and Pramanik16]. In this paper, we work in the opposite direction, showing that most sets supporting measures with a certain type of Fourier decay do not contain certain configurations. More precisely, for certain sets
$Z \subset (\operatorname{\mathbb{T}}^d)^n$, and
$s \gt 0$, we show that a ‘generic’ compact set
$E \subset \operatorname{\mathbb{T}}^d$ supporting a measure
$\mu$ such that
$|\widehat{\mu}(\xi)| \lesssim |\xi|^{\varepsilon -s/2}$ for any
$\varepsilon \gt 0$ also avoids the pattern defined by
$Z$, in the sense that for any distinct points
$x_1,\dots,x_n \in E$,
$(x_1,\dots,x_n) \not \in Z$. As an example, one might have
$Z = \{(x_1,\dots,x_n) \in \operatorname{\mathbb{T}}^{dn}: x_n = f(x_1,\dots,x_{n-1}) \}$, in which case a set
$E$ avoids
$Z$ precisely when, for any distinct points
$x_1,\dots,x_n \in E$,
$x_n \neq f(x_1,\dots,x_{n-1})$, i.e.
$E$ does not contain nontrivial solutions to the equation
$x_n = f(x_1,\dots,x_{n-1})$.
A useful statistic associated with any Borel set
$E \subset \operatorname{\mathbb{T}}^d$ is its Fourier dimension; given a finite Borel measure
$\mu$, its Fourier dimension
$\operatorname{dim}_{\mathbb{F}}(\mu)$ is the supremum of all
$s \in [0,d]$ such that
$\sup_{\xi \in \operatorname{\mathbb{Z}}^d} |\widehat{\mu}(\xi)| |\xi|^{s/2} \lt \infty$. The Fourier dimension of a compact Borel set
$E$ is then the supremum of
$\operatorname{dim}_{\mathbb{F}}(\mu)$, where
$\mu$ ranges over all Borel probability measures
$\mu$ with
$\text{supp}(\mu) \subset E$. A particularly tractable family of sets in this scheme are Salem sets, sets whose Fourier dimension agrees with their Hausdorff dimension. Often constructions in the pattern avoidance literature focus on constructing sets with large Hausdorff dimension which avoid patterns [Reference Denson, Pramanik and Zahl3, Reference Fraser and Pramanik5, Reference Máthé11], but it is not necessarily possible to use the same techniques to construct large Salem sets avoiding patterns, or even to construct sets with positive Fourier dimension, since having large Fourier dimension is often a much stronger property than having large Hausdorff dimension. Nonetheless, in this paper we are able to provide techniques proving the existence of large Salem pattern avoiding sets.
One heuristic reason why Fourier decay can guarantee patterns where other notions of dimension is that the existence of a given pattern can often be detected via understanding a ‘blending’ of
$\mu$ with itself (e.g. for linear patterns, this operation is convolution). Such blending operations often magnify Fourier decay, which acts as a kind of ‘uniform measure of smoothness at all frequency scales’. Such magnification is not necessarily true without the assumption of uniform Fourier decay. For instance, if
$\text{supp}(\mu)$ is contained in the middle thirds Cantor set (with Hausdorff dimension
$\log_3(2)$), the convolution
$\mu * \mu * \mu * \mu$ must be singular, concentrated near numbers with particular ternary expansions (reflecting the fact that
$\widehat{\mu}(3^n)$ cannot tend to zero as
$n \to \infty$), whereas for a measure
$\mu$ with
$|\widehat{\mu}(\xi)| \leq |\xi|^{-\log_3(2)/2}$, the convolution
$\mu * \mu * \mu * \mu$ is always a continuous function, guaranteeing this lack of concentration. Nonetheless, the distinction between Hausdorff and Fourier dimension in pattern avoidance can be subtle, and this is especially so when studying non-linear patterns.
Our paper is part of a growing body of literature on pattern avoidance problems: given a set
$Z \subset \operatorname{\mathbb{T}}^{dn}$, the pattern avoidance problem for
$Z$ asks to construct a pattern avoiding set
$E \subset \operatorname{\mathbb{T}}^d$, a set such that for any distinct
$x_1,\dots,x_n \in E$,
$(x_1,\dots,x_n) \not \in Z$, which is as large as possible with respect to some particular statistic relevant to the problem, such as the Hausdorff or Fourier dimension. The main inspiration for the results of this paper were the results of [Reference Denson, Pramanik and Zahl3] on ‘rough’ patterns, which constructed, for any set
$Z \subset \operatorname{\mathbb{T}}^{dn}$ formed from the countable union of closed sets with lower Minkowski dimension at most
$\alpha$, a set
$E$ avoiding
$Z$ with
\begin{equation}
\operatorname{dim}_{\mathbb{H}}(E) = \min\left( \frac{dn - \alpha}{n - 1}, d \right).
\end{equation} While the sets
$E$ constructed using this method are not guaranteed to be Salem, the techniques in the method (an iterative random selection procedure) seemed amenable to produce Salem sets, since random methods are often very useful for constructing Salem sets (randomly constructed functions are often uncorrelated with oscillation at any particular frequency, and thus tend to have good Fourier decay properties). Our goal was thus to modify the construction of [Reference Denson, Pramanik and Zahl3] in order to ensure the resulting sets constructed were Salem.
Conjecture 1.1. For any set
$Z \subset \operatorname{\mathbb{T}}^{dn}$ which is the countable union of closed sets of lower Minkowski dimension
$\alpha$, there exists a set
$E \subset \operatorname{\mathbb{T}}^d$ with
\begin{equation*} \operatorname{dim}_{\mathbb{F}}(E) = \min \left( \frac{dn - \alpha}{n-1}, d \right), \end{equation*} such that for any distinct
$x_1,\dots,x_n \in E$,
$(x_1,\dots,x_n) \not \in Z$.
Theorem 38 of [Reference Denson2] provides a baseline for this conjecture, constructing a Salem set
$E$ avoiding
$Z$ with
\begin{equation}
\operatorname{dim}_{\mathbb{F}}(E) = \min \left( \frac{dn - \alpha}{n}, d \right).
\end{equation} In this paper, we are only able to prove Conjecture 1.1 when
$Z$ exhibits a weak translation symmetry, as detailed in Theorem 1.4 of this paper. But even for more general sets, we are able to improve upon the dimension given by (1.2), as detailed in Theorems 1.2 and 1.3.
The assumption in Conjecture 1.1 that
$Z$ is formed from sets with a given lower Minkowski dimension is not the only way one might measure the ‘dimension’ of a pattern. For instance, one might instead use the Hausdorff dimension of
$Z$ (a weaker assumption than lower Minkowski dimension) or perhaps even the Fourier dimension of
$Z$ itself (a stronger assumption). But lower Minkowski dimension provides the most effective way of discretizing the pattern at each scale in our arguments. We are unsure whether changing these assumptions changes the Fourier dimension of the set
$E$ constructed in Conjecture 1.1.
The methods in this paper are generic, in the sense of the Baire category theorem; we define a complete metric space
$\mathcal{X}_\beta$ for each
$\beta \in (0,d]$, which consists of all pairs
$(E,\mu)$, where
$E$ is a compact set, and
$\mu$ is a Borel probability measure supported on
$E$ with
$\operatorname{dim}_{\mathbb{F}}(\mu) \geq \beta$. We then show that for an appropriate choice of
$\beta$, the family of all pairs
$(E,\mu) \in \mathcal{X}_\beta$ such that
$E$ is Salem and avoids a pattern is comeager, or generic in
$\mathcal{X}_\beta$ (the complement of a set of first category). In particular, the Baire category theorem then guarantees that there exists many sets
$E$ with
$\operatorname{dim}_{\mathbb{F}}(E) = \beta$ which avoid a given pattern.
Many other approaches [Reference Denson, Pramanik and Zahl3, Reference Fraser and Pramanik5, Reference Keleti9] to the pattern avoidance problem construct large pattern avoiding sets explicitly, exchanging nonconstructive Baire category type methods for various constructive queuing techniques. The approaches in this paper can be modified to give a constructive queuing argument for the existence of sets with the properties guaranteed in Theorems 1.2, 1.3, and 1.4. But using Baire category techniques allows us to avoid the technical numerology that goes into these constructions, so that we can focus on the more novel aspects of our analysis. See Remark 4.3 for a brief outline of how one might obtain such a queuing construction.
Let us now introduce the three primary results of this paper. Theorem 1.2 has the weakest conclusions and its proof relies on a small modifications of previous techniques that exist in the literature, e.g. in [Reference Körner7]. But the theorem works for the most general family of patterns, and its proof acts as a good foundation on which to build the more novel and powerful theorems in this paper.
Theorem 1.2. Fix
$0 \leq \alpha \lt dn$, and let
$Z \subset \operatorname{\mathbb{T}}^{dn}$ be a compact set with lower Minkowski dimension at most
$\alpha$. Set
\begin{equation*} \beta_0 = \min \left( \frac{dn - \alpha}{n-1/2}, d \right). \end{equation*} Then there exists a compact Salem set
$E \subset \operatorname{\mathbb{T}}^d$ with
$\operatorname{dim}_{\mathbb{F}}(E) = \beta_0$, such that for any distinct points
$x_1, \dots, x_n \in E$,
$(x_1, \dots, x_n) \not \in Z$. Moreover, if
$\beta \leq \beta_0$, then the family of all pairs
$(E,\mu) \in \mathcal{X}_\beta$ such that
$E$ is Salem and avoids the pattern generated by
$Z$ is comeager.
The remaining two results give a more novel analysis, improving upon the result of Theorem 1.2 when the pattern
$Z$ satisfies additional regularity conditions. In these latter theorems, we focus on patterns specified by equations of the form
i.e. where
$Z = \{(x_1,\dots,x_n): x_n = f(x_1,\dots,x_{n-1}) \}$, and our assumptions will be on structure of the function
$f$. Under the assumption that
$f$ is smooth, and satisfies a regularity condition geometrically equivalent to the graph of
$f$ being transverse to any axis-oriented
$(n-2) \times d$ dimensional plane in
$(\operatorname{\mathbb{R}}^d)^{n-1}$, we are able to improve the Fourier dimension bound obtained using previous construction techniques, though not quite enough to match the Hausdorff dimension bound obtained in [Reference Fraser and Pramanik5], except in the fairly trivial case where
$n = 2$ (though see the fourth remark below).
Theorem 1.3. Consider a smooth function
$f: V \to \operatorname{\mathbb{T}}^d$, where
$V$ is an open subset of
$\operatorname{\mathbb{T}}^{d(n-1)}$, such that for each
$k \in \{1, \dots, n-1 \}$, the matrix
\begin{equation*} D_{x_k} f(x_1,\dots,x_{n-1}) = \begin{pmatrix} \frac{\partial f_i}{\partial (x_k)_j} \end{pmatrix}_{1 \leq i,j \leq d} \end{equation*} is invertible whenever
$x_1,\dots,x_{n-1}$ are distinct and
$(x_1,\dots,x_{n-1}) \in V$. Then there exists a compact Salem set
$E \subset \operatorname{\mathbb{T}}^d$ with dimension
\begin{equation*} \beta_0 = \begin{cases} d &: n = 2 \\ d/(n - 3/4) &: n \geq 3 \end{cases} \end{equation*} such that for any distinct points
$x_1, \dots, x_n \in E$, with
$x_1,\dots,x_{n-1} \in V$,
Moreover, if
$\beta \leq \beta_0$, then the family of pairs
$(E,\mu) \in \mathcal{X}_\beta$ such that
$E$ is Salem and does not contain any solutions to the equation
$x_n = f(x_1,\dots,x_{n-1})$ is comeager.
Finally, we consider patterns defined by equations which feature some translation invariance. Here we can construct Salem sets with dimension exactly matching the Hausdorff dimension results obtained in [Reference Denson, Pramanik and Zahl3]. The simplest example of such a pattern is that specified by a family of equations of the form
$m_1x_1 + \dots + m_nx_n = s$, where at least two of the integers
$m_1,\dots,m_n \in \operatorname{\mathbb{Z}}$ is nonzero, and
$s$ ranges over a low dimension set in
$\operatorname{\mathbb{T}}^d$. But we can also consider more nonlinear patterns, such as those formed by solutions to an equation
for a non-zero rational number
$a$, and a locally Lipschitz function
$f$. Even in the case where
$f$ is linear, this theorem gives new results.
Theorem 1.4. Fix
$d(n-1) \leq \alpha \lt dn$, a non-zero rational number
$a$, and a locally Lipschitz function
$f: V \to \operatorname{\mathbb{R}}$, where
$V$ is an open subset of
$\operatorname{\mathbb{T}}^{d(n-2)}$. Suppose
$F \subset \operatorname{\mathbb{R}}$ is a compact set with lower Minkowski dimension at most
$\alpha - d(n-1)$. Set
\begin{equation*} \beta_0 = \frac{dn - \alpha}{n - 1}. \end{equation*} Then there exists a compact Salem set
$E \subset \operatorname{\mathbb{T}}^d$ with
$\operatorname{dim}_{\mathbb{F}}(E) = \beta_0$ such that for any distinct points
$x_1,\dots,x_n \in E$, with
$(x_1,\dots,x_{n-2}) \in V$,
Moreover, if
$\beta \leq \beta_0$, then the family of all pairs
$(E,\mu) \in \mathcal{X}_\beta$ such that
$E$ is Salem and does not contain distinct points satisfying the equation above is comeager in
$\mathcal{X}_\beta$.
Under the assumptions of Theorem 1.4, the set
\begin{equation*} Z = \{(x_1,\dots,x_n) \in \operatorname{\mathbb{T}}^d: x_n - ax_{n-1} - f(x_1,\dots,x_{n-2}) \in F \} \end{equation*} has lower Minkowski dimension at most
$\alpha$. Observing the value
$\beta_0$ in that Theorem, we see that in this setting, we see we have verified Conjecture 1.1 for such patterns.
Theorem
(1) Because we are using Baire category techniques, the results we obtain remain true when, instead of avoiding a single pattern, we avoid a countable family of patterns. This is because the countable union of sets of first category is a set of first category. As an example of this property, we note that the conclusion of Theorem 1.2 holds when
$Z$ is replaced by a countable union of compact sets, each with lower Minkowski dimension at most
$\alpha$. Similar generalizations apply to Theorems 1.3 and 1.4.(2) It is quite surprising that we are able to generically improve the bound (1.2) in the sense of the Baire category theorem. In [Reference Shmerkin and Suomala14], for each
$\beta$, a natural probability measure, the fractal percolation, is constructed on the space of all
$\beta$-dimensional subsets of
$\operatorname{\mathbb{T}}^d$, and the question of for which
$\beta$ a fractal percolation almost surely contains a given pattern is studied. Such fractal percolations will almost surely be Salem sets, which connects such questions to our problem. The patterns they study in their paper are slightly different from the regimes of this paper under which sharper results are obtained, i.e. as given in Theorems 1.3 and 1.4. But even using the more general result of Theorem 1.2 we see that the parameters under which a random
$\beta$ dimensional set almost surely contains a pattern behave quite differently to the parameters under which a generic
$\beta$ dimensional set in
$\mathcal{X}_\beta$ contains a pattern.More precisely, consider the set
$Z \subset \operatorname{\mathbb{T}}^{dn}$ defined to be the family of all translations and dilations of a fixed tuple of points
$(y_1,\dots,y_n) \in \operatorname{\mathbb{T}}^{dn}$. Then
$Z$ has dimension
$d+1$, which we will label
$\alpha$. Theorem 1.1 of [Reference Shmerkin and Suomala14] implies that for any
$\beta \gt (dn - \alpha)/n$, a random fractal percolation
$E \subset \operatorname{\mathbb{T}}^d$ will almost surely contain points
$x_1,\dots,x_n \in E$ such that
$(x_1,\dots,x_n) \in Z$. On the other hand, Theorem 1.2 of this paper implies that for any
$\beta \leq (dn - \alpha)/(n-1/2)$, a generic element of
$\mathcal{X}_\beta$ does not contain any instances of the pattern specified by
$Z$. Thus the theorems in this paper imply that generic pattern avoidance in a Baire category sense differs from generic pattern avoidance in a probabilistic sense, at least in the range
$(dn - \alpha)/n \lt \beta \leq (dn - \alpha)/(n - 1/2)$.(3) If
$Z \subset \operatorname{\mathbb{T}}^{dn}$ is a compact set with lower Minkowski dimension
$\alpha$ with
$0 \leq \alpha \lt d$, then for any
$1 \leq i \leq d$, the set
$\operatorname{\mathbb{T}}^d - \pi_i(Z)$ has full Hausdorff dimension
$d$, where
$\pi_i: \operatorname{\mathbb{T}}^{dn} \to \operatorname{\mathbb{T}}^d$ is given by
$\pi_i(x_1,\dots,x_n) = x_i$. Thus the pattern avoidance problem is trivial in this case for Hausdorff dimension. This is no longer true when studying Fourier dimension, since
$\operatorname{\mathbb{T}}^d - \pi_i(Z)$ need not be a Salem set, nor even have particularly large Fourier dimension, making the results of this paper still interesting in this range.That this is true is hinted at in Example 8 of [Reference Ekström, Persson and Schmeling4], where it is shown that there exists a set
$X \subset \operatorname{\mathbb{T}}$ which is the countable union of compact sets
$\{X_k \}$, with
$\sup_k \operatorname{dim}_{\mathbb{M}}(X_k) \leq 3/4$, such that
$\operatorname{dim}_{\mathbb{F}}(\operatorname{\mathbb{T}} - X) \leq 3/4$. Thus
$\operatorname{\mathbb{T}} - X$ is not a Salem set, since
$\operatorname{\mathbb{T}}$ has Hausdorff dimension one. The pattern
\begin{equation*} Z = \bigcup_{i = 0}^{n-1} \mathbb{Q}^i \times X \times \mathbb{Q}^{n-i-1}, \end{equation*}is a countable union of compact sets, each with Minkowski dimension at most
$3/4$. On the other hand, for each
$i \in \{1, \dots, n \}$, we find that
\begin{equation*} \operatorname{dim}_{\mathbb{F}}(\operatorname{\mathbb{T}} - \pi_i(Z)) \leq \operatorname{dim}_{\mathbb{F}}(\operatorname{\mathbb{T}} - X) \leq 3/4. \end{equation*}Thus the trivial solution obtained by removing a projection of
$Z$ onto a particular coordinate axis does not necessarily give a pattern avoiding set with optimal Fourier dimension in this setting. Applying Theorem 1.2 naively to the pattern
$Z$ shows that a generic Salem set
$E \subset \operatorname{\mathbb{T}}$ of dimension
$(n-3/4)/(n-1/2)$ avoids
$Z$, which exceeds the dimension of the trivial construction for all
$n \gt 1$. In fact, a generic Salem set
$E \subset \operatorname{\mathbb{T}}$ of dimension
$1$ will avoid
$Z$, since any subset of
$\operatorname{\mathbb{T}} - X$ will avoid
$Z$, and Theorem 1.2 applied with
$Z = X$ implies that a generic Salem set
$E$ of dimension 1 will be contained in
$\operatorname{\mathbb{T}} - X$.(4) If
$n = 2$, the problem of constructing a set which avoids solutions to the equation
$y = f(x)$ for a continuous function
$f: V \to \operatorname{\mathbb{T}}^d$ is essentially trivial. If there exists
$x \in \operatorname{\mathbb{T}}^d$ such that
$f(x) \neq x$, there there exists an open set
$U$ around
$x$ such that
$U \cap f(U) = \emptyset$. Then
$U$ has full Fourier dimension, and avoids solutions to the equation
$y = f(x)$. On the other hand, if
$f(x) = x$ for all
$x$, then there are no distinct
$x$ and
$y$ in
$[0,1]$ such that
$y = f(x)$, and so the problem is also trivial. But it is a less trivial to argue that a generic set with full Fourier dimension avoids this pattern, which is proved in Theorem 1.3, so we still obtain nontrivial information in this case.(5) Working on patterns on
$\operatorname{\mathbb{R}}^d$ is not significantly different from working over
$\operatorname{\mathbb{T}}^d$. For our purposes, the latter domain has several notational advantages, which is why in this paper we have chosen to work with the pattern avoidance pattern in this setting. But there is no theoretical obstacle in applying the techniques described here to prove the existence of pattern avoiding sets in
$\operatorname{\mathbb{R}}^d$. Let us briefly describe how this can be done. Given a finite Borel measure
$\mu$ on
$\operatorname{\mathbb{R}}^d$, we define the Fourier dimension
$\operatorname{dim}_{\mathbb{F}}(\mu)$ of
$\mu$ to be the supremum of all
$s \in [0,d]$ such that
$\sup_{\xi \in \operatorname{\mathbb{R}}^d} |\widehat{\mu}(\xi)| |\xi|^{s/2} \lt \infty$, and define
$\operatorname{dim}_{\mathbb{F}}(E)$ for a Borel set
$E \subset \operatorname{\mathbb{R}}^d$ to be the supremum of the quantities
$\operatorname{dim}_{\mathbb{F}}(\mu)$, taken over finite Borel measures
$\mu$ supported on
$E$. It is a simple consequence of the Poisson summation formula that if
$\mu$ is a compactly supported finite measure on
$\operatorname{\mathbb{R}}^d$, and we consider the periodization
$\mu^*$ of
$\mu$, i.e. the finite Borel measure on
$\operatorname{\mathbb{T}}^d$ such that for any
$f \in C(\operatorname{\mathbb{T}}^d)$,
(1.3)
\begin{equation}
\int_{\operatorname{\mathbb{T}}^d} f(x)\; d\mu^*(x) = \int_{\operatorname{\mathbb{R}}^d} f(x)\; d\mu(x),
\end{equation}then
$\operatorname{dim}_{\mathbb{F}}(\mu^*) = \operatorname{dim}_{\mathbb{F}}(\mu)$. A proof is given in Lemma 39 of [Reference Denson2]. Since
$\mu$ is compactly supported, it is also simple to see that
$\operatorname{dim}_{\mathbb{H}}(\mu^*) = \operatorname{dim}_{\mathbb{H}}(\mu)$. It follows that if
$E$ is a compact subset of
$[0,1)^d$, and
$\pi: [0,1)^d \to \operatorname{\mathbb{T}}^d$ is the natural projection map, then
$\operatorname{dim}_{\mathbb{F}}(E) = \operatorname{dim}_{\mathbb{F}}(\pi(E))$ and
$\operatorname{dim}_{\mathbb{H}}(E) = \operatorname{dim}_{\mathbb{H}}(\pi(E))$. These results therefore imply we can reduce the study of patterns on
$\operatorname{\mathbb{R}}^{dn}$ to patterns on
$\operatorname{\mathbb{T}}^{dn}$, and thus obtain analogous results to Theorems 1.2, 1.3, and 1.4 for the construction of Salem sets avoiding patterns in
$\operatorname{\mathbb{R}}^d$, where every instance of
$\operatorname{\mathbb{T}}$ in the statement of those theorems is replaced by
$[0,1]$.
2. Notation
• Given a metric space
$X$, a point
$x \in X$, and a positive number
$\varepsilon \gt 0$, we let
$B_\varepsilon(x)$ denote the open ball of radius
$\varepsilon$ around
$x$. For
$x \in X$, we let
$\delta_x$ denote the Dirac delta measure at
$x$. For a set
$E \subset X$ and
$\varepsilon \gt 0$, we let
$N(E,\varepsilon) = \bigcup_{x \in E} B_\varepsilon(x)$ denote the
$\varepsilon$-neighborhood of the set
$E$. For two sets
$E_1,E_2 \subset X$, we let
\begin{equation*} d(E_1,E_2) = \inf \{d(x_1,x_2) : x_1 \in E_1, x_2 \in E_2 \}, \end{equation*}and then define the Hausdorff distance
\begin{equation*} d_{\mathbb{H}}(E_1,E_2) = \max \left( \sup_{x_1 \in E_1} d(x_1,E_2), \sup_{x_2 \in E_2} d(E_1,x_2) \right). \end{equation*}If
$d_{\mathbb{H}}(E_1,E_2) \lt \varepsilon$, then
$E_1 \subset N(E_2,\varepsilon)$ and
$E_2 \subset N(E_1,\varepsilon)$, and the Hausdorff distance can be described as the infimum of such
$\varepsilon$.• A subset of a metric space
$X$ is of first category, or meager in
$X$ if it is a subset of the countable union of closed sets with empty interior, and is comeager if it is the complement of such a set (it contains a countable intersection of open, dense sets). We say a property holds quasi-always, or a property is generic in
$X$, or quasi-all elements of
$X$ have a property, if the set of points in
$X$ satisfying that property is comeager. The Baire category theorem then states precisely that any comeager set in a complete metric space is dense.• We let
$\operatorname{\mathbb{T}}^d = \operatorname{\mathbb{R}}^d/\operatorname{\mathbb{Z}}^d$. Given
$x \in \operatorname{\mathbb{T}}$, we let
$|x|$ denote the minimal absolute value of an element of
$\operatorname{\mathbb{R}}$ lying in the coset of
$x$. For
$x \in \operatorname{\mathbb{T}}^d$, we let
$|x| = \sqrt{|x_1|^2 + \dots + |x_d|^2}$. The canonical metric on
$\operatorname{\mathbb{T}}^d$ is then given by
$d(x,y) = |x - y|$, for
$x,y \in \operatorname{\mathbb{T}}^d$.For an axis-oriented cube
$Q$ in
$\operatorname{\mathbb{T}}^d$, and some
$t \gt 0$, we let
$tQ$ be the axis-oriented cube in
$\operatorname{\mathbb{T}}^d$ with the same centre and
$t$ times the sidelength.We say a family of subsets
$\mathcal{A}$ of
$\operatorname{\mathbb{T}}^d$ is downward closed if, whenever
$E \in \mathcal{A}$, any subset of
$E$ is also an element of
$\mathcal{A}$. The quintessential downward closed family for our purposes, given a set
$Z \subset \operatorname{\mathbb{T}}^{dn}$, is the collection of all sets
$E \subset \operatorname{\mathbb{T}}^d$ that avoid the pattern
$Z$, i.e. such that for any distinct points
$x_1,\dots,x_n \in E$,
$(x_1,\dots,x_n) \not \in Z$.• For
$\alpha \in [0,d]$ and
$\delta \gt 0$, the
$(\alpha,\delta)$ Hausdorff content of a Borel set
$E \subset \operatorname{\mathbb{T}}^d$ is
\begin{equation*}
H^\alpha_\delta(E) = \inf \left\{\sum_{k = 1}^\infty \varepsilon_k^\alpha : E \subset \bigcup_{k = 1}^\infty B_{\varepsilon_k}(x_k)\ \text{and}\ 0 \lt \varepsilon_k \leq \delta\ \text{for all}\ k \geq 1 \right\}.
\end{equation*}The
$\alpha$ dimensional Hausdorff measure of
$E$ is equal to
\begin{equation*} H^\alpha(E) = \lim_{\delta \to 0} H^\alpha_\delta(E). \end{equation*}The Hausdorff dimension
$\operatorname{dim}_{\mathbb{H}}(E)$ of a Borel set
$E$ is then the supremum over all
$s \in [0,d]$ such that
$H^s(E) = \infty$, or equivalently, the infimum over all
$s \in [0,d]$ with
$H^s(E) = 0$.For a measurable set
$E \subset \operatorname{\mathbb{T}}^d$, we let
$|E|$ denote its Lebesgue measure. We define the lower Minkowski dimension of a compact Borel set
$E \subset \operatorname{\mathbb{T}}^d$ as
\begin{equation*} \operatorname{\underline{\dim}_{\mathbb{M}}}(E) = \liminf_{r \to 0} d - \log_r|N(E,r)|. \end{equation*}Thus
$\operatorname{\underline{\dim}_{\mathbb{M}}}(E)$ is the largest number such that for
$\alpha \lt \operatorname{\underline{\dim}_{\mathbb{M}}}(E)$, there exists a decreasing sequence
$\{r_i \}$ with
$\lim_{i \to \infty} r_i = 0$ and
$|N(E,r_i)| \leq r_i^{d - \alpha}$ for each
$i$.• At several points in this paper we will need to employ probabilistic concentration bounds. In particular, we use McDiarmid’s inequality. Let
$S$ be a set, let
$\{X_1, \dots, X_N \}$ be an independent family of
$S$-valued random variables, and consider a function
$f: S^N \to \operatorname{\mathbb{C}}$. Suppose that for each
$i \in \{1, \dots, N \}$, there exists a constant
$A_i \gt 0$ such that for any
$x_1, \dots, x_{i-1}, x_{i+1}, \dots, x_N \in S$, and for each
$x_i, x_i' \in S$,
\begin{equation*} |f(x_1, \dots, x_i, \dots, x_N) - f(x_1, \dots, x_i', \dots, x_N)| \leq A_i. \end{equation*}Then McDiarmid’s inequality guarantees that for all
$t \geq 0$,
\begin{equation*} \operatorname{\mathbb{P}} \left( |f(X_1, \dots, X_N) - \operatorname{\mathbb{E}}(f(X_1, \dots, X_N))| \geq t \right) \leq 4 \exp \left( \frac{-2t^2}{A_1^2 + \dots + A_N^2} \right). \end{equation*}Proofs of McDiarmid’s inequality for real-valued functions are given in many probability texts, for instance, in Theorem 3.11 of [Reference van Handel15], but can be trivially extended to the complex-valued case by taking a union bound to the inequality for real and imaginary values of
$f$.A special case of McDiarmid’s inequality is Hoeffding’s Inequality. Hoeffding’s inequality is often stated in slightly different ways depending on the context; In this paper we use the following formulation: if
$\{X_1, \dots, X_N \}$ is a family of independent random variables, such that for each
$i$, there exists a constant
$A_i \geq 0$ such that
$|X_i| \leq A_i$ almost surely, then for each
$t \geq 0$,
\begin{align*} & \operatorname{\mathbb{P}} \left( |(X_1 + \dots + X_N) - \operatorname{\mathbb{E}}(X_1 + \dots + X_N)| \geq t \right) \\ &\quad\leq 4 \exp \left(\frac{-t^2}{2(A_1^2 + \dots + A_N^2)} \right). \end{align*}
3. Applications of our results
3.1. Arithmetic patterns
An important problem in current research on pattern avoidance is to construct sets
$E$ which avoid linear patterns, i.e. sets
$E$ which avoid solutions to equations of the form
for distinct points
$x_1,\dots,x_n \in E$. This is one scenario in which we know upper bounds on the Fourier dimension of pattern avoiding sets. It is simple to prove that if
$E \subset \operatorname{\mathbb{T}}^d$,
$\operatorname{dim}_{\mathbb{F}}(E) \gt 2d/n$, and
$m_1,\dots,m_n$ are non-zero integers, then
$m_1 E + \dots + m_n E$ is an open subset of
$\operatorname{\mathbb{T}}^d$. This is obtained by a simple modification of the argument of [Reference Mattila12, Proposition 3.14]. Thus there exists some choice of integers
$m_1,\dots,m_n$ and distinct points
$x_1,\dots,x_n \in E$ such that
$m_1x_1 + \dots + m_nx_n = 0$. Recently, under the same assumptions, Liang and Pramanik [Reference Liang and Pramanik10] have shown that for
$d = 1$, one can choose these integers
$m_1,\dots,m_n$ to satisfy
$m_1 + \dots + m_n = 0$. These results drastically contrasts the Hausdorff dimension setting, where there exists sets
$E \subset \operatorname{\mathbb{T}}^d$ with
$\operatorname{dim}_{\mathbb{H}}(E) = d$ which are linearly independent over the rational numbers, and thus avoid nontrivial solutions to integer equations of an arbitrary size (see [Reference Keleti9] for a discussion of the case where
$d = 1$, whose proof can be adapted to higher dimensions).
If
$\operatorname{dim}_{\mathbb{F}}(E) \gt d/n$, and
$m_1,\dots,m_n \neq 0$, then the set
$m_1 E + \dots + m_n E$ has positive Lebesgue measure [Reference Mattila12, Proposition 3.14]. This does not necessarily mean that
$E$ will contain solutions to the equation
$m_1x_1 + \dots + m_nx_n = 0$, but indicates why it might be difficult to push past the current Fourier dimension estimates obtained in (1.2), which construct sets with Fourier dimension
$d/n$ avoiding such patterns. The first success in pushing past this barrier was the main result of [Reference Körner8], which showed that for each
$n \gt 0$, there exists a set
$E \subset \operatorname{\mathbb{T}}$ with Fourier dimension
$1/(n-1)$ such that for any integers
$m_1,\dots,m_n \in \operatorname{\mathbb{Z}}$, not all zero, and any distinct
$x_1,\dots,x_n \in \operatorname{\mathbb{T}}^d$,
$m_1x_1 + \dots + m_nx_n \neq 0$. The technique used to control Fourier decay in that paper (bounding first derivatives of distribution functions associated with the construction of a random family of pattern avoiding sets) relies heavily on the one dimensional nature of the problem, which makes it difficult to generalize the proof technique to higher dimensions. The results of this paper imply a more robust
$d$-dimensional generalization of the result of [Reference Körner8].
Theorem 3.1. Suppose
$F \subset \operatorname{\mathbb{T}}^d$ is the countable union of a family of compact sets, each with lower Minkowski dimension
$\alpha$. Then there exists a Salem set
$E \subset \operatorname{\mathbb{T}}^d$ of dimension
$(d - \alpha)/(n-1)$ such that for any
$x \in F$, any distinct
$x_1,\dots,x_n \in E$, and any integers
$m_1,\dots,m_n \in \operatorname{\mathbb{Z}}$,
$m_1x_1 + \dots + m_nx_n \neq x$. Moreover, if
$\beta_0 = d/(n-1)$, then for any
$\beta \leq \beta_0$, and for a generic set
$(E,\mu) \in \mathcal{X}_\beta$, the set
$E$ has this property.
Proof. It will suffice to show that a generic set
$(E,\mu) \in \mathcal{X}_\beta$ is Salem and avoids solutions to equations of the form
with
$a_2,\dots,a_n \in \operatorname{\mathbb{Q}}$,
$x \in F$, and where either
$a_{n-1} \neq 0$, or
$a_2 = a_3 = \dots = a_n = 0$. Without loss of generality, we may assume
$F$ is compact and has lower Minkowski dimension
$\alpha$. If
$a_{n-1} \neq 0$, then Theorem 1.4 applies directly to the equation
where
$f(x_1,\dots,x_{n-2}) = a_1x_1 + \dots + a_{n-2}x_{n-2}$. Applying Theorem 1.4, we conclude that the set of
$(E,\mu) \in \mathcal{X}_\beta$ such that
$E$ is Salem and avoids solutions to (3.3) is comeager. On the other hand, if
$a_2 = a_3 = \dots = a_n = 0$, then the equation we must avoid is precisely
and it follows from Theorem 1.2 with
$Z = S$ and
$n = 1$ that the set of
$(E,\mu) \in \mathcal{X}_\beta$ such that
$E$ is Salem and avoids solutions to (3.4) is comeager. Taking countable unions ranging over the choices of coefficients
$a_2,\dots,a_n$ shows that the set of all
$(E,\mu) \in \mathcal{X}_\beta$ such that
$E$ is Salem and avoids all
$n$-variable linear equations is comeager, which completes the proof.
Remark 3.2. For particular linear patterns, it is certainly possible to improve the result of Theorem 3.1. For instance, Shmerkin [Reference Shmerkin13] constructed a set
$E \subset \operatorname{\mathbb{T}}$ with
$\operatorname{dim}_{\mathbb{F}}(E) = 1$ which contains no three numbers forming an arithmetic progressions, i.e. solutions to the linear equation
This equation can also be written as
Liang and Pramanik [Reference Liang and Pramanik10] generalized this technique by constructing, for any finite family of translation-invariant linear functions
$\{f_i \}$, a set
$E \subset \operatorname{\mathbb{T}}$ with
$\operatorname{dim}_{\mathbb{F}}(E) = 1$ such that for distinct
$x_1,\dots,x_n \in E$, and any index
$i$,
$f_i(x_1,\dots,x_n) \neq 0$. This same paper even constructs a set with Fourier dimension close to one avoiding an uncountable family of translation-invariant linear functions, though only those that are of a very special form. The advantage of Theorem 3.1 is that it applies to a very general family of uncountably many linear equations, though one does not obtain as high a Fourier dimension bound as those obtained in [Reference Liang and Pramanik10] and [Reference Shmerkin13]. Nonetheless, though we construct sets of dimension
$d/(n-1)$, we still remain quite far away from the best known upper bound
$2d/n$ of the Fourier dimension of a set avoiding general integer linear equations.
The arguments in this paper are heavily inspired by the techniques of [Reference Körner8], but augmented with some more robust probabilistic concentration inequalities and oscillatory integral techniques, which enables us to push the results of [Reference Körner8] to a much more general family of patterns. In particular, Theorem 1.4 shows that the results of that paper do not depend on the rich arithmetic structure of the equation
$m_1x_1 + \dots + m_nx_n = 0$, but rather only on a very weak translation invariance property of the pattern. We are unable to close the gap between the upper bound
$2d/n$ of sets avoiding
$n$-variable linear equations for
$n \geq 3$, which would seem to require utilizing the full linear nature of the equations involved much more heavily than the very weak linearity assumption that Theorem 3.1 requires.
3.2. Isosceles triangles on curves, and other nonlinear patterns
Theorems 1.2, 1.3, and 1.4 can be applied to find sets avoiding linear patterns, but the main power of these results that they can be applied to ‘nonlinear’ patterns which cannot be analysed quite as easily via the Fourier transform. As a result, relatively few results exist showing that sets with large Fourier dimension avoid patterns, though some partial results are given in [Reference Henriot, Łaba and Pramanik6] for a slightly different regime than that considered here; we do not even know if sets
$E \subset \operatorname{\mathbb{T}}$ with Fourier dimension one contain one of the simplest nonlinear patterns
$\{x, y, z, z + (y - x)^2 \}$, and results like that of [Reference Shmerkin13] show that even in the linear setting it is difficult to conjecture what might be optimal in this setting. However, Theorem 1.4 applies to this pattern, since the pattern is specified by the equation
$z - x = (y - w)^2$, obtaining a Salem set
$E \subset \operatorname{\mathbb{T}}$ of dimension
$1/3$ avoiding this pattern, the first proof of the existence of a set with positive Fourier dimension avoiding isosceles triangles in the literature.
In this section, we consider a more geometric example of a non-linear pattern, finding subsets of curves which do not contain isosceles triangles; given a simple segment of a curve given by a smooth map
$\gamma : [0,1] \to \operatorname{\mathbb{R}}^d$, we say a set
$E \subset [0,1]$ avoids isosceles triangles on
$\gamma$ if for any distinct values
$t_1,t_2,t_3 \in [0,1]$,
$|\gamma(t_1) - \gamma(t_2)| \neq |\gamma(t_2) - \gamma(t_3)|$, i.e. if
$\gamma(E)$ does not contain any three points forming the vertices of an isosceles triangle. In [Reference Fraser and Pramanik5], methods are provided to construct sets
$E \subset [0,1]$ with
$\operatorname{dim}_{\mathbb{H}}(E) = \log_3 2 \approx 0.63$ such that
$\gamma(E)$ does not contain any isosceles triangles, but
$E$ is not guaranteed to be Salem. Using the results of this paper, we can now construct Salem sets
$E \subset [0,1]$ with
$\operatorname{dim}_{\mathbb{F}}(E) = 4/9 \approx 0.44$, such that
$\gamma(E)$ does not contain any isosceles triangles.
Theorem 3.3. For any smooth map
$\gamma: [0,1] \to \operatorname{\mathbb{R}}^d$ with
$\gamma'(t) \neq 0$ for all
$t \in [0,1]$, there exists a Salem set
$E \subset [0,1]$ with
$\operatorname{dim}_{\mathbb{F}}(E) = 4/9$ which avoids isosceles triangles on
$\gamma$.
Proof. Assume without loss of generality (working on a smaller portion of the curve if necessary and then rescaling) that there exists a constant
$C \geq 1$ such that for any
$t,s \in [0,1]$,
and
Let
$\varepsilon = 1/2C^3$, and let
A simple calculation using (3.5) and (3.6) reveals that for
$0 \leq t_1,t_2 \leq \varepsilon$,
\begin{equation}
\left| \frac{\partial F}{\partial t_1} \right| = 2 \left| (\gamma(t_1) - \gamma(t_2)) \cdot \gamma'(t_1) \right| \geq (2/C) |t_2 - t_1| - 2C |t_2 - t_1|^2 \geq (1/C) |t_2 - t_1|.
\end{equation} This means that
$\partial F / \partial t_1 \neq 0$ unless
$t_1 = t_2$. Thus the implicit function theorem implies that there exists a countable family of smooth functions
$\{f_i: U_i \to [0,1] \}$, where
$U_i \subset [0,\varepsilon]^2$ for each
$i$ and
$f_i(t_2,t_3) \neq t_3$ for any
$(t_2,t_3) \in U_i$, such that if
$F(t_1,t_2,t_3) = 0$ for distinct points
$t_1,t_2,t_3 \in [0,\varepsilon]$, then there exists an index
$i$ with
$(t_2,t_3) \in U_i$ and
$t_1 = f_i(t_2,t_3)$.
Differentiating both sides of the equation
in
$t_2$ and
$t_3$ shows that
\begin{equation}
\frac{\partial f_i}{\partial t_2}(t_2,t_3) = \frac{(\gamma(f_i(t_2,t_3)) - \gamma(t_3)) \cdot \gamma'(t_2)}{(\gamma(f_i(t_2,t_3)) - \gamma(t_2)) \cdot \gamma'(f_i(t_2,t_3))},
\end{equation}and
\begin{equation}
\frac{\partial f_i}{\partial t_3}(t_2,t_3) = \frac{- (\gamma(t_2) - \gamma(t_3)) \cdot \gamma'(t_3)}{(\gamma(f_i(t_2,t_3)) - \gamma(t_2)) \cdot \gamma'(f_i(t_2,t_3))}.
\end{equation} In order to apply Theorem 1.3, we must show that the partial derivatives in 3.11 and 3.12 are both non-vanishing for
$t_2,t_3 \in [0,\varepsilon]$. We calculate using (3.5), (3.6) and (3.7) that
\begin{align}
\begin{split}
|(\gamma(f_i(t_2,t_3)) - \gamma(t_3)) \cdot \gamma'(t_2)| &\geq |(\gamma(f_i(t_2,t_3)) - \gamma(t_3)) \cdot \gamma'(t_3)|\\
&\quad\quad + |(\gamma(f_i(t_2,t_3)) - \gamma(t_3)) \cdot (\gamma'(t_2) - \gamma'(t_3))|\\
&\geq (1/C) |f_i(t_2,t_3) - t_3| - C^2 |f_i(t_2,t_3) - t_3||t_2 - t_3|\\
&\geq (1/C - C^2 \varepsilon) |f_i(t_2,t_3) - t_3|\\
&\geq (1/2) |f_i(t_2,t_3) - t_3|.
\end{split}
\end{align} Since
$f_i(t_2,t_3) \neq t_3$ for all
$(t_2,t_3) \in U_i$, it follows from (3.11) and (3.13) that if
$(t_2,t_3) \in U_i$ with
$t_2 \neq t_3$,
\begin{equation}
\frac{\partial f_i}{\partial t_2}(t_2,t_3) \neq 0.
\end{equation} A similar calculation to (3.9) shows that for
$t_2,t_3 \in [0,\varepsilon]$,
Combining (3.12) with (3.15) shows that for
$t_2 \neq t_3$ with
$(t_2,t_3) \in U_i$,
\begin{equation}
\frac{\partial f_i}{\partial t_3}(t_2,t_3) \neq 0,
\end{equation}verifying the partial derivatives are non-vanishing.
Now (3.14) and (3.16) imply that each function in the family
$\{f_i \}$ satisfy the hypothesis of Theorem 1.3. Thus that theorem implies that for
$\beta = 4/9$, each index
$i$, and a generic element of
$(E,\mu) \in \mathcal{X}_\beta$, the set
$E$ is Salem and for any distinct
$t_1,t_2,t_3 \in E \cap [0,\varepsilon]$,
$f_i(t_1,t_2,t_3) \neq 0$. This means precisely that
$|\gamma(t_1) - \gamma(t_2)| \neq |\gamma(t_2) - \gamma(t_3)|$ for any distinct
$t_1,t_2,t_3 \in E$. Thus we conclude we can find a Salem set
$E \subset [0,\varepsilon]$ with
$\operatorname{dim}_{\mathbb{F}}(E) = 4/9$ such that
$\gamma(E)$ does not contain the vertices of any isosceles triangles.
Theorem 1.2 can also be used to construct sets with a slightly smaller dimension avoiding isosceles triangles on a rougher family of curves. If we consider a Lipschitz function
$\gamma: [0,1] \to \operatorname{\mathbb{R}}^{d-1}$, where there exists
$M \lt 1$ with
$|\gamma(t) - \gamma(s)| \leq M |t - s|$ for each
$t,s \in [0,1]$, then Theorem 3 of [Reference Denson, Pramanik and Zahl3] guarantees that the set
\begin{equation*} Z = \left\{(x_1,x_2,x_3) \in [0,1]^3 : \begin{array}{c}
\text{$(x_1,\gamma(x_1)), (x_2,\gamma(x_2)), (x_3,\gamma(x_3))$}\\
\text{form the vertices of an isosceles triangle.} \end{array} \right\} \end{equation*} has lower Minkowski dimension at most two. Thus Theorem 1.2 guarantees that there exists a Salem set
$E \subset [0,1]$ with
$\operatorname{dim}_{\mathbb{F}}(E) = 2/5 = 0.4$ such that
$\gamma(E)$ avoids all isosceles triangles. The main result of [Reference Fraser and Pramanik5] constructs a set
$E \subset [0,1]$ with
$\operatorname{dim}_{\mathbb{H}}(E) = 0.5$ such that
$\gamma(E)$ avoids all isosceles triangles, but this set is not guaranteed to be Salem.
4. A metric space controlling Fourier dimension
In order to work with a Baire category type argument, we must construct an appropriate metric space appropriate for our task, and establish a set of tools for obtaining convergence in this metric space. In later sections we will fix a specific choice of
$\beta$ to avoid a particular pattern. But in this section we let
$\beta$ be an arbitrary fixed number in
$(0,d]$. Our approach in this section is heavily influenced by [Reference Körner8]. However, we employ a Fréchet space construction instead of the Banach space construction used in [Reference Körner8], which enables us to use softer estimates in our arguments, with the disadvantage that we can obtain only Fourier dimension bounds in Theorems 1.2, 1.3, and 1.4 at the endpoint dimensions
$\beta_0$ considered in the theorems, rather than the explicit decay estimates as is obtained, for instance, in Theorem 2.4 of [Reference Körner8]:
• We let
$\mathcal{E}$ denote the family of all compact subsets of
$\operatorname{\mathbb{T}}^d$. If we consider the Hausdorff distance
$d_{\mathbb{H}}$ between sets, then
$(\mathcal{E},d_{\mathbb{H}})$ forms a complete metric space.• We let
$M_*(\beta)$ consist of all finite Borel measures
$\mu$ on
$\operatorname{\mathbb{T}}^d$ such that for each
$\lambda \in [0,\beta)$,
\begin{equation*} \| \mu \|_{M(\lambda)} = \sup_{\xi \in \operatorname{\mathbb{Z}}^d} |\widehat{\mu}(\xi)| |\xi|^{\lambda/2} \end{equation*}is finite. Then
$\| \cdot \|_{M(\lambda)}$ is a seminorm on
$M_*(\beta)$ for each
$\lambda \in [0,\beta)$, and the collection of all such seminorms gives
$M_*(\beta)$ the structure of a Frechét space (most importantly, this means
$M_*(\beta)$ is a complete metric space). Under this topology, a sequence of probability measures
$\{\mu_k \}$ converges to a probability measure
$\mu$ in
$M_*(\beta)$ if and only if for any
$\lambda \in [0,\beta)$,
$\lim_{k \to \infty} \| \mu_k - \mu \|_{M(\lambda)} = 0$.
We now let
$\mathcal{X}_\beta$ be the collection of all pairs
$(E,\mu) \in \mathcal{E} \times M_*(\beta)$, where
$\mu$ is a probability measure such that
$\text{supp}(\mu) \subset E$. Then
$\mathcal{X}_\beta$ is a closed subset of
$\mathcal{E} \times M_*(\beta)$ under the product metric, and thus a complete metrizable space. We remark that for any
$\lambda \in [0,\beta)$ and
$(E,\mu) \in \mathcal{X}_\beta$,
\begin{equation}
\lim_{|\xi| \to \infty} |\widehat{\mu}(\xi)| |\xi|^{\lambda/2} = 0.
\end{equation} Thus
$\operatorname{dim}_{\mathbb{F}}(E) \geq \operatorname{dim}_{\mathbb{F}}(\mu) \geq \beta$ for each
$(E,\mu) \in \mathcal{X}_\beta$. This means that
$\mathcal{X}_\beta$ can be thought of as a space of compact sets, augmented with a measure providing a certification guaranteeing the set’s Fourier dimension is at least
$\beta$.
Lemma 4.1 allows us to reduce the proof of density arguments in
$\mathcal{X}_\beta$ to the construction of large discrete subsets in
$\operatorname{\mathbb{T}}^d$ with well-behaved Fourier analytic properties. We recall that a family
$\mathcal{A}$ of subsets of
$\operatorname{\mathbb{T}}^d$ is downward closed if, whenever
$E \subset \mathcal{A}$, any subset of
$E$ is also contained in
$\mathcal{A}$.
Lemma 4.1. Let
$\mathcal{A}$ be a downward closed family of subsets of
$\operatorname{\mathbb{T}}^d$. Fix
$\beta \gt 0$,
$\kappa \gt 0$, and a large constant
$C \gt 0$. Suppose that for all small
$\delta \gt 0$, and all
$\lambda \in [0,\beta)$, there are arbitrarily large integers
$N \gt 0$ for which there exists a finite set
$S = \{x_1, \dots, x_N \}$, positive numbers
$\{a_1, \dots, a_N \}$ such that
$\sum_{k = 1}^N a_k = N$, and a quantity
$r \geq (2N)^{-1/\lambda}$ such that the following properties hold:
(1)
$N(S,r) \in \mathcal{A}$.(2) For each
$\xi \in \operatorname{\mathbb{Z}}^d - \{0 \}$ with
$|\xi| \leq (1/r)^{1 + \kappa}$,
\begin{equation*} \left| \frac{1}{N} \sum_{k = 1}^N a_k e^{2 \pi i \xi \cdot x_k} \right| \leq C N^{-1/2} \log(N) + \delta |\xi|^{-\lambda/2} \end{equation*}
Then
$\{(E,\mu) \in \mathcal{X}_\beta : E \in \mathcal{A} \}$ is dense in
$\mathcal{X}_\beta$.
Because the proof of Lemma 4.1 is somewhat technical, we relegate it to an appendix found at the end of this paper. In the remainder of this section, we use Lemma 4.1, together with a probabilistic argument to argue some general results about
$\mathcal{X}_\beta$ which will be useful in the proofs of Theorems 1.2, 1.3, and 1.4.
Remark 4.2. We will be able to take
$\delta = 0$ for the applications of Lemma 4.1 in Theorems 1.2 and 1.4 by applying purely probabilistic arguments which give rise to square root cancellation in the number of random variables involved. We only need to take
$\delta \gt 0$ when applying this result to Theorem 1.3, because we must apply some oscillatory integral bounds which give an additional decaying factor as
$\xi \to \infty$.
Let us comment on the intuition underlying Lemma 4.1. Consider a large integer
$N$, and suppose there is a discrete family of
$N$ points
$S = \{x_1, \dots, x_N \}$ such that
$N(S,r)$ does not contain any incidences of a particular pattern. Then
$N(S,r)$ is a union of
$N$ balls of radius
$r$, so if
$N \approx r^{-\beta}$, and these balls do not overlap too much, we might expect
$N(S,r)$ to behave like an
$r$-thickening of a
$\beta$-dimensional set. The Fourier analytic properties of
$S$ can be understood by taking exponential sums, i.e. considering quantities of the form
\begin{equation*} \frac{1}{N} \sum_{k = 1}^N a_k e^{2 \pi i \xi \cdot x_k}. \end{equation*} where
$a_1,\dots,a_N$ are non-negative and sum to one as in Lemma 4.1. For any set
$S$, taking in absolute values gives a trivial bound on the exponential sum
\begin{equation}
\left| \frac{1}{N} \sum_{k = 1}^N a(x_k) e^{2 \pi i \xi \cdot x_k} \right| \leq 1,
\end{equation} and this bound can be tight for general sets
$S$, for instance, if
$S$ behaves like an arithmetic progression with a frequency
$\xi$, as happens for
\begin{equation*} S = \left\{k \frac{\xi}{|\xi|^2} : 1 \leq k \leq N \right\}. \end{equation*} If one can significantly improve upon this bound, one therefore thinks of
$S$ as having additional regularity from the perspective of Fourier analysis. The best case we can hope to hold for a ‘generic’ choice of
$S$ is a square root cancellation bound of the form
\begin{equation}
\left| \frac{1}{N} \sum_{k = 1}^N a_k e^{2 \pi i \xi \cdot x_k} \right| \leq C N^{-1/2},
\end{equation} which, roughly speaking, holds if the points
$\{x_k \}$ are not significantly periodic at the frequency
$\xi$. If
$\kappa \gt 0$ is fixed, and equation (4.3) holds for all
$|\xi| \lesssim (1/r)^{1 + \kappa}$, then one might therefore think of
$N(S,r)$ as behaving like an
$r$-thickening of a Salem set with dimension
$\beta$. Note that, up to a logarithmic constant, and a negligible term which decays as
$|\xi| \to \infty$, Lemma 4.1 obtains such a square root cancellation bound. Since the assumptions of Lemma 4.1 guarantee that we can construct such sets at arbitrarily small scales
$r$, it makes sense that one should be able to use such assumptions to construct Salem sets avoiding patterns.
Lemma 4.1 not only guarantees that we can construct Salem sets avoiding patterns, but also guarantees that such pattern avoiding sets are generic among all Salem sets. The reason such a result is possible given the assumptions of Lemma 4.1 is that the points in the discrete set
$S$ are not significantly periodic at any frequency
$\xi$, which heuristically (à la Weyl’s equidistribution theorem) implies such points are evenly distributed in
$\operatorname{\mathbb{T}}^d$. We do not need equidistribution here, but we are able to obtain in Lemma A.4 that for any
$\varepsilon \gt 0$, if
$N$ is taken appropriately large, then the set
$S$ guaranteed by the assumptions of Lemma 4.1 is
$\varepsilon$-dense in
$\operatorname{\mathbb{T}}^d$, i.e.
$\operatorname{\mathbb{T}}^d = N(S,\varepsilon)$. This will allow us to approximate an arbitrary compact set of dimension
$\beta$ by subsets of
$S$, and thus approximate an arbitrary element of
$\mathcal{X}_\beta$ by an element which is pattern avoiding, thus guaranteeing that sets containing patterns are of first category.
Remark 4.3. In each of Theorems 1.2, 1.3, and 1.4 we will prove that quasi-all sets in
$\mathcal{X}_\beta$ avoid a particular pattern by constructing a countable family
$\{\mathcal{A}_k : k \geq 1 \}$ of downward closed sets, each of which being open in
$\mathcal{X}_\beta$, such that all sets in
$\bigcap \mathcal{A}_k$ avoid a particular pattern. By verifying that the hypotheses of Lemma 4.1 hold for each set
$\mathcal{A}_k$, we obtain density of
$\mathcal{A}_k$ in
$\mathcal{X}_\beta$, so that their intersection is comeager in
$\mathcal{X}_\beta$.
Each such argument can be converted into an explicit ‘queuing construction’ of a pattern avoiding set via a method we very briefly outline. In particular, one can obtain a nested family of sets
$\{E_k \}$, and a family of non-negative measures
$\{\mu_k \}$, with
$\mu_k$ supported on
$E_k$ and with
$\mu_k(\operatorname{\mathbb{T}}^d)$ uniformly bounded from below in
$k$, so that
$E = \bigcap E_k$ is pattern avoiding, has Hausdorff dimension at most
$\beta$, and any weak limit of the measures (which necessarily must be supported on
$E$) has Fourier decay
$|\xi|^{\varepsilon -\beta/2}$ for all
$\varepsilon \gt 0$, which will justify that
$E$ is Salem of dimension
$\beta$.
To obtain
$\{E_k \}$ and
$\{\mu_k \}$, we choose
$\{\lambda_k \}$ converging rapidly to
$\beta$ from below,
$\{\delta_k \}$ and
$\{\kappa_k \}$ converging rapidly to zero, and
$\{N_k \}$ converging rapidly to infinity. We set
$E_0 = \operatorname{\mathbb{T}}^d$, and let
$\mu_0 = \mathbb{I}_{\operatorname{\mathbb{T}}^d}$. For
$k \gt 0$, we use the hypotheses of Lemma 4.1 to find a set
$S_k$, coefficients
$a_k: S_k \to [0,\infty)$, such that if
$r_k = (2N_k)^{-1/\lambda_k}$, then
$N(S_k,r_k) \in \mathcal{A}_k$ and the exponential sums
$\sum a_k(x) e^{2 \pi i \xi \cdot x}$ satisfy the square root cancellation bounds above for
$|\xi| \leq (1/r_k)^{1 + \kappa_k}$. We then define
$E_k = E_{k-1} \cap N(S_k,r_k)$ and let
$\mu_k = (A * \phi_r) \mu_{k-1}$, where
$A = \sum_{x \in S} a(x) \delta_x$, and
$\phi_r$ is a smooth bump function supported in a
$r$ neighborhood of the origin. If
$\delta_k$ and
$N_k$ are sufficiently small, then a Plancherel argument à la the methods of Lemma A.4 can be used to guarantee that
$\mu_k(\operatorname{\mathbb{T}}^d) \geq (1 - 2^{-k}) \mu_{k-1}(\operatorname{\mathbb{T}}^d)$. And one can also recursively justify that there is
$C_\varepsilon \gt 0$ for each
$\varepsilon \gt 0$ so that, uniformly in
$k$,
$|\widehat{\mu}_k(\xi)| \leq C_\varepsilon |\xi|^{\varepsilon-\beta/2}$, so that any weak limit
$\mu$ of the measures
$\mu_k$ satisfies
$|\widehat{\mu}(\xi)| \leq C_\varepsilon |\xi|^{\varepsilon-\beta/2}$ for each
$\varepsilon \gt 0$. Since one also has
$\mu(\operatorname{\mathbb{T}}^d) \geq \prod (1 - 2^{-k}) \geq e^{-2}$ this proves
$E$ has Fourier dimension at least
$\beta$. Moreover,
$E$ is covered by
$N_k$ balls of radius
$(2N_k)^{-1/\lambda_k}$ for each
$k$, so that
$E$ has lower Minkowski dimension at most
$\beta$, and thus has Hausdorff dimension at most
$\beta$, which justifies that
$E$ is a Salem set of dimension
$\beta$.
It is a useful heuristic that in a metric space whose elements are sets, and with distance defined in terms of the Hausdorff distance, quasi-all elements are as ‘thin as possible’ (by quasi-all we mean in a Baire category sense, i.e. the family of all sets which fail to be as thin as possible is contained in a countable union of closed sets with non-empty interior). In particular, we should expect the Hausdorff dimension and Fourier dimension of a generic element of
$\mathcal{X}_\beta$ to be as low as possible. For each
$(E,\mu) \in \mathcal{X}_\beta$, the condition that
$\mu \in M_*(\beta)$ implies that
$\operatorname{dim}_{\mathbb{F}}(\mu) \geq \beta$, so
$\operatorname{dim}_{\mathbb{H}}(E) \geq \operatorname{dim}_{\mathbb{F}}(E) \geq \beta$. Since the Fourier dimension and Hausdorff dimension are lower bounded by
$\beta$, and this bound is tight, our heuristic thus leads us to believe that for quasi-all
$(E,\mu) \in M_*(\beta)$, the set
$E$ has both Hausdorff dimension and Fourier dimension equal to
$\beta$, i.e.
$E$ is a Salem set of dimension
$\beta$. We will finish this section with a proof of this fact. This will also give some more elementary variants of the kinds of probabilistic arguments we will later use to prove Theorems 1.2, 1.3, and 1.4, which will allow us to become more comfortable with these techniques in preparation for the proofs of these theorems.
Lemma 4.4. Fix a positive integer
$N$, and
$\kappa \gt 0$. Let
$X_1, \dots, X_N$ be independent random variables on
$\operatorname{\mathbb{T}}^d$, such that for each
$\xi \in \operatorname{\mathbb{Z}}^d - \{0 \}$,
\begin{equation}
\sum_{k = 1}^N \operatorname{\mathbb{E}} \left( e^{2 \pi i \xi \cdot X_k} \right) = 0.
\end{equation} Then there exists a constant
$C$ depending on
$d$ and
$\kappa$ such that
\begin{equation*} \operatorname{\mathbb{P}} \left( \sup_{|\xi| \leq N^{1 + \kappa}} \left| \frac{1}{N} \sum_{k = 1}^N e^{2 \pi i \xi \cdot X_k} \right| \geq C N^{-1/2} \log(N)^{1/2} \right) \leq 1/10. \end{equation*}Remark 4.5. In particular, the assumptions of Lemma 4.4 hold if the random variables
$\{X_1,\dots, X_N \}$ are uniformly distributed on
$\operatorname{\mathbb{T}}^d$, since then
$\operatorname{\mathbb{E}}(e^{2 \pi i \xi \cdot X_k}) = 0$ for all
$\xi \in \operatorname{\mathbb{Z}}^d - \{0 \}$ and
$1 \leq k \leq N$.
Proof. For each
$\xi \in \operatorname{\mathbb{Z}}^d$ and
$k \in \{1, \dots, N \}$, consider the random variable
Then for each
$\xi \in \operatorname{\mathbb{Z}}^d$,
\begin{equation}
\sum_{k = 1}^N Y(\xi,k) = \frac{1}{N} \sum_{k = 1}^N e^{2 \pi i \xi \cdot X_k}.
\end{equation} We also note that for each
$\xi \in \operatorname{\mathbb{Z}}^d$ and
$k \in \{1, \dots, N \}$,
Moreover,
\begin{equation}
\begin{split}
\sum_{k = 1}^N \operatorname{\mathbb{E}}(Y(\xi,k)) = 0.
\end{split}
\end{equation} Since the family of random variables
$\{Y(\xi,k) \}$ is independent for a fixed
$\xi$, we can apply Hoeffding’s inequality together with (4.5) and (4.6) to conclude that for all
$t \geq 0$,
\begin{equation}
\operatorname{\mathbb{P}} \left( \left| \frac{1}{N} \sum_{k = 1}^N e^{2 \pi i \xi \cdot X_k} \right| \geq t \right) \leq 2 e^{-Nt^2/2}.
\end{equation} Taking a union bound obtained by applying (4.8) over all
$|\xi| \leq N^{1 + \kappa}$ gives the existence of a constant
$C \geq 10$ depending on
$d$ and
$\kappa$ such that
\begin{equation}
\operatorname{\mathbb{P}} \left( \sup_{|\xi| \leq N^{1 + \kappa}} \left| \frac{1}{N} \sum_{k = 1}^N e^{2 \pi i \xi \cdot X_k} \right| \geq t \right) \leq \exp \left( C \log(N) - \frac{5N t^2}{C} \right).
\end{equation} But then setting
$t = CN^{-1/2} \log(N)^{1/2}$ in (4.9) completes the proof.
Lemma 4.6. For quasi-all
$(E,\mu) \in \mathcal{X}_\beta$,
$E$ is a Salem set of dimension
$\beta$.
Proof. We shall assume
$\beta \lt d$ in the proof, since when
$\beta = d$,
$E$ is a Salem set for any
$(E,\mu) \in \mathcal{X}_\beta$, and thus the result is trivial. Since the Hausdorff dimension of a measure is an upper bound for the Fourier dimension, it suffices to show that for quasi-all
$(E,\mu) \in \mathcal{X}_\beta$,
$E$ has Hausdorff dimension at most
$\beta$. For each
$\alpha \gt \beta$ and
$\delta, s \gt 0$, we let
\begin{equation*} \mathcal{A}(\alpha,\delta,s) = \{E \subset \operatorname{\mathbb{T}}^d: H^\alpha_\delta(E) \lt s \}. \end{equation*}and set
Then
$A(\alpha,\delta,s)$ is an open subset of
$\mathcal{X}_\beta$, and
\begin{equation}
\bigcap_{n = 1}^\infty \bigcap_{m = 1}^\infty \bigcap_{k = 1}^\infty A(\beta + 1/n, 1/m, 1/k),
\end{equation} is precisely the family of
$(E,\mu) \in \mathcal{X}_\beta$ such that
$E$ has Hausdorff dimension at most
$\beta$. Thus it suffices to show that
$A(\alpha,\delta,s)$ is dense in
$\mathcal{X}_\beta$ for all
$\alpha \gt \beta$, all
$\delta \gt 0$, and all
$s \gt 0$. Since
$\mathcal{A}(\alpha,\delta,s)$ is a downward closed family of subsets of
$\operatorname{\mathbb{T}}^d$, we may apply Lemma 4.1. Fix a large integer
$N$, and set
$r = N^{-1/\beta}$, so that
$N \geq (1/2) r^{-\lambda}$ satisfies the condition for Lemma 4.1 to apply to these quantities. Lemma 4.4 shows that there exists a constant
$C$ depending on
$\beta$ and
$d$, as well as
$N$ points
$S = \{x_1, \dots, x_N \} \subset \operatorname{\mathbb{T}}^d$ such that for each
$|\xi| \leq N^{1 + \kappa}$,
\begin{equation}
\left| \frac{1}{N} \sum_{k = 1}^N e^{2 \pi i \xi \cdot x_k} \right| \leq C N^{-1/2} \log(N)^{1/2}.
\end{equation} Now
$N(S,r)$ is a union of
$N$ balls of radius
$r$, and thus if
$r \leq \delta$,
\begin{equation}
H^\alpha_\delta(N(S,r)) \leq N r^\alpha = N^{1 - \alpha / \beta}.
\end{equation} Since
$\alpha \gt \beta$, taking
$N$ appropriately large gives a set
$N(S,r)$ with
Thus
$N(S,r) \in \mathcal{A}(\alpha,\delta,s)$ for sufficiently large integers
$N$. But together with (4.11), this justifies that the hypothesis of Lemma 4.1 applies to this scenario. Thus that lemma implies that
$A(\alpha,\delta,s)$ is dense in
$\mathcal{X}_\beta$, completing the proof.
This concludes the setup to the proof of Theorems 1.2, 1.3, and 1.4. All that remains is to show that quasi-all elements of
$\mathcal{X}_\beta$ avoid the given set
$Z$ for a suitable parameter
$\beta$; it then follows from Lemma 4.6 that quasi-all elements of
$\mathcal{X}_\beta$ are Salem and avoid the given set
$Z$ (since the intersection of two generic subsets of
$\mathcal{X}_\beta$ is also generic). The advantage of Lemma 4.1, combined with a Baire category argument, is that we can now reduce our calculations to finding suitable finite families of points with nice Fourier analytic properties.
5. Random avoiding sets for rough patterns
We begin by proving Theorem 1.2, which requires simpler calculations than Theorem 1.3 and Theorem 1.4. In the last section, our results held for an arbitrary
$\beta \in (0,d]$. But in this section, we assume
\begin{equation*} \beta \leq \min \left( d, \frac{dn - \alpha}{n - 1/2} \right), \end{equation*} where
$\alpha$ and
$n$ are as in the statement of Theorem 1.2. Then
$\beta$ is small enough to show that the pattern
$Z$ described in Theorem 1.2 is avoided by a generic element of
$\mathcal{X}_\beta$. The construction here is very similar to the construction in [Reference Denson, Pramanik and Zahl3], albeit in a Baire category setting, and with modified parameters to ensure a Fourier dimension bound rather than just a Hausdorff dimension bound.
Lemma 5.1. Let
$Z \subset \operatorname{\mathbb{T}}^{dn}$ be a compact set with lower Minkowski dimension at most
$\alpha$. Then for quasi-all
$(E,\mu) \in \mathcal{X}_\beta$, for any distinct points
$x_1, \dots, x_n \in E$,
$(x_1, \dots, x_n) \not \in Z$.
Proof. For any
$s \gt 0$, consider the set
\begin{equation*} \mathcal{B}(Z,s) = \left\{E \subset \operatorname{\mathbb{T}}^d: \begin{array}{c}
\text{for all $x_1, \dots, x_n \in E$ such that}\\
\text{$|x_i - x_j| \geq s$ for $i \neq j$, $(x_1, \dots, x_n) \not \in Z$}
\end{array} \right\}, \end{equation*}and
Then
$B(Z,s)$ is open in
$\mathcal{X}_\beta$, and
\begin{equation}
\bigcap_{k = 1}^\infty B(Z,1/k)
\end{equation} consists of the family of sets
$(E,\mu)$ such that for distinct
$x_1, \dots, x_n \in E$,
$(x_1, \dots, x_n) \not \in Z$. Now for each
$k$, the set
$\mathcal{B}(Z,1/k)$ is a downward closed family, which means that, after we verify the appropriate hypotheses, we can apply Lemma 4.1 to prove
$B(Z,s)$ is dense in
$\mathcal{X}_\beta$ for each
$s \gt 0$, which would complete the proof. Thus we must construct a set
$S = \{x_1, \dots, x_N \}$ for
$N$ which can be made arbitrarily large, such that
$N(S,r) \in \mathcal{B}(Z,s)$ and associate with the set an exponential sum satisfying a square root cancellation bound.
Let
$c = 2 n^{1/2}$. Since
$Z$ has lower Minkowski dimension at most
$\alpha$, for any
$\gamma \in (\alpha,dn]$, we can find arbitrarily small
$r \in (0,1)$ such that
Pick
$\lambda \in [0,(dn - \gamma)/(n-1/2))$, and suppose that we can find an integer
$M \geq 10$ with
Let
$X_1, \dots, X_M$ be independent and uniformly distributed on
$\operatorname{\mathbb{T}}^d$. For each distinct set of indices
$k_1, \dots, k_n \in \{1, \dots, M \}$, the random vector
$X_k = (X_{k_1}, \dots, X_{k_n})$ is uniformly distributed on
$\operatorname{\mathbb{T}}^{nd}$, and so (5.2) and (5.3) imply that
\begin{equation}
\operatorname{\mathbb{P}}(d(X_k,Z) \leq cr) \leq |N(Z,cr)| \leq r^{dn - \gamma} \lesssim M^{- \frac{dn - \gamma}{\lambda}} \leq M^{-(n-1/2)},
\end{equation} If
$M_0$ denotes the number of indices
$k$ such that
$d(X_k,Z) \leq cr$, then by linearity of expectation, since there are at most
$M^n$ such indices, we conclude from (5.4) that there is a constant
$C \gt 0$ such that
\begin{equation}
\operatorname{\mathbb{E}}(M_0) \leq (C/10) M^{1/2}.
\end{equation}Applying Markov’s inequality to (5.5), we conclude that
\begin{equation}
\operatorname{\mathbb{P}}(M_0 \geq C M^{1/2}) \leq 1/10.
\end{equation} Fix some small
$\kappa \gt 0$. Taking a union bound to (5.6) and the results of Lemma 4.4, we conclude that if
$M$ is sufficiently large, there exists
$M$ distinct points
$x_1, \dots, x_M \in \operatorname{\mathbb{T}}^d$ and a constant
$C \gt 0$ such that the following two statements hold:
(1) Let
$I$ be the set of indices
$k_n \in \{1, \dots, M \}$ with the property that we can find distinct indices
$k_1, \dots, k_{n-1} \in \{1, \dots, M \}$ such that if
$X = (X_{k_1}, \dots, X_{k_n})$, then
$d(X,Z) \leq cr$. Then
$\#(I) \leq C M^{1/2}$.(2) For
$0 \lt |\xi| \leq M^{1 + \kappa}$,
\begin{equation*} \left| \frac{1}{M} \sum_{k = 1}^M e^{2 \pi i \xi \cdot x_k} \right| \leq C M^{-1/2} \log(M)^{1/2}. \end{equation*}
We now use this information to construct the set
$S$ required to verify the hypotheses of Lemma 4.1.
Now set
$S = \{x_k : k \not \in I \}$ and let
$N = \#(S)$. Then Property (1) implies that
Thus for
$M \geq 4C^2$,
Property (1) and (2) imply that for
$0 \lt |\xi| \leq N^{1 + \kappa}$,
\begin{equation}
\begin{split}
\left| \frac{1}{N} \sum_{x \in S} e^{2 \pi i \xi \cdot x} \right| &\leq \left| \frac{1}{N} \sum_{k = 1}^N e^{2 \pi i \xi \cdot x_k} \right| + \left| \frac{1}{N} \sum_{k \in I} e^{2 \pi i \xi \cdot x_k} \right|\\
&\leq 2C M^{-1/2} \log(M)^{1/2} + \#(I)/N \\
&\lesssim N^{-1/2} \log(N)^{1/2} + N^{-1/2}\\
&\lesssim N^{-1/2} \log(N)^{1/2}.
\end{split}
\end{equation} As long as we can show that
$N(S,r) \in \mathcal{B}(Z,s)$, then (5.8) and (5.9) allows us to apply Lemma 4.1, completing the proof that
$B(Z,s)$ is dense. To check this, consider
$n$ points
$y_1,\dots,y_n \in N(S,r)$, with
$|y_i - y_j| \geq s$ for any two indices
$i \neq j$. Provided that
$s \gt 10r$, we can therefore find distinct indices
$k_1, \dots, k_n \not \in I$ such that for each
$i \in \{1, \dots, n \}$,
$|x_{k_i} - y_i| \leq r$, which means if we set
$x = (x_{k_1}, \dots, x_{k_n})$ and
$y = (y_1, \dots, y_n)$, then
Since
$k_n \not \in I$,
$d(x,Z) \geq cr$, which combined with (5.10) implies
Thus in particular, we conclude
$y \not \in Z$. But this means we have proved precisely that
$N(S,r) \in \mathcal{B}(Z,s)$. Thus Lemma 4.1 implies that
$B(Z,s)$ is dense in
$\mathcal{X}_\beta$ for each
$s \gt 0$, completing the proof.
The Baire category theorem, applied to the result of Lemma 5.1, shows that a pattern avoiding set exists in
$\mathcal{X}_\beta$, completing the proof of Theorem 1.2. Before we move onto the proof of Theorem 1.3, let us discuss the main obstacle which prevents us from finding Salem sets with dimension
\begin{equation}
\frac{dn - \alpha}{n - 1},
\end{equation} avoiding the pattern
$Z$, instead only obtaining Salem sets with dimension at most
\begin{equation}
\frac{dn - \alpha}{n-1/2}.
\end{equation} We begin by noting that the quantity (5.12) is the maximum dimension one can obtain using the randomized selection method used in Lemma 5.1 to choose the set
$S$, since if one alters the parameters used in the proof so that
$M \gt rsim r^{-\frac{dn-\alpha}{n-1}}$, then the expectation bounds used in the proof above to control
$M_0$ cannot even guarantee that
$S$ is non-empty with positive probability. More precisely, for general parameters
$M$ and
$r$, one can guarantee with high probability that
$\#(I) \lesssim M^n r^{dn - \alpha}$. For
$M \gg r^{-\frac{dn-\alpha}{n-1}}$, one also expects to have
$\#(I) \gg M$, and in this situation we will have
$S = \emptyset$, so the construction above does not work at all. In this proof however, we were forced to choose
$M$ much smaller than
$r^{-\frac{dn-\alpha}{n-1}}$, i.e. we chose
$M \approx r^{-\frac{dn - \alpha}{n - 1/2}}$ so that we could guarantee that
$\#(I) \lesssim N^{1/2}$. The importance of this is that the trivial bound
\begin{equation}
\left| \sum_{k \in I} e^{2 \pi i (\xi \cdot X_k)} \right| \leq \#(I),
\end{equation} obtained by the triangle inequality was then enough to obtain the square root cancellation bound in equation (5.9). On the other hand, if we were able to show that the set
$I$ itself satisfied a square root cancellation bound of the form
\begin{equation}
\left| \sum_{k \in I} e^{2 \pi i (\xi \cdot X_k)} \right| \lesssim \#(I)^{1/2},
\end{equation} then there would be no barrier to choosing
$M \approx r^{-\frac{dn-\alpha}{n-1}}$, which would allow us to prove the existence of a pattern avoiding set with Fourier dimension matching the quantity in equation (5.13), matching that of the Hausdorff dimension bound obtained in [Reference Denson, Pramanik and Zahl3]. Under stronger assumptions on the pattern we are trying to avoid, which form the hypotheses of Theorem 1.3, we are able to justify that some kind of square root cancellation, like that of (5.15) takes place, though with an additional term that we are only able to bound appropriately for
$n \gt 2$ using an inclusion-exclusion argument combined with some oscillatory integrals if we set
$N \approx r^{-\frac{dn-\alpha}{n-3/4}}$. Under the hypothesis of Theorem 1.4, we are able to make this additional term vanish completely, which will enable us to set
$N \approx r^{-\frac{dn - \alpha}{n - 1}}$, thus obtaining sets with Fourier dimension matching the quantity in equation (5.13), and completely recovering the dimension bound of [Reference Denson, Pramanik and Zahl3] in the setting of Salem sets.
6. Concentration bounds for smooth surfaces
In this section we prove Theorem 1.3 using some more robust probability concentration calculations, which allow us to justify the kinds of square root cancellation alluded to at the end of the last section. We set
\begin{equation*} \beta \leq \begin{cases} d &: n = 2 \\ d/(n - 3/4) &: n \geq 3 \end{cases}. \end{equation*} For such
$\beta$, we now prove that elements of the space
$\mathcal{X}_\beta$ will generically avoid patterns given by an equation
$x_n = f(x_1,\dots,x_{n-1})$, where
$f$ satisfies the hypotheses of Theorem 1.3.
Lemma 6.1. Suppose
$f: V \to \operatorname{\mathbb{T}}^d$ satisfy the hypothesis of Theorem 1.3. Then for quasi-all
$(E,\mu) \in \mathcal{X}_\beta$, and for any distinct points
$x_1,\dots,x_n \in E$,
$x_n \neq f(x_1,\dots,x_{n-1})$.
Proof. Given any family of disjoint, closed cubes
$R_1,\dots,R_n \subset \operatorname{\mathbb{T}}^d$ such that
$(R_1 \times \dots \times R_n) \cap V$ is a closed set, we let
\begin{equation*} \mathcal{H}(R_1,\dots,R_n) = \{E \subset \operatorname{\mathbb{T}}^d: \text{for all $x_i \in R_i \cap E$, $x_n \neq f(x_1,\dots,x_{n-1})$} \}, \end{equation*}and let
Then
$H(R_1,\dots,R_n)$ is an open subset of
$\mathcal{X}_\beta$. For the purpose of a Baire category argument, this proof will follow by showing
$H(R_1,\dots,R_n)$ is dense in
$\mathcal{X}_\beta$ for any family of disjoint cubes
$\{R_1,\dots, R_n \}$, each having common sidelength
$s$ for some
$s \gt 0$, such that if
$Q_i = 2R_i$ for each
$i$, then
$Q_2 \times \dots \times Q_n \subset V$, and
$d(R_i,R_j) \geq 10s$ for each
$i \neq j$. Since
$\mathcal{H}(R_1,\dots,R_n)$ is a downward closed family of sets, we will prove this result by applying Lemma 4.1. Thus for a suitable choice of
$r \gt 0$, we must construct a large discrete set
$S$ such that
$N(S,r) \in \mathcal{H}(R_1,\dots,R_n)$ and whose exponential sums exhibit square root cancellation.
Since
$f$ is smooth, we can fix a constant
$L \geq 0$ such that for any
$x,y \in Q_1 \times \dots \times Q_{n-1}$,
Fix a family of non-negative bump functions
$\psi_0,\psi_1,\dots,\psi_n \in C^\infty(\operatorname{\mathbb{T}}^d)$, such that for
$i \in \{1,\dots,n \}$,
$\psi_i(x) = 1$ for
$x \in 1.5 \cdot Q_i$,
$\psi_i(x) = 0$ for
$x \not \in R_i$, and
$\psi_0(x) + \dots + \psi_n(x) = 1$ for
$x \in \operatorname{\mathbb{T}}^d$. For
$i \in \{0, \dots, n \}$, let
$A_i = \int \psi_i(x)\; dx$ denote the total mass of
$\psi_i$. Now fix a large integer
$M \gt 0$, and consider a family of independent random variables
where the random variable
$X_i(k)$ is chosen with respect to the probability density function
$A_i^{-1} \psi_i$. Fix
$\lambda \in [0,\beta)$, and set
$r = M^{-1/\lambda}$, i.e. so that
$M = r^\lambda$. Let
$c = 2 n^{1/2} (L+1)$, and define
$I$ to be the set of all indices
$k_n \in \{1, \dots, M \}$ such that there are indices
$k_1,\dots,k_{n-1} \in \{1,\dots,M \}$ with the property that
Now (6.2) implies that if
$k_n \not \in I$, then for any
$k_1,\dots,k_{n-1} \in \{1, \dots, M \}$,
We now use these points to construct a set
$S$ verifying the hypotheses of Lemma 4.1.
Set
Then we claim that
$N(S,r) \in \mathcal{H}(R_1,\dots,R_n)$ for suitably small
$r$; to see this, suppose there were distinct
$y_1,\dots,y_n \in N(S,r)$ such that
$y_1 \in R_1, \dots, y_n \in R_n$, and
$y_n = f(y_1,\dots,y_{n-1})$. We may pick
$x_1,\dots,x_n \in S$ such that
$|x_i - y_i| \leq r$ for each
$i$. Since
$d(R_i,R_j) = 10s$ for
$i \neq j$, if
$r \leq s$, then it cannot be true that
$x_i = X_j(k)$ for some
$j \in \{1, \dots, n \}$ and
$k \in \{1, \dots, M \}$. Since
$\psi_i(x) = 1$ on
$1.5 R_i$, if
$r \lt 0.5 s$, we have
$d(\text{supp}(\psi_0), R_i) \geq 0.5 s$ and so it also cannot be true that
$x_i = X_0(k)$ for some
$k \in \{1, \dots, M \}$. Thus there must be
$k_i \in \{1, \dots, M \}$ such that
$x_i = X_i(k_i)$. But by assumption
$k_n \not \in I$, so we have
Thus (6.1) and (6.4) imply that
which gives a contradiction, proving that
$N(S,r) \in \mathcal{H}(R_1,\dots,R_n)$. The remainder of the proof focuses on bounding exponential sums associated with
$S$, so that we may apply Lemma 4.1 and thus prove the conclusion of the theorem.
Consider the random exponential sums
\begin{equation*} F(\xi) = \sum_{i \in \{0, \dots, n-1 \}} \sum_{k = 1}^M A_i e^{2 \pi i \xi \cdot X_i(k)} + \sum_{k \not \in I} A_n e^{2 \pi i \xi \cdot X_n(k)}. \end{equation*} Controlling
$|F(\xi)|$ with high probability will justify an application of Lemma 4.1. To analyze
$F$, introduce the auxiliary exponential sums
\begin{equation*} G(\xi) = \sum_{i = 0}^n \sum_{k = 1}^M A_i e^{2 \pi i \xi \cdot X_i(k)} \end{equation*}and
\begin{equation*} H(\xi) = \sum_{k \in I} A_n e^{2 \pi i \xi \cdot X_n(k)}. \end{equation*} Then
$F(\xi) = G(\xi) - H(\xi)$.
Obtaining a bound on
$G(\xi)$ is simple since it is a sum of
$(n+1) \cdot M$ independent random variables. For non-zero
$\xi \in \operatorname{\mathbb{Z}}^d$,
\begin{equation}
\begin{split}
\operatorname{\mathbb{E}}(G(\xi)) &= \sum_{i = 0}^n M A_i \int (\psi_i(x) / A_i) e^{2 \pi i \xi \cdot x}\; dx\\
&= M \sum_{i = 0}^n \int \psi_i(x) e^{2 \pi i \xi \cdot x}\; dx\\
&= M \int_{\operatorname{\mathbb{T}}^d} e^{2 \pi i \xi \cdot x}\; dx = 0.
\end{split}
\end{equation} Applying Lemma 4.4, we conclude that for any fixed
$\kappa \gt 0$, there is
$C \gt 0$ such that
\begin{equation}
\operatorname{\mathbb{P}} \left( \sup_{|\xi| \leq N^{1 + \kappa}} |G(\xi)| \geq C M^{1/2} \log(M)^{1/2} \right) \leq 1/10,
\end{equation} This bound guarantees
$G$ is sufficiently small with high probability.
Analyzing
$H(\xi)$ requires a more subtle concentration bound, which we delegate to a series of lemmas following this proof:
• In Lemma 6.2, we will employ some concentration bounds to show that
(6.8)
\begin{equation}
\operatorname{\mathbb{P}} \left( \sup_{|\xi| \leq N^{1 + \kappa}} | H(\xi) - \operatorname{\mathbb{E}}(H(\xi)) | \geq C M^{1/2} \log(M)^{1/2} \right) \leq 1/10.
\end{equation}• In Lemma 6.3 we will show that for any
$\delta \gt 0$, there exists
$r_1 \gt 0$ such that for
$r \leq r_1$ and any nonzero
$\xi \in \operatorname{\mathbb{Z}}^d$,
(6.9)
\begin{equation}
|\operatorname{\mathbb{E}}(H(\xi))| \leq \delta M |\xi|^{-\beta/2} + O(M^{1/2}).
\end{equation}
These bounds together guarantee that
$H$ is sufficiently small with high probability.
Analogous to equation (5.4) in Lemma 5.1, for any indices
$k_1,\dots,k_n \in \{1, \dots, M \}$, we have
\begin{equation}
\operatorname{\mathbb{P}} \Big( |X_n(k_n) - f(X_1(k_1),\dots,X_{n-1}(k_{n-1}))| \leq 2 n^{1/2} \cdot (L+1) \cdot r \Big) \lesssim_{n,L} r^d \lesssim M^{-d/\lambda}.
\end{equation} Thus if
$M_0$ denotes the number of tuples of indices
$(k_1,\dots,k_n)$ such that (6.2) holds, then (6.10) implies that
Applying Markov’s inequality to (6.11), we conclude that there exists a constant
$C \gt 0$ such that
Taking a union bound to (6.7), (6.8), and (6.12), and then applying (6.9), we conclude that there exists
$C \gt 0$ and a particular instantiation of the random variables
$\{X_i(k) \}$ such that for any
$0 \lt |\xi| \leq M^{1 + \kappa}$,
and
And
Since
$\lambda \lt \beta_0 \lt d/(n-1)$, the inequality
$d/\lambda - n \lt 1$ holds. Thus (6.15) implies that for sufficiently large
$M \gt 0$, if
$N = \#(S)$, then
Putting (6.9), (6.13), (6.14), and the fact that
$F(\xi) = G(\xi) + H(\xi)$ together, if we set
$\tilde{a}(X_i(k)) = A_i$ for each
$i$ and
$k$, then
\begin{equation}
\left| \frac{1}{N} \sum_{x \in S} \tilde{a}(x) e^{2 \pi i \xi \cdot x} \right| \lesssim C N^{1/2} \log(N)^{1/2} + \delta |\xi|^{-\beta/2}.
\end{equation} Since
$\sum_{x \in S} \tilde{a}(x) \geq N$, if we set
\begin{equation*} a(x) = N \cdot \frac{\tilde{a}(x)}{\sum_{x \in S} \tilde{a}(x)}, \end{equation*} then (6.17), (6.16) and the fact that
$N(S,r) \in \mathcal{H}(R_1,\dots,R_n)$ imply that the sum
\begin{equation*} \frac{1}{N} \sum_{x \in S} a(x) e^{2 \pi i \xi \cdot x} \end{equation*} satisfies the assumptions of Lemma 4.1 for arbitrarily large
$N$. We therefore conclude by that Lemma that
$H(R_1,\dots,R_n)$ is dense in
$\mathcal{X}_\beta$.
Our proof of Theorem 1.3 will be complete once we prove (6.8) and (6.9), i.e. once we prove Lemmas 6.2 and 6.3.
Lemma 6.2. Let
$H(\xi)$ be the random exponential sum described in Lemma 6.1. Then
\begin{equation*} \operatorname{\mathbb{P}} \left( \sup_{|\xi| \leq M^{1 + \kappa}} | H(\xi) - \operatorname{\mathbb{E}}(H(\xi)) | \geq C M^{1/2} \log(M)^{1/2} \right) \leq 1/10. \end{equation*} for some universal constant
$C \gt 0$.
Before we prove this Lemma, let us describe the idea behind the proof. The result is a concentration bound for the random quantity
$H(\xi)$, a standard topic in the theory of high dimensional probability. The basic heuristic of this topic is that for an arbitrary function
$F(X_1,\dots,X_N)$ of many independent random inputs, we have
$F(X_1,\dots,X_N) \approx \operatorname{\mathbb{E}}(F(X_1,\dots,X_N))$ with high probability, provided that each of the random inputs
$\{X_i \}$ has a small influence on the overall output of
$f$. McDiarmid’s and Hoeffding’s inequalities, described in the notation section of this paper, are two classic results in this theory. A major difference between the two inequalities is that McDiarmid’s inequality can be applied to nonlinear functions
$F$, whereas Hoeffding’s inequality can only be applied when
$F$ is linear.

Figure 1. The two diagrams displayed indicate two instances of the set
$S$ for the function
$f(x_1,x_2) = (x_1 - 1/2)^2 + (x_2 - 1/2)^2$. Here
$M = 3$,
$n = 3$, the values on the
$x$-axis represent the values
$X_1(1), X_1(2),$ and
$X_1(3)$, the values on the
$y$-axis represent the values
$X_2(1), X_2(2)$, and
$X_2(3)$, the dark points represent the family of all pairs
$(X_1(k_1),X_2(k_2))$, and the annuli represent the
$O(r)$-neighbourhoods of
$f^{-1}(X_3(1)), f^{-1}(X_3(2))$, and
$f^{-1}(X_3(3))$. In this setup,
$S$ consists of all of the values
$\{X_2(k) \}$, as well as all values of
$X_1(k)$ such that none of the dark points on the vertical line above
$X_1(k)$ intersect any of the annuli. The two diagrams only differ as a result of adjusting a single variable
$X_2(k_0)$, indicated by the shaded value on the
$y$-axis. For the values represented in the left diagram,
$I = \emptyset$, whereas for the values represented in the right diagram,
$I$ contains every index, and this completely alters the exponential sums associated with
$S$.
Since
$H(\xi)$ is a nonlinear function of the independent random quantities
$\{X_i(k) \}$, McDiarmid’s inequality presents itself as a useful concentration bound. However, a naive application of McDiarmid’s inequality fails here, because changing a single random variable
$X_i(k)$ for
$1 \leq i \leq n-1$ while fixing all other random variables can change the indices contained in the set
$I$ by as much as
$O(M)$, and thus change
$H(\xi)$ by as much as
$O(M)$ as a result (see Figure 1 for an example of this phenomenon). McDiarmid’s inequality then gives that
$|H(\xi) - \operatorname{\mathbb{E}}(H(\xi))| \lesssim M$ with high probability, which is not tight enough to obtain square root cancellation like what we obtained in (6.7). On the other hand, it seems that a single variable
$X_i(k)$ only changes
$H(\xi)$ by
$O(M)$ when the other random variables
$\{X_1(k) \}$ are configured in a very particular way, which is unlikely to happen. Thus we should expect that adjusting a single random variable
$X_i(k)$ does not influence the value of
$H(\xi)$ much when averaged over the possible choices of
$\{X_n(k) \}$. This leads us to first average over this first set of random variables, and then apply McDiarmid’s inequality, which yields the correct concentration result.
Proof of Lemma 6.2
Consider the random set
$\Omega$ of values
$x_n \in Q_n$ such that there are
$k_1,\dots,k_{n-1} \in \{1,\dots,M \}$ with
Then
\begin{equation}
H(\xi) = A_n \sum_{k = 1}^M Z(k,\xi).
\end{equation}where
\begin{equation*} Z(k,\xi) = \begin{cases} e^{2 \pi i \xi \cdot X_n(k)} &: X_n(k) \in \Omega, \\ 0 &: X_n(k) \not \in \Omega \end{cases}. \end{equation*} If
$\Sigma$ is the
$\sigma$-algebra generated by the random variables
then
$\Omega$ is measurable with respect to
$\Sigma$. Thus the random variables
$\{Z(k,\xi) \}$ are conditionally independent given
$\Sigma$. Since we have
$|Z(k,\xi)| \leq 1$ almost surely, Hoeffding’s inequality thus implies that for all
$t \geq 0$,
\begin{equation}
\operatorname{\mathbb{P}} \left( \left| H(\xi) - \operatorname{\mathbb{E}}(H(\xi)|\Sigma) \right| \geq t \right) \leq 4 \exp \left( \frac{-t^2}{2M} \right).
\end{equation}It is simple to see that
\begin{equation}
\operatorname{\mathbb{E}}(H(\xi) | \Sigma) = A_n M \int_\Omega \psi_n(x) e^{2 \pi i \xi \cdot x}\; dx.
\end{equation}Since
\begin{equation}
\Omega = \bigcup \left\{N(f(X_1(k_1),\dots,X_{n-1}(k_{n-1})), cr) : 1 \leq k_1,\dots,k_{n-1} \leq N \right\}.
\end{equation} we therefore see that varying each random variable
$X_i(k)$, for
$1 \leq i \leq n-1$ while fixing the other random variables adjusts at most
$M^{n-2}$ of the balls forming
$\Omega$, and thus varying
$X_i(k)$ while fixing the other random variables changes
$\operatorname{\mathbb{E}}(H(\xi)|\Sigma)$ by at most
\begin{equation}
M \cdot 2 \cdot (2cr)^d \cdot M^{n-2} \lesssim_{n,d,L} r^d M^{n-1} \lesssim 1.
\end{equation} Thus McDiarmid’s inequality shows that there exists a constant
$C$ depending on
$d$,
$n$, and
$L$, such that for any
$t \geq 0$,
\begin{equation}
\operatorname{\mathbb{P}} \left( |\operatorname{\mathbb{E}}(H(\xi)|\Sigma) - \operatorname{\mathbb{E}}(H(\xi))| \geq t \right) \leq 4 \exp \left( \frac{-t^2}{C M} \right).
\end{equation} Combining (6.20) and (6.24), we conclude that there exists a constant
$C \gt 0$ such that for each
$\xi \in \operatorname{\mathbb{Z}}^d$,
\begin{equation}
\operatorname{\mathbb{P}} \left( | H(\xi) - \operatorname{\mathbb{E}}(H(\xi)) | \geq t \right) \leq 8 \exp \left( \frac{-M t^2}{C} \right).
\end{equation} Applying a union bound to (6.25) over all
$0 \lt |\xi| \leq M^{1 + \kappa}$ shows that there exists a constant
$C \gt 0$ such that
\begin{equation*} \operatorname{\mathbb{P}} \left( \sup_{|\xi| \leq M^{1 + \kappa}} | H(\xi) - \operatorname{\mathbb{E}}(H(\xi)) | \geq C M^{1/2} \log(M)^{1/2} \right) \leq 1/10. \end{equation*} The analysis of (6.9) requires a different class of probabilistic techniques. For any set
$E \subset \operatorname{\mathbb{T}}^{d(n-1)}$, let
$A(E)$ denote the event that there exists indices
$k_1,\dots, k_{n-1}$ such that
Understanding the quantity
$\operatorname{\mathbb{E}}(H(\xi))$ will follow from an analysis of the values
$\operatorname{\mathbb{P}}(A(E))$. To see why this is true, note that because the random variables
$\{X_n(k) \}$ are all identically distributed, for a fixed
$\xi$, all the quantities
$\{Z(k,\xi) \}$ are identically distributed, and so
\begin{align*}
\operatorname{\mathbb{E}}(H(\xi)) &= M A_n \cdot \operatorname{\mathbb{E}}(Z(1,\xi))\\
&= M A_n \cdot \operatorname{\mathbb{E}}( \mathbb{I}(X_n(1) \in \Omega) e^{2 \pi i \xi \cdot X_n(1)} )\\
&= M A_n \cdot \int \operatorname{\mathbb{P}}(X_n(1) \in \Omega | X_n(1) = x_n) e^{2 \pi i \xi \cdot x_n} d\operatorname{\mathbb{P}}(X_n(1) = x_n)\\
&= M A_n \cdot \int \psi_n(x_n) \operatorname{\mathbb{P}}(1 \in I | X_n(1) = x_n) e^{2 \pi i \xi \cdot x_n}\; dx_n.
\end{align*} If we write
$E_{x_n} = f^{-1}(B_{cr}(x_n))$, then
$\operatorname{\mathbb{P}}(1 \in I | X_n(1) = x_n) = \operatorname{\mathbb{P}}(A(E_{x_n}))$, so that
\begin{equation}
\operatorname{\mathbb{E}}(H(\xi)) = MA_n \cdot \int \psi_n(x_n) \operatorname{\mathbb{P}}(A(E_{x_n})) \cdot e^{2 \pi i \xi \cdot x_n}\; dx_n.
\end{equation} If
$n = 2$, then we can explicitly calculate
$\operatorname{\mathbb{P}}(A(E)) \approx 1 - (1 - |E|)^M$ for any set
$E$, which makes this analysis of
$\operatorname{\mathbb{E}}[H(\xi)]$ more tractable. If
$n \gt 2$, the random vectors
are not independent of one another, which makes an analysis of the quantities
$\operatorname{\mathbb{P}}(A(E))$ difficult. An exception to this is when
$E = E_1 \times \dots \times E_{n-1}$ is a Cartesian product, in which case
\begin{equation*} \operatorname{\mathbb{P}}(A(E)) = \prod_{i = 1}^{n-1} \operatorname{\mathbb{P}}(\text{There is $k$ such that}\ X_i(k) \in E_i) \approx \prod_{i = 1}^{n-1} (1 - (1 - |E_i|)^M). \end{equation*} Our strategy to understanding the sets
$E_{x_n}$ for
$n \gt 2$ is therefore to apply the Whitney decomposition Lemma, writing
$E_{x_n} = \bigcup_i Q_i$ for a family of almost disjoint, axis-oriented cubes
$Q_i$. Then
$A(E_{x_n}) = \bigcup_i A(Q_i)$, so we can approximate
$\operatorname{\mathbb{P}}(A(E_{x_n}))$ by applying the inclusion-exclusion principle. This leads to a sufficient approximation to the values
$\operatorname{\mathbb{P}}(A(E_{x_n}))$ provided that
$M \ll r^{-d/(n-3/4)}$. This is the only part of the proof of Theorem 1.3 where the dimension bound becomes tight; increasing the dimension bound in Theorem 1.3 will be immediate if we can improve the following Lemma, i.e. finding a better analysis of
$\operatorname{\mathbb{E}}[H(\xi)]$.
Lemma 6.3. Let
$H(\xi)$ be the random exponential sum described in Lemma 6.1. Then there exists
$C \gt 0$ such that for any
$\delta \gt 0$, there exists
$M_0 \gt 0$ such that for
$M \geq M_0$,
Proof. We break the analysis of
$\operatorname{\mathbb{E}}(H(\xi))$ into two cases, depending on whether
$n = 2$ or
$n \gt 2$. Let’s start with the case
$n = 2$, in which case our assumptions imply that
$f$ is a diffeomorphism if the cubes
$R_1$ and
$R_2$ in which we are choosing random points are chosen small enough. For each
$x_2 \in \operatorname{\mathbb{T}}^d$, a change of variables shows that
\begin{equation}
\begin{split}
\operatorname{\mathbb{P}}(A(E_{x_2})) &= 1 - \left( 1 - \int_{f^{-1}(B_{cr}(x_2))} \psi_1(x_1)\; dx_1 \right)^M\\
&= 1 - \left( 1 - \int_{B_{cr}(x_2)} \frac{(\psi_1 \circ f^{-1})(x_1)}{|\det(Df)(f^{-1}(x_1))|}\; dx_1 \right)^M\\
&= 1 - \left( 1 - \int_{B_{cr}(x_2)} \tilde{\psi_1}(x_1)\; dx_1 \right)^M,
\end{split}
\end{equation}where
\begin{equation*} \tilde{\psi_1}(x_1) = \frac{(\psi_1 \circ f^{-1})(x_1)}{|\det(Df)(f^{-1}(x_1))|}. \end{equation*} If we define
$g(x_2) = \operatorname{\mathbb{P}}(A(E_{x_2}))$, then
$\operatorname{\mathbb{E}}(H(\xi)) = M A_2 \cdot \widehat{\psi_2 g}(\xi)$. We can obtain a bound on
$\operatorname{\mathbb{E}}(H(\xi))$ by bounding the partial derivatives of
$\psi_2 g$. Bernoulli’s inequality implies that
\begin{equation}
g(x) = 1 - \left( 1 - \int_{B_{cr}(x)} \tilde{\psi_1}(x_1)\; dx_1 \right)^M \lesssim_L M r^d \lesssim M^{1 - d/\lambda}.
\end{equation} On the other hand, for any multi-index
$\alpha$ with
$|\alpha| \gt 0$,
$\partial^\alpha g(x)$ is a sum of terms of the form
\begin{equation}
(-1)^m \frac{M!}{(M-m)!} \left( 1 - \int_{B_{cr}(x)} \tilde{\psi_1}(x_1)\; dx_1 \right)^{M-m} \left( \prod_{i = 1}^{m} \int_{B_{cr}(x)} \partial_{\alpha_i} \tilde{\psi_1}(x_1)\; dx_1 \right),
\end{equation} where
$\alpha_i \neq 0$ for any
$i$ and
$\alpha = \alpha_1 + \dots + \alpha_m$. This implies
$0 \lt m \leq |\alpha|$ for any terms in the sum. Now the bound
$|\partial_{\alpha_i} \tilde{\psi_1}(x_2)| \lesssim_{\alpha_i} 1$ implies that
\begin{equation}
\left| \int_{B_{cr}(x)} \partial_{\alpha_i} \tilde{\psi_1}(x_2)\; dx_2 \right| \lesssim_{\alpha_i} r^d.
\end{equation}Applying (6.30) to (6.29) enables us to conclude that
\begin{equation}
|\partial_\alpha g(x)| \lesssim_\alpha \max_{0 \lt m \leq |\alpha|} M^m r^{md} \leq M^{1 - d/\lambda},
\end{equation} Since the fact that
$\psi_2 \in C^\infty(\mathbb{T}^d)$ implies that
$\| \partial_\alpha \psi_2 \|_{L^\infty(\operatorname{\mathbb{T}}^d)} \lesssim_\alpha 1$ for any multi-index
$\alpha$, the product rule applied to (6.31) implies that
$\| \partial_\alpha (\psi_2 g) \|_{L^\infty(\operatorname{\mathbb{T}}^d)} \lesssim_\alpha M^{1 - d/\lambda}$ for all
$\alpha \gt 0$, which means that for any
$T \gt 0$ and
$\xi \neq 0$,
Since
$\lambda \lt d$,
$2 - d / \lambda \lt 1$, so setting
$T = \beta/2$, fixing
$\delta \gt 0$, and then choosing
$M_0$ appropriately, if
$M \geq M_0$, (6.32) shows that
This completes the proof in the case
$n = 2$.
Now we move on to the case where
$n \geq 3$, which is made more complicated by the lack of an explicit formula for
$\operatorname{\mathbb{P}}(A(E_{x_n}))$. For any cube
$Q \in \operatorname{\mathbb{T}}^{d(n-1)}$ and any indices
$1 \leq k_1,\dots,k_{n-1} \leq K$, set
$k = (k_1,\dots,k_{n-1})$ and let
$A(Q;k)$ denote the event that
$(X_1(k_1),\dots,X_{n-1}(k_{n-1})) \in Q$. Then
\begin{equation}
A(Q) = \bigcup_k A(Q;k).
\end{equation} For any cube
$Q$ and index
$k$,
\begin{equation}
\operatorname{\mathbb{P}}(A(Q;k)) = \int_Q \psi_1(x_1) \dots \psi_{n-1}(x_{n-1})\; dx_1 \dots dx_{n-1},
\end{equation}and so
\begin{equation}
\sum_k \operatorname{\mathbb{P}}(A(Q;k)) = M^{n-1} \int_Q \psi_1(x_1) \cdots \psi_{n-1}(x_{n-1})\; dx_1 \dots dx_{n-1}.
\end{equation}An application of inclusion exclusion to (6.36) thus shows that
\begin{equation}
\begin{split}
&\left| \operatorname{\mathbb{P}}(A(Q)) - M^{n-1} \int_Q \psi_1(x_1) \cdots \psi_{n-1}(x_{n-1})\; dx_1 \dots dx_{n-1} \right|\\
&\quad\quad\quad\leq \sum_{k \neq k'} \operatorname{\mathbb{P}}(A(Q;k) \cap A(Q;k')).
\end{split}
\end{equation} For each
$k,k'$, the quantity
$\operatorname{\mathbb{P}}(A(Q;k) \cap A(Q;k'))$ depends on the number of indices
$i$ such that
$k_i = k_i'$. In particular, if
$I \subset \{1, \dots, n-1 \}$ is the set of indices where the quantity agrees, then
\begin{equation}
\begin{split}
\operatorname{\mathbb{P}}(A(Q;k) \cap A(Q;k')) = \left( \prod_{i \in I} \int_{Q_i} \psi_i(x)\; dx \right) \cdot \left( \prod_{i \not \in I} \left( \int_{Q_i} \psi_i(x)\; dx \right)^2 \right).
\end{split}
\end{equation} In particular, if
$Q$ has sidelength
$l$ and
$\#(I) = m$, then
$\operatorname{\mathbb{P}}(A(Q;k) \cap A(Q;k')) \lesssim l^{d(2n - m - 2)}$. For each
$m$, there are at most
$M^{2n - m - 2}$ pairs
$k$ and
$k'$ with
$\#(I) = m$. And so provided
$l^d \leq 1/M$,
\begin{equation}
\sum_{k \neq k'} \operatorname{\mathbb{P}}(A(Q;k) \cap A(Q;k')) \lesssim \sum_{m = 0}^{n-2} (M \cdot l^d)^{2n-m-2} \lesssim M^n l^{dn}.
\end{equation}Thus we conclude from (6.37) and (6.39) that
\begin{equation}
\operatorname{\mathbb{P}}(A(Q)) = M^{n-1} \int_Q \psi_1(x_1) \dots \psi_{n-1}(x_{n-1}) dx_1 \dots dx_{n-1} + O(M^n l^{dn}).
\end{equation} Since
$f$ is a submersion, for each
$x_n$,
$E_{x_n}$ is contained in a
$O(r)$-thickening of a
$d(n-2)$ dimensional surface in
$\operatorname{\mathbb{T}}^{d(n-1)}$. Applying the Whitney covering lemma, we can find a family of almost disjoint dyadic cubes
$\{Q_{ij} : j \geq 0 \}$ such that
\begin{equation}
E_{x_n} = \bigcup_{i = 0}^\infty \bigcup_{j = 1}^{n_i} Q_{ij}(x_n),
\end{equation} where for each
$i \geq 0$,
$Q_{ij}$ is a sidelength
$r/2^i$ cube, and
$n_i \lesssim (r/2^i)^{-d(n-2)}$. It follows from (6.41) that
\begin{equation}
A(E_{x_n}) = \bigcup_{i,j} A(Q_{ij}).
\end{equation} Since
$n \geq 3$, we can use (6.40) to calculate that
\begin{equation}
\begin{split}
&\left| \sum_{i,j} \operatorname{\mathbb{P}}(A(Q_{ij})) - M^{n-1} \int_{E_{x_n}} \psi_1(x_1) \dots \psi_{n-1}(x_{n-1})\; dx \right|\\
&\quad\quad\quad\lesssim \sum_{i = 0}^\infty (r/2^i)^{-d(n-2)} \cdot \left( M^n (r/2^i)^{dn} \right)\\
&\quad\quad\quad\lesssim r^{2d} M^n \leq M^{-1/2}.
\end{split}
\end{equation}Thus an inclusion exclusion bound together with (6.42) and (6.43) implies that
\begin{equation}
\begin{split}
&\Big| \operatorname{\mathbb{P}}(A(E_{x_n})) - M^{n-1} \int_{E_{x_n}} \psi_1(x_1) \dots \psi_{n-1}(x_{n-1})\; dx \Big|\\
&\quad\quad\quad \lesssim M^{-1/2} + \sum_{(i_1,j_1) \neq (i_2,j_2)} \operatorname{\mathbb{P}}(A(Q_{i_1j_1}) \cap A(Q_{i_2j_2})).
\end{split}
\end{equation} The quantity
$\operatorname{\mathbb{P}}(A(Q_{i_1j_1}) \cap A(Q_{i_2j_2}))$ depends on the relation between the various sides of
$Q_{i_1j_1}$ and
$Q_{i_2j_2}$. Without loss of generality, we may assume that
$i_1 \geq i_2$. If
$I(Q_{i_1j_1},Q_{i_2j_2})$ is the set of indices
$1 \leq k \leq n-1$ where
$Q_{i_1j_1k} \subset Q_{i_2j_2k}$, and
$\#(I(Q_{i_1j_1}, Q_{i_2j_2})) = m$, then
\begin{equation}
\begin{split}
\operatorname{\mathbb{P}}(A(Q_{i_1j_1}) \cap A(Q_{i_2j_2})) &\lesssim (M(r/2^{i_1})^d)^m \cdot (M(r/2^{i_1})^d \cdot M(r/2^{i_2})^d)^{n-m-1}\\
&= 2^{-d[(n-1)i_1 + (n-m-1)i_2]} (Mr^d)^{2n - m-2}.
\end{split}
\end{equation} The condition that
$D_{x_k} f$ is invertible for all
$k$ on the domain of
$f$ implies that any axis-oriented plane in
$\operatorname{\mathbb{T}}^{dn}$ intersects transversally with the level sets of
$f$. In particular, this means that the intersection of a
$O(r/2^{i_1})$ thickening of a codimension
$dm$ axis-oriented hyperplane intersects a
$O(r/2^{i_1})$ thickening of
$\partial E_{x_n}$ (which has codimension
$d$) in a set with volume
$O \left( (r/2^{i_1})^d (r/2^{i_1})^{dm} \right)$, and intersects a
$O(r/2^{i_2})$ thickening of
$\partial E_{x_n}$ in a set with volume
$O \left( (r/2^{i_2})^d (r/2^{i_1})^{dm} \right)$. As a particular example of this, for any distinct indices
$j_1,\dots,j_m \in \{1,\dots, n-1 \}$, and any family of integers
$0 \leq n_{11},\dots,n_{md} \leq 2^{i_1}/r$, the set
\begin{equation}
\left\{x \in E_{x_n} : \frac{n_{11}}{2^{i_1}} \leq x_{j_1 1} \leq \frac{(n_{11} + 1)}{2^{i_1}}, \dots, \frac{n_{md}}{2^{i_1}} \leq x_{j_m d} \leq \frac{n_{md} + 1}{2^{i_1}} \right\}
\end{equation}contains at most
\begin{equation}
O \left( (r/2^{i_1})^d (r/2^{i_1})^{dm} (r/2^{i_1})^{-d(n-1)} \right) = O \left( 2^{d(n-m-2)i_1} r^{-d(n-m-2)} \right)
\end{equation} sidelength
$r/2^{i_1}$ dyadic cubes in the decomposition of
$E_{x_n}$, and at most
\begin{equation}
O \left( (r/2^{i_2})^d (r/2^{i_1})^{dm} (r/2^{i_2})^{-d(n-1)} \right) = O \left( 2^{d(n-2) i_2 - (dm) i_1} r^{-d(n-m-2)} \right)
\end{equation} sidelength
$r/2^{i_2}$ dyadic cubes in the decomposition of
$E_{x_n}$. Letting the integers
$\{n_{kl} \}$ vary over all possible choices we conclude from (6.47) and (6.48) that for each
$i_1$ and
$i_2$ there are at most
\begin{equation}
\begin{split}
&O \left( (2^{i_1}/r)^{dm} \left( 2^{d(n-m-2)i_1} r^{-d(n-m-2)} \right) \left( 2^{d(n-2) i_2 - (dm) i_1} r^{-d(n-m-2)} \right) \right)\\
&\quad\quad = O \left( 2^{d(n - m - 2)i_1 + d(n-2) i_2} r^{-d(2n - m - 4)} \right)
\end{split}
\end{equation} pairs
$Q_{i_1j_1}$ and
$Q_{i_2j_2}$ with
$I(Q_{i_1j_1},Q_{i_2j_2}) = m$. Thus we conclude from (6.45) and (6.49) that
\begin{equation}
\begin{split}
\sum_{(i,j) \neq (i',j')}& \operatorname{\mathbb{P}}(A(Q_{ij}) \cap A(Q_{i'j'}))\\
&\lesssim \sum_{m = 0}^{n-2} \sum_{i_1 \geq i_2} \left( 2^{d(n-m-2) i_1 + d(n-2) i_2} r^{-d(2n-m-4)} \right)\\
&\quad\quad\quad\quad\quad\quad\quad\quad\left( 2^{-d((n-1)i_1 + (n-m-1) i_2)} (Mr^d)^{2n - m - 2} \right)\\
&\lesssim r^{2d} \sum_{m = 0}^{n-2} M^{2n-m-2} \sum_{i_1 \geq i_2} 2^{-d(m+1)i_1 + d(m-1)i_2}\\
&\lesssim \sum_{m = 0}^{n-2} M^{2n-m-2} r^{2d}\\
&\lesssim M^{2(n-1)} r^{2d} \lesssim M^{-1/2}.
\end{split}
\end{equation}Returning to the bound in (6.44), (6.50) implies that
\begin{equation}
\left| \operatorname{\mathbb{P}}(A(E_{x_n})) - M^{n-1} \int_{E_{x_n}} \psi_1(x_1) \dots \psi_{n-1}(x_{n-1})\; dx_1 \dots dx_{n-1} \right| \lesssim M^{-1/2}.
\end{equation} Returning even further back to (6.26) , recalling that
$E_{x_n} = f^{-1}(B_r(x_n))$, (6.51) implies
\begin{align}
&\left| \operatorname{\mathbb{E}}(H(\xi)) - A_n \cdot M^n \int_{\operatorname{\mathbb{T}}^d} \psi_n(x_n) \int_{f^{-1}(B_r(x_n))} \psi_1(x_1) \dots \psi_{n-1}(x_{n-1})\right.\nonumber\\ &\quad \times e^{2 \pi i \xi \cdot x_n}\; dx_1\; \dots\; dx_n \Bigg| \lesssim M^{1/2}.
\end{align} Applying the co-area formula, writing
$\psi(x) = \psi_1(x_1) \dots \psi_n(x_n)$, we find
\begin{equation}
\begin{split}
\int_{\operatorname{\mathbb{T}}^d} & \int_{f^{-1}(B_r(x_n))} \psi(x) e^{2 \pi i \xi \cdot x_n}\; dx_1\; \dots\; dx_n\\
&= \int_{B_r(0)} \int_{\operatorname{\mathbb{T}}^d} \int_{f^{-1}(x + v)} \psi(x) e^{2 \pi i \xi \cdot x_n}\; dH^{n-2}(x_1,\dots,x_{n-1})\; dx_n\; dv\\
&= \int_{B_r(0)} \int_{\operatorname{\mathbb{T}}^{d(n-1)}} \psi(x,f(x) - v) \cdot e^{2 \pi i \xi \cdot (f(x) - v)} |Jf(x)|\; dx\; dv\\
&= \int_{B_r(0)} \int_{\operatorname{\mathbb{T}}^{d(n-1)}} \tilde{\psi}(x,v) \cdot e^{2 \pi i \xi \cdot (f(x) - v)}\; dx\; dv.
\end{split}
\end{equation} where
$\tilde{\psi}(x,v) = \psi(x,f(x) - v) \cdot |Jf(x)|$, and
$Jf$ is the rank-
$d$ Jacobian of
$f$. A consequence of (6.53) in light of (6.52) is that it reduces the study of
$\operatorname{\mathbb{E}}(H(\xi))$ to a standard oscillatory integral. If we look at the phase
then we see that
$\nabla_x \phi(x,v) = Df(x)^T \xi$, which is only equal to zero if
$\xi = 0$ since
$Df$ is surjective on the domain of
$f$ (this is implied by the stronger assumption that
$f$ is a diffeomorphism on each variable). Thus the oscillatory integral above has no stationary points in the
$x$-variable. Integrating by parts in the
$x$-variable thus allows us to conclude that for all
$|v| \leq 1$ and
$T \gt 0$,
\begin{equation}
\left|\int_{\operatorname{\mathbb{T}}^{d(n-1)}} \tilde{\psi}(x,v) \cdot e^{2 \pi i \xi \cdot (f(x) - v)}\; dx \right| \lesssim_T |\xi|^{-T}.
\end{equation}Now the bound in (6.54) can be applied with (6.53) to conclude that
\begin{equation}
\left| \int_{\operatorname{\mathbb{T}}^d} \int_{f^{-1}(B_r(x))} \psi(x) e^{2 \pi i \xi \cdot x_n}\; dx_2\; \dots\; dx_n\; dx_1 \right| \lesssim_T r^d |\xi|^{-T}.
\end{equation} In particular, taking
$T = \beta/2$ here, combined with (6.52), (6.55), we find that
Thus there exists
$C \gt 0$ such that for any
$\delta \gt 0$, there is
$r_0 \gt 0$ such that for
$r \leq r_0$, and any nonzero
$\xi \in \operatorname{\mathbb{Z}}^d$,
The proof of Lemma 6.3 is the only obstacle preventing us from constructing a Salem set
$X$ avoiding the pattern defined by
$Z$ with
\begin{equation*} \operatorname{dim}_{\mathbb{F}}(X) = \frac{d}{n-1}. \end{equation*} All other aspects of the proof carry through for
$d/(n-3/4) \leq \beta \leq d/(n-1)$. The problem with Lemma 6.3 in this scenario is that if we try to repeat the proof when
$n \geq 3$ and
$M \gg r^{-d/(n-3/4)}$, there is too much ‘overlap’ between the various cubes we use in our covering argument in the various axis; thus the inclusion-exclusion argument found in this proof cannot be used to control
$\operatorname{\mathbb{E}}(H(\xi))$ in a significant way. We believe our method can construct Salem sets with Fourier dimension
$d/(n-1)$, but new tools are required to improve the estimates on
$\operatorname{\mathbb{E}}(H(\xi))$. In the next section, we are able to modify our construction for patterns satisfying a weak translation invariance by a simple trick: we will modify the analogous exponential sums
$H(\xi)$ so that
$\operatorname{\mathbb{E}}(H(\xi)) = 0$ for all
$\xi \neq 0$, so that the analogue of Lemma 6.3 is trivial in this setting, and we thus obtain a construction of Salem sets avoiding patterns with dimension exactly matching those obtained in the Hausdorff dimension setting.
7. Expectation bounds for translation-invariant patterns
The proof of Theorem 1.4 uses very similar arguments to Theorem 1.3. The concentration bound arguments will be very similar to those applied in the last section. The difference here is that the translation-invariance of the pattern can be used to bypass estimating the expected values like those which caused us the most difficulty in Theorem 1.4. We can therefore construct Salem sets avoiding the pattern with dimension exactly matching the Hausdorff dimension of the sets which would be constructed using the method of [Reference Denson, Pramanik and Zahl3]. In this section, let
\begin{equation*} \beta \leq \min \left( \frac{dn - \alpha}{n-1}, d \right). \end{equation*} We then show that generic elements of
$\mathcal{X}_\beta$ avoid patterns satisfying the assumptions of Theorem 1.4.
Lemma 7.1. Fix
$a \in \operatorname{\mathbb{Q}} - \{0 \}$, and let
$f: V \to \operatorname{\mathbb{R}}$ and
$F \subset \operatorname{\mathbb{R}}$ satisfy the assumptions of Theorem 1.4. Then for quasi-all
$(E,\mu) \in \mathcal{X}_\beta$, and any distinct points
$(x_1,\dots,x_n) \in E$,
Proof. Set
\begin{equation*} W = \{(x_1,\dots,x_n) \in \operatorname{\mathbb{T}}^{2d} \times V : x_n - ax_{n-1} - f(x_1,\dots,x_{n-2}) \in F \}. \end{equation*} The assumption that
$f$ is a locally Lipschitz map, and thus continuous, implies that for any disjoint, closed cubes
$R_1,\dots,R_n \subset \operatorname{\mathbb{T}}^d$ such that
$R_1 \times \dots \times R_{n-2} \subset V$,
$(R_1 \times \dots \times R_n) \cap W$ will be a closed set. It follows that if we set
\begin{equation*} \mathcal{H}(R_1,\dots,R_n) = \{E \subset \operatorname{\mathbb{T}}^d : (R_1 \times \cdots \times R_n) \cap W \cap E^n = \emptyset \} \end{equation*}and
then
$H(R_1,\dots,R_n)$ is an open subset of
$\mathcal{X}_\beta$, and
$\mathcal{H}(R_1,\dots,R_n)$ is a downward closed family of sets. The proof will be complete will be proved that for each positive integer
$m$, and any choice of cubes
$R_1,\dots,R_n$ with common sidelength
$1/2am$, with
$d(R_i,R_j) \geq 10/am$ for
$i \neq j$, and with
$Q_1 \times \dots \times Q_{n-2}$ a closed subset of
$V$, where
$Q_i = 2R_i$, then the set
$H(R_1,\dots,R_n)$ is dense in
$\mathcal{X}_\beta$. To prove
$H(R_1,\dots,R_n)$ is dense, we may assume without loss of generality that the set
$F$ is
$1/m$ periodic, i.e.
$F + k / m = F$ for any
$k \in \operatorname{\mathbb{Z}}^d$, by replacing
$F$ with a finite union of its translates. The set
$H(R_1,\dots,R_n)$ is downward closed, so we can apply Lemma 4.1.
Since
$f$ is a locally Lipschitz map, we may fix
$L \gt 0$ such that for
$x_1,x_2 \in R_1 \times \dots \times R_{n-2}$,
Fix a large integer
$M \gt 0$,
$\lambda \in [0,\beta)$,
$\gamma \in [0,\alpha)$ and pick
$r \gt 0$ such that
$r^{-\lambda} \leq M \leq r^{-\lambda} + 1$. If
$c = 2(1 + |a| + L n^{1/2})$, and
$r$ is suitably small, then
For
$1 \leq i \leq n$, consider a family of independent random variables
$\{X_i(k): 1 \leq k \leq M \}$, such that
$X_i(k)$ is uniformly distributed on
$Q_i$ for each
$i$, as well as another independent family of random variables
$\{X_0(k): 1 \leq k \leq M \}$ uniformly distributed in
$\operatorname{\mathbb{T}}^d - (Q_1 \cup \dots \cup Q_n)$. Let
$I$ be the set of indices
$k_n \in \{1, \dots, N \}$ such that there are indices
$k_1,\dots,k_{n-1} \in \{1, \dots, N \}$ with the property that
\begin{equation}
d_{\mathbb{H}} \Big( X_n(k_n) - a X_{n-1}(k_{n-1}) - f \big( X_1(k_1), \dots, X_{n-2}(k_{n-2}) \big), F \Big) \leq cr.
\end{equation}If
then (7.1) implies that
$N(S,r) \in \mathcal{H}(W;R_1,\dots,R_n)$.
We claim that for each
$x \in Q_n$, the quantity
is independent of
$x$. To see this, we note that because
$F$ is
$1/m$ periodic, the quantity
\begin{equation*} d_{\mathbb{H}} \Big( x - a x_{n-1} - f \big( x_1,\dots,x_{n-2} \big), F \Big) \end{equation*} depends only on
$x_1,\dots,x_{n-2}$, and the value of
$x - a x_{n-1}$ in
$\operatorname{\mathbb{T}}^d / (\operatorname{\mathbb{Z}}^d / m)$. Because
$a X_{n-1}(k_{n-1})$ is uniformly distributed in
$a Q_n$, which is an axis-oriented cube with sidelength
$1/m$, it follows that the distribution of the random variable
modulo
$\operatorname{\mathbb{T}}^d / (\operatorname{\mathbb{Z}}^d / m)$, is independent of
$x$, which yields the claim.
Let
$P$ denote the common quantity of the values
$P(x)$, and let
\begin{equation*} G(x_1,\dots,x_{n-1}) = d_{\mathbb{H}} \Big( - a x_{n-1} - f \big( x_1, \dots, x_{n-2}) \big), F \Big). \end{equation*}A union bound, together with (7.2), implies
\begin{equation}
\begin{split}
P &= \operatorname{\mathbb{P}} \left( \bigcup_{k_1,\dots,k_{n-1}} \Big\{G(X_1(k_1),\dots, X_{n-1}(k_{n-1})) \leq cr \Big\} \right)\\
&\leq \sum_{k_1,\dots,k_{n-1}} \operatorname{\mathbb{P}} \left( G(X_1(k_1), \dots, X_{n-1}(k_{n-1})) \leq cr \right) \\
&\lesssim M^{n-1} |N(F,cr)| \leq M^{n-1} r^{dn-\gamma} \lesssim M^{n - 1 - (dn - \gamma)/\lambda}
\end{split}
\end{equation} Because
$(n-1) - (dn - \gamma)/\lambda \lt 0$, (7.4) implies that for suitably large integers
$M$ depending on
$n$,
$d$,
$\lambda$, and
$\gamma$,
Set
$A_0 = (1 - P) |\operatorname{\mathbb{T}}^d - Q_1 - \dots - Q_n|$, set
$A_i = (1 - P) |Q_i|$ if
$i \in \{1, \dots, n - 1 \}$, and let
$A_n = |Q_n|$. Define
\begin{equation*} F(\xi) = \sum_{i = 0}^{n-1} \sum_{k = 1}^M A_i e^{2 \pi i \xi \cdot X_i(k)} + \sum_{k \not \in I} A_n e^{2 \pi i \xi \cdot X_n(k)}. \end{equation*} The choice of coefficients is made so that
$\sum A_i \gt rsim 1$, and for any
$\xi \neq 0$,
$\operatorname{\mathbb{E}}(F(\xi)) = 0$. Indeed, we have
\begin{align*}
\operatorname{\mathbb{E}}(F(\xi)) &= \frac{M A_0}{|\operatorname{\mathbb{T}}^d - Q_1 - \dots - Q_n|} \int_{\operatorname{\mathbb{T}}^d - Q_1 - \dots - Q_n} e^{2 \pi i \xi \cdot x}\; dx\\
&\quad\quad + \sum_{i = 1}^{n-1} \frac{M A_i}{|Q_i|} \int_{Q_i} e^{2 \pi i \xi \cdot x}\; dx\\
&\quad\quad + \frac{M A_n}{|Q_n|} \int_{R_n} [1 - P(x)] e^{2 \pi i \xi \cdot x}; dx\\
&= M (1 - P) \int_{\operatorname{\mathbb{T}}^d} e^{2 \pi i \xi \cdot x} = 0.
\end{align*} Split up
$F$ into the sum of two exponential sums
\begin{equation*} G(\xi) = \sum_{i = 0}^n \sum_{k = 1}^M A_i e^{2 \pi i \xi \cdot X_i(k)} \end{equation*}and
\begin{equation*} H(\xi) = \sum_{k \in I} A_n e^{2 \pi i \xi \cdot X_n(k)}. \end{equation*} Applying Lemma 4.4, we conclude that for any fixed
$\kappa \gt 0$, there is
$C \gt 0$ such that
\begin{equation}
\operatorname{\mathbb{P}} \left( \sup_{|\xi| \leq N^{1 + \kappa}} |G(\xi) - \operatorname{\mathbb{E}}(G(\xi))| \geq C M^{1/2} \log(M)^{1/2} \right) \leq 1/10.
\end{equation}Lemma 7.2, which follows from a very similar argument to Lemma 6.2 in the last section, implies that
\begin{equation}
\begin{split}
\operatorname{\mathbb{P}} & \left( \sup_{|\xi| \leq M^{1 + \kappa}} | H(\xi) - \operatorname{\mathbb{E}}(H(\xi)) | \geq C M^{1/2} \log(M)^{1/2} \right) \leq 1/10.
\end{split}
\end{equation} A union bound applied to (7.6) and (7.7) implies that, since
$\operatorname{\mathbb{E}}(F(\xi)) = 0$,
\begin{equation}
\operatorname{\mathbb{P}} \left( \sup_{|\xi| \leq M^{1 + \kappa}} | F(\xi) | \geq C M^{1/2} \log(M)^{1/2} \right) \leq 1/5
\end{equation} Set
$N = \#(S)$. Then
Now (7.6), (7.7), and (7.9) imply that there exists a constant
$C \gt 0$ and an instantiation of the random variables
$\{X_i(k) \}$ such that if
$\tilde{a}(X_i(k)) = A_i$, then
\begin{equation}
\left| \frac{1}{N} \sum_{x \in S} \tilde{a}(x) e^{2 \pi i \xi \cdot x} \right| \lesssim C N^{1/2} \log(N)^{1/2}.
\end{equation} Since
$\sum_{x \in S} \tilde{a}(x) \gt rsim N$, if we set
\begin{equation*} a(x) = N \cdot \frac{\tilde{a}(x)}{\sum_{x' \in S} a(x')}, \end{equation*}then the sum
\begin{equation*} \frac{1}{N} \sum_{x \in S} a(x) e^{2 \pi i \xi \cdot x} \end{equation*} satisfies the assumptions of Lemma 4.1 for arbitrarily large
$N$. We therefore conclude by that Lemma that
$H(R_1,\dots,R_n)$ is dense in
$\mathcal{X}_\beta$.
All that remains to prove Theorem 1.4 is to prove Lemma 7.2.
Lemma 7.2. For any
$\kappa \gt 0$, there exists
$C \gt 0$ such that
\begin{equation*} \operatorname{\mathbb{P}} \left( \sup_{|\xi| \leq M^{1 + \kappa}} | H(\xi) - \operatorname{\mathbb{E}}(H(\xi)) | \geq C M^{1/2} \log(M)^{1/2} \right) \leq 1/10 \end{equation*}Proof. Consider the random set
$\Omega$ of values
$x_n \in Q_n$ such that there are
$k_1,\dots,k_{n-1} \in \{1,\dots,M \}$ with
Then
\begin{equation}
H(\xi) = \frac{A_n}{M} \sum_{k = 1}^M Z(k,\xi).
\end{equation}where
\begin{equation*} Z(k,\xi) = \begin{cases} e^{2 \pi i \xi \cdot X_n(k)} &: X_n(k) \not \in \Omega, \\ 0 &: X_n(k) \in \Omega \end{cases}. \end{equation*} If
$\Sigma$ is the
$\sigma$ algebra generated by the random variables
then
$\Omega$ is measurable with respect to
$\Sigma$. Thus the random variables
$\{Z(k,\xi) \}$ are conditionally independent given
$\Sigma$. Since we have
$|Z(k,\xi)| \leq 1$ almost surely, Hoeffding’s inequality thus implies that for all
$t \geq 0$,
\begin{equation}
\operatorname{\mathbb{P}} \left( \left| H(\xi) - \operatorname{\mathbb{E}}(H(\xi)|\Sigma) \right| \geq t \right) \leq 4 \exp \left( \frac{-t^2}{2M} \right).
\end{equation}It is simple to see that
\begin{equation}
\operatorname{\mathbb{E}}(H(\xi) | \Sigma) = A_n M \int_\Omega \psi_n(x) e^{2 \pi i \xi \cdot x}\; dx.
\end{equation}Since
\begin{equation}
\Omega = \bigcup \left\{f(X_1(k_1), \dots, X_{n-2}(k_{n-2})) + N(F, cr) : 1 \leq k_1,\dots,k_{n-1} \leq K \right\}.
\end{equation} we see that varying each random variable
$X_i(k)$, for
$1 \leq i \leq n-1$ while fixing the other random variables adjusts at most
$M^{n-2}$ of the sets forming
$\Omega$, each of which having volume
$O_{d,n,L}(r^{dn - \alpha})$, and thus varying
$X_i(k)$ while fixing the other random variables changes
$\operatorname{\mathbb{E}}(H(\xi)|\Sigma)$ by at most
\begin{equation}
M \cdot O_{d,n,L}(r^{dn - \alpha}) \cdot M^{n-1} \lesssim 1.
\end{equation} Thus McDiarmid’s inequality shows that there exists
$C \gt 0$ such that for any
$t \geq 0$,
\begin{equation}
\operatorname{\mathbb{P}} \left( |\operatorname{\mathbb{E}}(H(\xi)|\Sigma) - \operatorname{\mathbb{E}}(H(\xi))| \geq t \right) \leq 4 \exp \left( \frac{-t^2}{CM} \right).
\end{equation} Combining (7.13) and (7.17), we conclude that there exists
$C \gt 0$ such that for each
$\xi \in \operatorname{\mathbb{Z}}^d$,
\begin{equation}
\operatorname{\mathbb{P}} \left( | H(\xi) - \operatorname{\mathbb{E}}(H(\xi)) | \geq t \right) \leq 8 \exp \left( \frac{-t^2}{CM} \right).
\end{equation} Applying a union bound to (7.18) over all
$0 \lt |\xi| \leq M^{1 + \kappa}$ shows that there exists a constant
$C \gt 0$ such that
\begin{equation*} \operatorname{\mathbb{P}} \left( \sup_{|\xi| \leq M^{1 + \kappa}} | H(\xi) - \operatorname{\mathbb{E}}(H(\xi)) | \geq C M^{1/2} \log(M)^{1/2} \right) \leq 1/10.
\end{equation*}Appendix A. Justifying discretization
The main goal of this appendix is a proof of Lemma 4.1. Throughout this section, we will apply mollification. So we fix a smooth, non-negative function
$\phi \in C^\infty(\operatorname{\mathbb{T}}^d)$ such that
$\phi(x) = 0$ for
$|x| \geq 2/5$ and
$\int_{\operatorname{\mathbb{T}}^d} \phi(x)\; dx = 1$.
For each
$r \in (0,1)$, we can then define
$\phi_r \in C^\infty(\operatorname{\mathbb{T}}^d)$ by writing
\begin{equation*} \phi_r(x) = \begin{cases} r^{-d} \phi(x/r) &: |x| \lt r, \\ 0 &: \text{otherwise}. \end{cases} \end{equation*}The following standard properties hold:
(1) For each
$r \in (0,1)$,
$\phi_r$ is a non-negative smooth function with
(A.1)
\begin{equation}
\int_{\operatorname{\mathbb{T}}^d} \phi_r(x)\; dx = 1,
\end{equation}and
$\phi_r(x) = 0$ for
$|x| \geq r$.(2) For any
$r \in (0,1)$,
(A.2)
\begin{equation}
\big\| \widehat{\phi}_r \big\|_{L^\infty(\operatorname{\mathbb{Z}}^d)} = 1.
\end{equation}(3) For each
$\xi \in \operatorname{\mathbb{Z}}^d$,
(A.3)
\begin{equation}
\lim_{r \to 0} \widehat{\phi}_r(\xi) = 1.
\end{equation}(4) For each
$T \gt 0$, for all
$r \gt 0$, and for any non-zero
$\xi \in \operatorname{\mathbb{Z}}^d$,
(A.4)
\begin{equation}
|\widehat{\phi}_r(\xi)| \lesssim_T r^{-T} |\xi|^{-T}.
\end{equation}
We will prove Lemma 4.1 after a series of more elementary lemmas which give results about the metric space
$\mathcal{X}_\beta$.
Lemma A.1. The set of all
$(E,\mu) \in \mathcal{X}_\beta$ with
$\mu \in C^\infty(\operatorname{\mathbb{T}}^d)$ and
$\text{supp}(\mu) = E$ is dense in
$\mathcal{X}_\beta$.
Proof. Let
\begin{equation}
\tilde{\mathcal{X}_\beta} = \{(E,\mu) \in \mathcal{X}_\beta : \text{supp}(\mu) = E \}.
\end{equation} We begin by proving that the set of all
$(E,\mu) \in \tilde{\mathcal{X}_\beta}$ such that
$\mu \in C^\infty(\operatorname{\mathbb{T}}^d)$ is dense in
$\tilde{\mathcal{X}_\beta}$. Fix
$(E_0,\mu_0) \in \tilde{\mathcal{X}_\beta}$. For each
$r \in (0,1)$, consider the convolved measure
$\mu_r = \mu_0 * \phi_r$. Then
$\mu_r \in C^\infty(\operatorname{\mathbb{T}}^d)$ and
$\text{supp}(\mu_r) = E_r$. We claim that
$\lim_{r \to 0} (E_r,\mu_r) = (E_0,\mu_0)$, which would complete the proof. Since
$d_{\mathbb{H}}(E_0,E_r) \leq r$, we find that
$\lim_{r \to 0} E_r = E_0$ holds with respect to the Hausdorff metric. Now fix
$\lambda \in (0,\beta]$ and
$\delta \gt 0$. For each
$\xi \in \operatorname{\mathbb{Z}}^d$,
$|\widehat{\mu}_r(\xi)| = |\widehat{\phi}_r(\xi)| |\widehat{\mu}_0(\xi)|$, so
\begin{equation}
|\xi|^{\lambda/2} |\widehat{\mu}_r(\xi) - \widehat{\mu}_0(\xi)| = |\xi|^{\lambda/2} |\widehat{\phi}_r(\xi) - 1| |\widehat{\mu}_0(\xi)|.
\end{equation} We control (A.6) using the fact that
$|\widehat{\mu}_0(\xi)|$ is small when
$\xi$ is large, and
$|\widehat{\phi}_r(\xi) - 1|$ is small when
$\xi$ is small. Since
$(E_0,\mu_0) \in \mathcal{X}_\beta$, we can apply (4.1) to find
$R \gt 0$ such that for
$|\xi| \geq R$,
Combining (A.6), (A.7), and (A.2), for
$|\xi| \geq R$ we find that
On the other hand, (A.3) shows that there exists
$r_0 \gt 0$ such that for
$r \leq r_0$ and
$|\xi| \leq R$,
\begin{equation}
|\xi|^{\lambda/2} |\widehat{\phi}_r(\xi) - 1| \leq \delta.
\end{equation} The
$(L^1,L^\infty)$ bound for the Fourier transform implies that
$|\widehat{\mu}_0(\xi)| \leq \mu_0(\operatorname{\mathbb{T}}^d) = 1$, which combined with (A.9) gives that for
$r \leq r_0$ and
$|\xi| \leq R$,
Putting together (A.8) and (A.10) shows that for
$r \leq r_0$,
$\| \mu_r - \mu_0 \|_{M(\lambda)} \leq \delta$. Since
$\delta$ and
$\lambda$ were arbitrary, we conclude that
$\lim_{r \to 0} \mu_r = \mu_0$. Thus the set of all pairs
$(E,\mu) \in \tilde{\mathcal{X}_\beta}$ with
$\mu \in C^\infty(\operatorname{\mathbb{T}}^d)$ is dense in
$\tilde{\mathcal{X}_\beta}$.
Our proof will therefore be complete if we can show that
$\tilde{\mathcal{X}_\beta}$ is dense in
$\mathcal{X}_\beta$. We prove this using a Baire category argument. For each closed cube
$Q \subset \operatorname{\mathbb{T}}^d$, let
\begin{equation*} A(Q) = \{(E,\mu) \in \operatorname{\mathbb{T}}^d: (E \cap Q) = \emptyset\ \text{or}\ \mu(Q) \gt 0 \}. \end{equation*} Then
$A(Q)$ is an open set. If
$\{Q_k \}$ is a countable sequence enumerating all cubes with rational corners in
$\operatorname{\mathbb{T}}^d$, then
\begin{equation}
\bigcap_{k = 1}^\infty A(Q_k) = \tilde{\mathcal{X}_\beta}.
\end{equation} Thus it suffices to show that
$A(Q)$ is dense in
$\mathcal{X}_\beta$ for each closed cube
$Q$. To do this, we fix
$(E_0,\mu_0) \in \mathcal{X}_\beta - A(Q)$,
$\lambda \in [0,\beta)$, and
$\varepsilon \gt 0$, and try and find
$(E,\mu) \in A(Q)$ with
$d_{\mathbb{H}}(E,E_0) \leq \varepsilon$ and
$\| \mu_0 - \mu \|_{M(\lambda)} \leq \varepsilon$.
Because
$(E_0,\mu_0) \in \mathcal{X}_\beta - A(Q)$, we know
$E_0 \cap Q \neq \emptyset$ and
$\mu_0(Q) = 0$. Find a smooth probability measure
$\nu$ supported on
$N(E_0,\varepsilon) \cap Q$ and, for
$t \in (0,1)$, define
$\mu_t = (1 - t) \mu_0 + t \nu$. Then
$\text{supp}(\mu_t) \subset N(E_0,\varepsilon)$, so if we let
$E = \text{supp}(\nu) \cup \text{supp}(\mu)$, then
$d_{\mathbb{H}}(E,E_0) \leq \varepsilon$. Clearly
$(E,\mu_t) \in A(Q)$ for
$t \gt 0$. And
\begin{equation}
\| \mu_t - \mu_0 \|_{M(\lambda)} \leq t \left( \| \mu_0 \|_{M(\lambda)} + \| \nu \|_{M(\lambda)} \right),
\end{equation} so if we choose
$t \leq \varepsilon \cdot (\| \mu \|_{M(\lambda)} + \| \nu \|_{M(\lambda)})^{-1}$ we find
$\| \mu_t - \mu \|_{M(\lambda)} \leq \varepsilon$. Since
$\varepsilon$ was arbitrary, we conclude
$A(Q)$ is dense in
$\mathcal{X}_\beta$.
Remark A.2. The reason we must work with the metric space
$\mathcal{X}_\beta$ rather than the smaller space
$\tilde{\mathcal{X}_\beta} \subset \mathcal{X}_\beta$ is that
$\tilde{\mathcal{X}_\beta}$ is not a closed subset of
$\mathcal{X}_\beta$, and so is not a complete metric space, preventing the use of the Baire category theorem. However, the latter arguments in the proof of Lemma A.1 shows that quasi-all elements of
$\mathcal{X}_\beta$ belong to
$\tilde{\mathcal{X}_\beta}$, so that one can think of
$\mathcal{X}_\beta$ and
$\tilde{\mathcal{X}_\beta}$ as being equal ‘generically’.
The density argument of Lemma 4.1 requires constructing approximations to an arbitrary element of
$(E_0,\mu_0) \in \mathcal{X}_\beta$ by
$(E,\mu) \in \mathcal{X}_\beta$ such that
$E \in \mathcal{A}$. We do this by multiplying
$\mu_0$ by a smooth function
$f \in C^\infty(\operatorname{\mathbb{T}}^d)$ which cuts off parts of
$\mu_0$ which cause the support of
$\mu_0$ to fail to be in
$\mathcal{A}$. As long as
$\mu_0$ is appropriately smooth, and the Fourier transform of
$f$ decays appropriately quickly, the next lemma shows that
$f \mu_0 \approx \mu_0$.
Lemma A.3. Consider a finite measure
$\mu_0$ on
$\operatorname{\mathbb{T}}^d$, as well as a smooth probability density function
$f \in C^\infty(\operatorname{\mathbb{T}}^d)$. If we define
$\mu = f \mu_0$, then for any
$\lambda \in [0,d)$,
Proof. Since
$\widehat{\mu} = \widehat{f} * \widehat{\mu_0}$, and
$\widehat{f}(0) = 1$, for each
$\xi \in \operatorname{\mathbb{Z}}^d$ we have
\begin{equation}
\begin{split}
|\xi|^{\lambda/2} |\widehat{\mu}(\xi) - \widehat{\mu}_0(\xi)| &= |\xi|^{\lambda/2} \left| \sum_{\eta \neq \xi} \widehat{f}(\xi - \eta) \widehat{\mu}_0(\eta) \right|.
\end{split}
\end{equation} If
$|\eta| \leq |\xi|/2$, then
$|\xi|/2 \leq |\xi - \eta| \leq 2 |\xi|$, so
\begin{equation}
|\xi|^{\lambda/2} |\widehat{f}(\xi - \eta)| \leq \| f \|_{M(\lambda)} |\xi|^{\lambda/2} |\xi-\eta|^{-\lambda/2} \leq 2^{\lambda/2} \| f \|_{M(\lambda)} \lesssim_d \| f \|_{M(\lambda)}.
\end{equation}Thus the bound (A.14) implies
\begin{equation}
\begin{split}
|\xi|^{\lambda/2} \left| \sum_{0 \leq |\eta| \leq |\xi|/2} \widehat{f}(\xi - \eta) \widehat{\mu}_0(\eta) \right| &\lesssim_{\mu_0,d} \| \mu_0 \|_{M(d+1)} \| f \|_{M(\lambda)} \left( 1 + \sum_{0 \lt |\eta| \leq |\xi|/2} \frac{1}{|\eta|^{d+1}} \right)\\
&\lesssim_d \| \mu_0 \|_{M(d+1)} \| f \|_{M(\lambda)} \leq \| \mu_0 \|_{M(d+1)} \| f \|_{M(\lambda)}.
\end{split}
\end{equation} On the other hand, for all
$\eta \neq \xi$,
\begin{equation}
\begin{split}
|\widehat{f}(\xi - \eta)| \leq \| f \|_{M(\lambda)} |\xi - \eta|^{-\lambda} \leq \| f \|_{M(\lambda)}.
\end{split}
\end{equation}Thus we calculate that
\begin{equation}
\begin{split}
|\xi|^{\lambda/2} \left| \sum_{\substack{|\eta| \gt |\xi|/2\\ \eta \neq \xi}} \widehat{f}(\xi - \eta) \widehat{\mu}_0(\eta) \right| &\lesssim_{d,\mu_0} \| \mu_0 \|_{M(d+1)} \| f \|_{M(\lambda)} \cdot |\xi|^{\lambda/2} \sum_{|\eta| \gt |\xi|/2} \frac{1}{|\eta|^{d+1}}\\
&\lesssim_d \| \mu_0 \|_{M(d+1)} \| f \|_{M(\lambda)}.
\end{split}
\end{equation} The bound in Lemma A.3, if
$\| f \|_{M(\lambda)}$ is taken appropriately small, also implies that the Hausdorff distance between the supports of
$\mu$ and
$\mu_0$ are small.
Lemma A.4. Fix a probability measure
$\mu_0 \in C^\infty(\operatorname{\mathbb{T}}^d)$ and
$\lambda \in [0,d)$. For any
$\varepsilon \gt 0$, there exists
$\delta \gt 0$ such that if
$\mu \in C^\infty(\operatorname{\mathbb{T}}^d)$,
$\text{supp}(\mu) \subset \text{supp}(\mu_0)$, and
$\| \mu_0 - \mu \|_{M(\lambda)} \leq \delta$, then
$d_{\mathbb{H}}(\text{supp}(\mu),\text{supp}(\mu_0)) \leq \varepsilon$.
Proof. Consider any cover of
$\text{supp}(\mu_0)$ by a family of radius
$\varepsilon/3$ balls
$\{B_1,\dots,B_N \}$, and for each
$i \in \{1, \dots, N \}$, consider a smooth function
$f_i \in C_c^\infty(B_i)$ such that there is
$s \gt 0$ with
\begin{equation}
\int f_i(x) d\mu_0(x) \geq s
\end{equation} for each
$i \in \{1, \dots, N \}$. Fix
$A \gt 0$ with
\begin{equation}
\sum_{\xi \neq 0} |\widehat{f}_i(\xi)| \leq A
\end{equation} for all
$i \in \{1, \dots, N \}$ as well. Set
$\delta = s/2A$. If
$\| \mu_0 - \mu \|_{M(\lambda)} \leq \delta$, we apply Plancherel’s theorem together with (A.18) and (A.19) to conclude that
\begin{equation}
\begin{split}
\left| \int f_i(x) d\mu(x)\; dx - \int f_i(x) d\mu_0(x) \right| &= \left| \sum_{\xi \in \operatorname{\mathbb{Z}}^d} \widehat{f}_i(\xi) \left( \widehat{\mu}(\xi) - \widehat{\mu}_0(\xi) \right) \right|\\
&\leq A \| \mu_0 - \mu \|_{M(\lambda)}\\
&\leq s/2.
\end{split}
\end{equation}Thus we conclude from (A.18) and (A.20) that
\begin{equation}
\int f_i(x) d\mu(x)\; dx \geq \int f_i(x) d\mu_0(x) - s/2 \geq s/2 \gt 0.
\end{equation} Since equation (A.21) holds for each
$i \in \{1,\dots, N \}$, the support of
$\mu$ intersects every ball in
$\{B_1, \dots, B_N \}$. Combined with the assumption that
$\text{supp}(\mu) \subset \text{supp}(\mu_0)$, this implies that
$d_{\mathbb{H}}(\mu_0,\mu) \leq \varepsilon$.
Now we have the technology to prove Lemma 4.1.
Proof of Lemma 4.1
Fix
$(E_0,\mu_0) \in \mathcal{X}_\beta$. By Lemma A.1, without loss of generality, we may assume that
$\mu_0 \in C^\infty(\operatorname{\mathbb{T}}^d)$ and that
$\text{supp}(\mu_0) = E_0$. Our goal, for any
$\lambda \in [0,\beta)$ and
$\delta_0 \gt 0$, is to find
$(E,\mu) \in \mathcal{X}_\beta$ such that
$E \in \mathcal{A}$,
$d_{\mathbb{H}}(E,E_0) \leq \delta_0$, and
$\| \mu - \mu_0 \|_{M(\gamma)} \leq \delta_0$.
Fix
$\delta \gt 0$,
$\varepsilon \gt 0$, and
$\lambda \in (\gamma,\beta)$, and consider a set
$S = \{x_1, \dots, x_N \}$ and
$\{a_1, \dots, a_N \}$ satisfying the assumptions of the Lemma. If we set
\begin{equation*} \eta = \frac{1}{N} \sum_{k = 1}^N a_k \delta_{x_k}, \end{equation*} then
$\eta$ is a probability measure, and Property (2) implies that for
$|\xi| \leq (1/r)^{1 + \kappa}$,
Consider the function
$f = \eta * \phi_r$, where
$\phi_r$ is the mollifier defined in the notation section. For each
$\xi \in \operatorname{\mathbb{Z}}^d$,
\begin{equation}
\widehat{f}(\xi) = \widehat{\eta}(\xi) \widehat{\phi}_r(\xi).
\end{equation} For
$|\xi| \leq 1/r$, (A.22) and (A.2) together with (A.23) imply that there is
$\kappa_1 \gt 0$ depending on
$\beta,\lambda$, and
$\gamma$ such that
\begin{equation}
|\widehat{f}(\xi)| \leq C N^{-1/2} \log(N) + \delta |\xi|^{-\lambda/2} \leq (C N^{-\kappa_1} + \delta) |\xi|^{-\gamma/2}.
\end{equation} Thus if
$N$ is suitably large, we conclude that for
$|\xi| \leq 1/r$,
\begin{equation}
|\widehat{f}(\xi)| \leq 2 \delta |\xi|^{-\gamma/2}.
\end{equation} If
$(1/r) \leq |\xi| \leq (1/r)^{1+\kappa}$, (A.4) implies
$|\widehat{\phi}_r(\xi)| \lesssim_\beta r^{-\beta/2} |\xi|^{-\beta/2}$, which together with (A.2) and (A.22) applied to (A.23) allows us to conclude that there is
$\kappa_2 \gt 0$ depending on
$\beta$,
$\lambda$, and
$\gamma$, such that
\begin{equation}
\begin{split}
|\widehat{f}(\xi)| &= \delta |\xi|^{-\lambda/2} + O_\beta \left( N^{-1/2} \log(N) \cdot r^{-\beta/2} |\xi|^{-\beta/2} \right)\\
&\leq \left( \delta + O_{\beta,\kappa} \left( N^{-1/2} \log(N) \cdot r^{-\beta/2} |\xi|^{-(\beta - \lambda)/2} \right) \right) |\xi|^{- \lambda/2}\\
&\leq \left( \delta + O_{\beta,\kappa} \left( N^{-1/2} \log(N) r^{-\lambda/2} \right) \right) |\xi|^{-\lambda/2}\\
&\leq ( \delta + O_{\beta,\kappa} (N^{-\kappa_2(\beta,\varepsilon)}) ) |\xi|^{-\gamma/2}.
\end{split}
\end{equation} Thus if
$N$ is sufficiently large, then for
$(1/r) \leq |\xi| \leq (1/r)^{1+\kappa}$,
\begin{equation}
|\widehat{f}(\xi)| \leq 2 \delta |\xi|^{-\gamma/2}.
\end{equation} Finally, if
$|\xi| \geq (1/r)^{1 + \kappa}$, we apply (A.4) for
$T \geq \beta/2$ together with the bound
$\| \widehat{\eta} \|_{L^\infty(\operatorname{\mathbb{Z}}^d)} = 1$, which follows because
$\eta$ is a probability measure, to conclude that
\begin{equation}
\begin{split}
|\widehat{f}(\xi)| &\lesssim_T r^{-T} |\xi|^{-T}\\
&= r^{-T} |\xi|^{\beta/2 - T} \cdot |\xi|^{-\beta/2}\\
&\leq r^{-T} (1/r)^{(\beta/2 - T)(1 + \kappa)} \cdot |\xi|^{-\beta/2}\\
&= r^{\kappa T - (\beta/2)(1 + \kappa)} \cdot |\xi|^{-\beta/2}.
\end{split}
\end{equation} If we choose
$T \gt (\beta/2)(1 + 1/\kappa)$, then as
$r \to 0$,
$r^{\kappa T - (\beta/2)(1 + \kappa)} \to 0$. Thus we conclude from (A.28) that if
$N$ is sufficiently large, then for
$|\xi| \geq (1/r)^{1 + \kappa}$
\begin{equation}
|\widehat{f}(\xi)| \leq 2 \delta |\xi|^{-\beta/2} \leq 2 \delta |\xi|^{- \gamma/2}.
\end{equation} Combining (A.25), (A.27) and (A.29) shows that if
$N$ is sufficiently large,
Intuitively, if
$\delta \ll 1$, then the Fourier transform of
$f$ approximately looks like the Dirac delta function at the origin in
$\operatorname{\mathbb{T}}^d$, so we should expect
$f \approx 1$ on
$\operatorname{\mathbb{T}}^d$. In particular, we should expect that
$f \mu_0 \approx \mu_0$. Since
$\text{supp}(f \mu_0) \subset \text{supp}(f) \subset N(S,r)$, we know that
$\text{supp}(f \mu_0) \in \mathcal{A}$. Carrying out all these details numerically will complete the proof of density.
We start by applying Lemma A.3 using (A.30), which implies that if
$\rho = f \mu_0$, then
Using (A.30) and the fact that
$\widehat{\mu}_0 \in L^1(\operatorname{\mathbb{Z}}^d)$ because
$\mu_0 \in C^\infty(\operatorname{\mathbb{T}}^d)$, we find that
\begin{equation}
\begin{split}
\rho(\operatorname{\mathbb{T}}^d) &= (\widehat{f} * \widehat{\mu}_0)(0) \geq 1 - \sum_{|\xi| \neq 0} |\widehat{f}(\xi)| |\widehat{\mu}_0(-\xi)| \geq 1 - O_{\mu_0}(2 \delta).
\end{split}
\end{equation} Thus if we define
$\mu = \rho / \rho(\operatorname{\mathbb{T}}^d)$, then for
$\delta \leq 1$,
\begin{equation}
\begin{split}
\| \mu - \mu_0 \|_{M(\gamma)} &\leq \| \mu - \rho \|_{M(\gamma)} + \| \rho - \mu_0 \|_{M(\gamma)}\\
&= (1 / \rho(\operatorname{\mathbb{T}}^d) - 1) \cdot \| \rho \|_{M(\gamma)} + \| \rho - \mu_0 \|_{M(\gamma)}\\
&\lesssim_{\mu_0} \delta \| \rho \|_{M(\gamma)} + \| \rho - \mu_0 \|_{M(\gamma)}\\
&\lesssim \delta \| \mu_0 \|_{M(\gamma)} + \| \rho - \mu_0 \|_{M(\gamma)} \lesssim \delta.
\end{split}
\end{equation} If we take
$\delta$ suitably small, (A.33) implies that
$\| \mu - \mu_0 \|_{M(\gamma)} \leq \delta_0$. Since
$\text{supp}(\mu) \subset \text{supp}(\mu_0)$, Lemma A.4 implies that if
$\delta$ is taken even smaller, then
$d_{\mathbb{H}}(E, E_0) \leq \delta_0$. Thus if we set
$E = \text{supp}(\mu)$, then
$E \in \mathcal{A}$ since
$E \subset N(S,r)$ and
$\mathcal{A}$ is downward closed, and since
$\delta_0$ and
$\gamma$ were arbitrary, this completes the proof of density.










































