To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
This chapter gives an overview of the remainder of the book. We first provide commonsense and social science examples of reliability and validity, two necessary conditions that data must posses to have trustworthy conclusions based upon it. We next introduce Benford’s law and offer a brief overview of other social science studies that have employed it to check the accuracy of their data. We then turn to an overview of our Benford agreement analysis procedure and introduce the concept of Benford validity. The chapter concludes with a plan for the remainder of the book.
Here we develop a discussion of the concept of validity in the social sciences. We first highlight the history of validity and how it has been conceptualized and measured over time. Next we discuss a type of social science data that is often overlooked in the validity measurement and assessment literature: data that are based on self-reporting. Despite the widespread use of self-reported data in various social science disciplines such as economics, political science, and sociology, there are still few reported attempts to check data accuracy. By way of giving examples, we overview self-reported data in four areas: (1) US prison population data, (2) COVID-19 case data, (3) toxic releases, and (4) fish landings. We then discuss the need for a tool and for an established workflow for assessing the accuracy and validity of quantitative self-reported data in the social sciences. We suggest that applying Benford’s law to these types of data can provide a measure of validity assessment for data that would otherwise not be assessed for accuracy; then we briefly introduce the concept of Benford validity. We conclude the chapter with a short review of existing studies that have applied Benford’s law to social science data in some manner.
Chapter 8 concludes that Benford agreement analyses are a proper process for assessing the validity of self-reported data when these data meet certain characteristics. The Benford agreement analysis workflow developed in previous chapters is summarized. Recommendations as to when researchers should use Benford agreement analyses to assess their data for Benford validity are discussed. The chapter concludes with some thoughts on the utility of Benford validity analyses in the social sciences.
A spread-out lattice animal is a finite connected set of edges in $\{\{x,y\}\subset \mathbb{Z}^d\;:\;0\lt \|x-y\|\le L\}$. A lattice tree is a lattice animal with no loops. The best estimate on the critical point $p_{\textrm{c}}$ so far was achieved by Penrose (J. Stat. Phys. 77, 3–15, 1994) : $p_{\textrm{c}}=1/e+O(L^{-2d/7}\log L)$ for both models for all $d\ge 1$. In this paper, we show that $p_{\textrm{c}}=1/e+CL^{-d}+O(L^{-d-1})$ for all $d\gt 8$, where the model-dependent constant $C$ has the random-walk representation
where $U^{*n}$ is the $n$-fold convolution of the uniform distribution on the $d$-dimensional ball $\{x\in{\mathbb R}^d\;: \|x\|\le 1\}$. The proof is based on a novel use of the lace expansion for the 2-point function and detailed analysis of the 1-point function at a certain value of $p$ that is designed to make the analysis extremely simple.
In this paper, we consider subgeometric (specifically, polynomial) ergodicity of univariate nonlinear autoregressions with autoregressive conditional heteroskedasticity (ARCH). The notion of subgeometric ergodicity was introduced in the Markov chain literature in the 1980s, and it means that the transition probability measures converge to the stationary measure at a rate slower than geometric; this rate is also closely related to the convergence rate of $\beta $-mixing coefficients. While the existing literature on subgeometrically ergodic autoregressions assumes a homoskedastic error term, this paper provides an extension to the case of conditionally heteroskedastic ARCH-type errors, considerably widening the scope of potential applications. Specifically, we consider suitably defined higher-order nonlinear autoregressions with possibly nonlinear ARCH errors and show that they are, under appropriate conditions, subgeometrically ergodic at a polynomial rate. An empirical example using energy sector volatility index data illustrates the use of subgeometrically ergodic AR–ARCH models.
The coronavirus disease-2019 (COVID-19) pandemic and the mobility restrictions governments imposed to prevent its spread changed the cities’ ways of living. Transport systems suffered the consequences of the falling travel demand, and readjustments were made in many cities to prevent the complete shutdown of services. In Córdoba, the second largest city in Argentina, the Municipality dictated route cuts and reduced frequencies to sustain the buses and trolleys system. In 2022, Martinazzo and Falavigna assessed potential accessibility to hospitals before (2019) and during the pandemic (2021). Overall, the study indicated that average travel times increased by 20% and that the gap between less vulnerable and more vulnerable population quintiles reached almost 8 points. In this paper, potential accessibility to public hospitals in 2022 and 2023 is calculated using Martinazzo and Falavigna’s (2022) work as a baseline to compare, considering that neither cutting the services during the pandemic nor recovering the service after the pandemic the Municipality performed an accessibility assessment. The main results showed that, despite the system having almost recovered its extension by 2023, it maintained the regressive tendency between less vulnerable and more vulnerable population quintiles, as the difference in average travel time between these two groups reached up to 14 min, while the cumulative opportunities measure for the high-income groups was up to 68% higher than the most vulnerable households.
We consider the holder of an individual tontine retirement account, with maximum and minimum withdrawal amounts (per year) specified. The tontine account holder initiates the account at age 65 and earns mortality credits while alive, but forfeits all wealth in the account upon death. The holder wants to maximize total withdrawals and minimize expected shortfall at the end of the retirement horizon of 30 years (i.e., it is assumed that the holder survives to age 95). The holder controls the amount withdrawn each year and the fraction of the retirement portfolio invested in stocks and bonds. The optimal controls are determined based on a parametric model fitted to almost a century of market data. The optimal control algorithm is based on dynamic programming and the solution of a partial integro differential equation (PIDE) using Fourier methods. The optimal strategy (based on the parametric model) is tested out of sample using stationary block bootstrap resampling of the historical data. In terms of an expected total withdrawal, expected shortfall (EW-ES) efficient frontier, the tontine overlay dramatically outperforms an optimal strategy (without the tontine overlay), which in turn outperforms a constant weight strategy with withdrawals based on the ubiquitous four per cent rule.
The factorially normalized Bernoulli polynomials $b_n(x) = B_n(x)/n!$ are known to be characterized by $b_0(x) = 1$ and $b_n(x)$ for $n \gt 0$ is the anti-derivative of $b_{n-1}(x)$ subject to $\int _0^1 b_n(x) dx = 0$. We offer a related characterization: $b_1(x) = x - 1/2$ and $({-}1)^{n-1} b_n(x)$ for $n \gt 0$ is the $n$-fold circular convolution of $b_1(x)$ with itself. Equivalently, $1 - 2^n b_n(x)$ is the probability density at $x \in (0,1)$ of the fractional part of a sum of $n$ independent random variables, each with the beta$(1,2)$ probability density $2(1-x)$ at $x \in (0,1)$. This result has a novel combinatorial analog, the Bernoulli clock: mark the hours of a $2 n$ hour clock by a uniformly random permutation of the multiset $\{1,1, 2,2, \ldots, n,n\}$, meaning pick two different hours uniformly at random from the $2 n$ hours and mark them $1$, then pick two different hours uniformly at random from the remaining $2 n - 2$ hours and mark them $2$, and so on. Starting from hour $0 = 2n$, move clockwise to the first hour marked $1$, continue clockwise to the first hour marked $2$, and so on, continuing clockwise around the Bernoulli clock until the first of the two hours marked $n$ is encountered, at a random hour $I_n$ between $1$ and $2n$. We show that for each positive integer $n$, the event $( I_n = 1)$ has probability $(1 - 2^n b_n(0))/(2n)$, where $n! b_n(0) = B_n(0)$ is the $n$th Bernoulli number. For $ 1 \le k \le 2 n$, the difference $\delta _n(k)\,:\!=\, 1/(2n) -{\mathbb{P}}( I_n = k)$ is a polynomial function of $k$ with the surprising symmetry $\delta _n( 2 n + 1 - k) = ({-}1)^n \delta _n(k)$, which is a combinatorial analog of the well-known symmetry of Bernoulli polynomials $b_n(1-x) = ({-}1)^n b_n(x)$.