Regularized quantile regression (QR) is a useful technique for analyzing heterogeneous data under potentially heavy-tailed error contamination in high dimensions. This paper provides a new analysis of the estimation/prediction error bounds of the global solution of -regularized QR (QR-LASSO) and the local solutions of nonconvex regularized QR (QR-NCP) when the number of covariates is greater than the sample size. Our results build upon and significantly generalize the earlier work in the literature. For certain heavy-tailed error distributions and a general class of design matrices, the least-squares-based LASSO cannot achieve the near-oracle rate derived under the normality assumption no matter the choice of the tuning parameter. In contrast, we establish that QR-LASSO achieves the near-oracle estimation error rate for a broad class of models under conditions weaker than those in the literature. For QR-NCP, we establish the novel results that all local optima within a feasible region have desirable estimation accuracy. Our analysis applies to not just the hard sparsity setting commonly used in the literature, but also to the soft sparsity setting which permits many small coefficients. Our approach relies on a unified characterization of the global/local solutions of regularized QR via subgradients using a generalized Karush–Kuhn–Tucker condition. The theory of the paper establishes a key property of the subdifferential of the quantile loss function in high dimensions, which is of independent interest for analyzing other high-dimensional nonsmooth problems.
]]>We propose a series-based nonparametric specification test for a regression function when data are spatially dependent, the “space” being of a general economic or social nature. Dependence can be parametric, parametric with increasing dimension, semiparametric or any combination thereof, thus covering a vast variety of settings. These include spatial error models of varying types and levels of complexity. Under a new smooth spatial dependence condition, our test statistic is asymptotically standard normal. To prove the latter property, we establish a central limit theorem for quadratic forms in linear processes in an increasing dimension setting. Finite sample performance is investigated in a simulation study, with a bootstrap method also justified and illustrated. Empirical examples illustrate the test with real-world data.
]]>We use mollification to regularize the problem of deconvolution of random variables. This regularization method offers a unifying and generalizing framework in order to compare the benefits of various filter-type techniques like deconvolution kernels, Tikhonov, or spectral cutoff methods. In particular, the mollifier approach allows to relax some restrictive assumptions required for the deconvolution kernels, and has better stabilizing properties compared with spectral cutoff or Tikhonov. We show that this approach achieves optimal rates of convergence for both finitely and infinitely smoothing convolution operators under Besov and Sobolev smoothness assumptions on the unknown probability density. The qualification can be arbitrarily high depending on the choice of the mollifier function. We propose an adaptive choice of the regularization parameter using the Lepskiĭ method, and we provide simulations to compare the finite sample properties of our estimator with respect to the well-known regularization methods.
]]>In a dynamic panel data model, the number of moment conditions increases rapidly with the time dimension, resulting in a large dimensional covariance matrix of the instruments. As a consequence, the generalized method of moments (GMM) estimator exhibits a large bias in small samples, especially when the autoregressive parameter is close to unity. To address this issue, we propose a regularized version of the one-step GMM estimator using three regularization schemes based on three different ways of inverting the covariance matrix of the instruments. Under double asymptotics, we show that our regularized estimators are consistent and asymptotically normal. These regularization schemes involve a tuning or regularization parameter which needs to be chosen. We derive a data-driven selection of this regularization parameter based on an approximation of the higher-order mean square error and show its optimality. As an empirical application, we estimate a model of income dynamics.
]]>This paper considers the estimation of panel data models with interactive fixed effects where the idiosyncratic errors are subject to conditional quantile restrictions. An easy-to-implement two-step estimator is proposed for the coefficients of the observed regressors. In the first step, the principal component analysis is applied to the cross-sectional averages of the regressors to estimate the latent factors. In the second step, the smoothed quantile regression is used to estimate the coefficients of the observed regressors and the factor loadings jointly. The consistency and asymptotic normality of the estimator are established under large asymptotics. It is found that the asymptotic distribution of the estimator suffers from asymptotic biases, and this paper shows how to correct the biases using both analytical and split-panel jackknife bias corrections. Simulation studies confirm that the proposed estimator performs well with moderate sample sizes.
]]>This paper proposes a jackknife Lagrange multiplier (JLM) test for instrumental variable regression models, which is robust to (i) many instruments, where the number of instruments may increase proportionally with the sample size, (ii) arbitrarily weak instruments, and (iii) heteroskedastic errors. In contrast to Crudu, Mellace, and Sándor (2021, Econometric Theory 37, 281–310) and Mikusheva and Sun (2021, Review of Economic Studies 89, 2663–2686), who proposed jackknife Anderson–Rubin tests that are also robust to (i)–(iii), we modify a score statistic by jackknifing and construct its heteroskedasticity robust variance estimator. Compared to the Lagrange multiplier tests by Kleibergen (2002, Econometrica 70, 1781–1803) and Moreira (2001, Tests with Correct Size when Instruments Can Be Arbitrarily Weak, Working paper) and their modification for many instruments by Hansen, Hausman, and Newey (2008, Journal of Business & Economic Statistics 26, 398–422), our JLM test is robust to heteroskedastic errors and may circumvent a possible decrease in the power function. Simulation results illustrate the desirable size and power properties of the proposed method.
]]>