• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Confidence intervals for low dimensional parameters in high dimensional linear models. (2013)

by C-H Zhang, S S Zhang
Venue:J. R. Stat. Soc. B,
Add To MetaCart

Tools

Sorted by:
Results 1 - 10 of 28
Next 10 →

Confidence intervals and hypothesis testing for highdimensional regression. arXiv: 1306.3171

by Adel Javanmard, Andrea Montanari
"... Fitting high-dimensional statistical models often requires the use of non-linear parameter estimation procedures. As a consequence, it is generally impossible to obtain an exact characterization of the probability distribution of the parameter estimates. This in turn implies that it is extremely cha ..."
Abstract - Cited by 31 (3 self) - Add to MetaCart
Fitting high-dimensional statistical models often requires the use of non-linear parameter estimation procedures. As a consequence, it is generally impossible to obtain an exact characterization of the probability distribution of the parameter estimates. This in turn implies that it is extremely challenging to quantify the uncertainty associated with a certain parameter estimate. Concretely, no commonly accepted procedure exists for computing classical measures of uncertainty and statistical significance as confidence intervals or p-values. We consider here a broad class regression problems, and propose an efficient algorithm for constructing confidence intervals and p-values. The resulting confidence intervals have nearly optimal size. When testing for the null hypothesis that a certain parameter is vanishing, our method has nearly optimal power. Our approach is based on constructing a ‘de-biased ’ version of regularized M-estimators. The new construction improves over recent work in the field in that it does not assume a special structure on the design matrix. Furthermore, proofs are remarkably simple. We test our method on a diabetes prediction problem. 1
(Show Context)

Citation Context

...BvdG11] for an in-depth introduction to this area. The problem of quantifying statistical significance in high-dimensional parameter estimation is, by comparison, far less understood. Zhang and Zhang =-=[ZZ14]-=-, and Bühlmann [Büh13] proposed hypothesis testing procedures under restricted eigenvalue or compatibility conditions [BvdG11]. These papers provide deterministic guarantees but –in order to achieve...

A significance test for the lasso

by Richard Lockhart, Jonathan Taylor, Ryan J. Tibshirani, Robert Tibshirani
"... In the sparse linear regression setting, we consider testing the significance of the predictor variable that enters the current lasso model, in the sequence of models visited along the lasso solution path. We propose a simple test statistic based on lasso fitted values, called the covariance test st ..."
Abstract - Cited by 20 (2 self) - Add to MetaCart
In the sparse linear regression setting, we consider testing the significance of the predictor variable that enters the current lasso model, in the sequence of models visited along the lasso solution path. We propose a simple test statistic based on lasso fitted values, called the covariance test statistic, and show that when the true model is linear, this statistic has an Exp(1) asymptotic distribution under the null hypothesis (the null being that all truly active variables are contained in the current lasso model). Our proof of this result for the special case of the first predictor to enter the model (i.e., testing for a single significant predictor variable against the global null) requires only weak assumptions on the predictor matrix X. On the other hand, our proof for a general step in the lasso path places further technical assumptions on X and the generative model, but still allows for the important high-dimensional case p> n, and does not necessarily require that the current lasso model achieves perfect recovery of the truly active variables. Of course, for testing the significance of an additional variable between two nested linear models, one typically uses the chi-squared test, comparing the drop in residual sum of squares (RSS) to a χ 2 1 distribution. But when this additional variable is not fixed, and has been chosen adaptively or greedily, this test is no longer appropriate: adaptivity makes the drop in RSS stochastically much larger than χ 2 1 under the null hypothesis. Our analysis explicitly accounts for adaptivity, as it must, since the lasso builds an adaptive sequence of linear models as the tuning parameter λ decreases. In this analysis, shrinkage plays a key role: though additional variables are chosen adaptively, the coefficients of lasso active variables are shrunken due to the ℓ1 penalty. Therefore the test statistic (which is based on lasso fitted values) is in a sense balanced by these two opposing properties—adaptivity and shrinkage—and its null distribution is tractable and asymptotically Exp(1).

Sparse methods for biomedical data

by Jieping Ye, Jun Liu - SIGKDD Explor. Newsl , 2012
"... Following recent technological revolutions, the investigation of massive biomedical data with growing scale, diversity, and complexity has taken a center stage in modern data analysis. Although complex, the underlying representations of many biomedical data are often sparse. For example, for a certa ..."
Abstract - Cited by 12 (2 self) - Add to MetaCart
Following recent technological revolutions, the investigation of massive biomedical data with growing scale, diversity, and complexity has taken a center stage in modern data analysis. Although complex, the underlying representations of many biomedical data are often sparse. For example, for a certain disease such as leukemia, even though humans have tens of thousands of genes, only a few genes are relevant to the disease; a gene network is sparse since a regulatory pathway involves only a small number of genes; many biomedical signals are sparse or compressible in the sense that they have concise representations when expressed in a proper basis. Therefore, finding sparse representations is fundamentally important for scientific discovery. Sparse methods based on the ℓ1 norm have attracted a great amount of research efforts in the past decade due to its sparsity-inducing property, convenient convexity, and strong theoretical guarantees. They have achieved great success in various applications such as biomarker selection, biological network construction, and magnetic resonance imaging. In this paper, we review state-of-the-art sparse methods and their applications to biomedical data.
(Show Context)

Citation Context

...methods. Most existing work on sparse learning focus on prediction, parameter estimation, and variable selection. Very few work address the problem of assigning statistical significance or confidence =-=[11; 118]-=-. However, such significance or confidence measures are crucial in biomedical applications where interpretation of parameters and variables is very important [11]. Most sparse methods in the literatur...

Uniform Post Selection Inference for LAD Regression and Other Z-Estimation Problems.” arXiv e-print

by Alexandre Belloni , Victor Chernozhukov , Kengo Kato , A Belloni , V Chernozhukov , K Kato , 2013
"... Abstract. We develop uniformly valid confidence regions for a regression coefficient in a high-dimensional sparse LAD (least absolute deviation or median) regression model. The setting is one where the number of regressors could be large in comparison to the sample size , but only ≪ of them are nee ..."
Abstract - Cited by 7 (5 self) - Add to MetaCart
Abstract. We develop uniformly valid confidence regions for a regression coefficient in a high-dimensional sparse LAD (least absolute deviation or median) regression model. The setting is one where the number of regressors could be large in comparison to the sample size , but only ≪ of them are needed to accurately describe the regression function. Our new methods are based on the instrumental LAD regression estimator that assembles the optimal estimating equation from either post ℓ 1 -penalized LAD regression or ℓ 1 -penalized LAD regression. The estimating equation is immunized against non-regular estimation of nuisance part of the regression function, in the sense of Neyman. We establish that in a homoscedastic regression model, under certain conditions, the instrumental LAD regression estimator of the regression coefficient is asymptotically root-normal uniformly with respect to the underlying sparse model. The resulting confidence regions are valid uniformly with respect to the underlying model. The new inference methods outperform the naive, "oracle based" inference methods, which are known to be not uniformly valid -with coverage property failing to hold uniformly with respect the underlying model -even in the setting with = 2. We also provide Monte-Carlo experiments which demonstrate that standard post-selection inference breaks down over large parts of the parameter space, and the proposed method does not.
(Show Context)

Citation Context

.... The latter feature is critical in achieving uniformity over a large class of data generating processes, similarly to the results for instrumental regression and mean regression studied in [2], [7], =-=[32]-=-, [6]. This allows us to overcome the impact of (moderate) model selection mistakes on inference, avoiding (in part) the criticisms in [19], who prove that the “oracle property” sometime achieved by t...

Exact Post Model Selection Inference for Marginal Screening

by Jason D. Lee, Jonathan E. Taylor , 2014
"... We develop a framework for post model selection inference, via marginal screening, in linear regression. At the core of this framework is a result that characterizes the exact distribution of linear functions of the response y, conditional on the model being selected (“condi-tion on selection ” fram ..."
Abstract - Cited by 4 (2 self) - Add to MetaCart
We develop a framework for post model selection inference, via marginal screening, in linear regression. At the core of this framework is a result that characterizes the exact distribution of linear functions of the response y, conditional on the model being selected (“condi-tion on selection ” framework). This allows us to construct valid con-fidence intervals and hypothesis tests for regression coefficients that account for the selection procedure. In contrast to recent work in high-dimensional statistics, our results are exact (non-asymptotic) and re-quire no eigenvalue-like assumptions on the design matrix X. Further-more, the computational cost of marginal regression, constructing con-fidence intervals and hypothesis testing is negligible compared to the cost of linear regression, thus making our methods particularly suitable for extremely large datasets. Although we focus on marginal screening to illustrate the applicability of the condition on selection framework, this framework is much more broadly applicable. We show how to ap-ply the proposed framework to several other selection procedures in-cluding orthogonal matching pursuit, non-negative least squares, and marginal screening+Lasso. 1
(Show Context)

Citation Context

...ly computationally intensive and currently only applicable for p ≤ 30. A separate line of work establishes the asymptotic normality of a corrected estimator obtained by “inverting” the KKT conditions =-=[29, 32, 14]-=-. The corrected estimator b̂ has the form b̂ = β̂ + λΘ̂ẑ, where ẑ is a subgradient of the penalty at β̂ and Θ̂ is an approximate inverse to the Gram matrix XTX. The two main drawbacks to this approa...

A split-and-conquer approach for analysis of extraordinarily large data. Statistica Sinica Forthcoming.

by Xueying Chen , Min-Ge Xie , 2014
"... Abstract: If there are datasets, too large to fit into a single computer or too expensive for a computationally intensive data analysis, what should we do? We propose a split-and-conquer approach and illustrate it using several computationally intensive penalized regression methods, along with a th ..."
Abstract - Cited by 3 (0 self) - Add to MetaCart
Abstract: If there are datasets, too large to fit into a single computer or too expensive for a computationally intensive data analysis, what should we do? We propose a split-and-conquer approach and illustrate it using several computationally intensive penalized regression methods, along with a theoretical support. We show that the split-and-conquer approach can substantially reduce computing time and computer memory requirements. The proposed methodology is illustrated numerically using both simulation and data examples.
(Show Context)

Citation Context

...Thus, the combined estimator β (c) A is asymptotically as efficient as β (a) A . Together with the fact that both estimators are model selection consistent, the combined estimator β (c) A is asymptotically equivalent to β (a) A . The signal strength and sparsity assumptions in Theorem 2 depend on the number of splits. When K = O(1), asymptotic equivalence holds without any additional requirements on either, but with K going to infinity, we pay the price that stronger conditions are needed. If stronger signal strength is a concern in a specific problem, the two-stage estimation approach of Zhang and Zhang (2014) can perhaps be used to weaken the requirement. The strengthened conditions ensure that each subset contains enough data to provide a meaningful inference for the unknown model parameters. 3.3. Error control We provide an upper bound of the expected number of falsely selected variables and a lower bound of the expected number of truly selected variables. In Theorem 3 below, s∗ = supk sk and s∗ = infk sk, where sk = E(|Ak|) be the average number of selected variables of the penalized estimator from the kth subset. SPLIT-AND-CONQUER 1665 A similar result is provided by Fan, Samworth, and Wu ...

Gaussian graphical model estimation with false discovery rate control

by Weidong Liu - Annals of Statistics , 2013
"... This paper studies the estimation of high dimensional Gaussian graphical model (GGM). Typically, the existing methods depend on regularization techniques. As a result, it is necessary to choose the regularized parameter. However, the precise relationship between the regularized parameter and the num ..."
Abstract - Cited by 3 (1 self) - Add to MetaCart
This paper studies the estimation of high dimensional Gaussian graphical model (GGM). Typically, the existing methods depend on regularization techniques. As a result, it is necessary to choose the regularized parameter. However, the precise relationship between the regularized parameter and the number of false edges in GGM estimation is unclear. Hence, it is impossible to evaluate their performance rigorously. In this paper, we propose an alternative method by a multiple testing procedure. Based on our new test statistics for conditional dependence, we pro-pose a simultaneous testing procedure for conditional dependence in GGM. Our method can control the false discovery rate (FDR) asymptotically. The numerical performance of the proposed method shows that our method works quite well. 1

Inference in High Dimensions with the Penalized Score Test.” posted on arXiv: 1401.2678v1

by Arend Voorman, Ali Shojaie, Daniela Witten , 2014
"... In recent years, there has been considerable theoretical development regarding variable selection consistency of penalized regression techniques, such as the lasso. However, there has been relatively little work on quantifying the uncertainty in these selection procedures. In this paper, we propose ..."
Abstract - Cited by 3 (0 self) - Add to MetaCart
In recent years, there has been considerable theoretical development regarding variable selection consistency of penalized regression techniques, such as the lasso. However, there has been relatively little work on quantifying the uncertainty in these selection procedures. In this paper, we propose a new method for inference in high dimensions using a score test based on penalized regression. In this test, we perform penalized regression of an outcome on all but a single feature, and test for correlation of the residuals with the held-out feature. This procedure is applied to each feature in turn. Interestingly, when an `1 penalty is used, the sparsity pattern of the lasso corresponds exactly to a decision based on the proposed test. Further, when an `2 penalty is used, the test corresponds precisely to a score test in a mixed effects model, in which the effects of all but one feature are assumed to be random. We formulate the hypothesis being tested as a compromise between the null hypotheses tested in simple linear regression on each feature and in multiple linear regression on all features, and develop reference distributions for some well-known penalties. We also examine the behavior of the test on real and simulated data.

Asymptotic properties of lasso+mls and lasso+ridge in sparse high-dimensional linear regression

by Hanzhong Liu, Bin Yu - Electronic Journal of Statistics , 2013
"... ar ..."
Abstract - Cited by 2 (0 self) - Add to MetaCart
Abstract not found
(Show Context)

Citation Context

... high-dimensional setting. The bootstrap is a popular alternative in this case. A summary of bootstrap methods in linear and generalized linear penalized regression models for fixed p can be found in =-=[46]-=-. We will consider the p ≫ n case by proposing a new inference procedure: residual bootstrap after two stage estimators. Our method allows p to grow at an exponential rate in n. In the context of a re...

ROBUST INFERENCE IN HIGH-DIMENSIONAL APPROXIMATELY SPARSE QUANTILE REGRESSION MODELS

by A. Belloni, V. Chernozhukov, K. Kato
"... ar ..."
Abstract - Cited by 2 (0 self) - Add to MetaCart
Abstract not found
(Show Context)

Citation Context

...ion, [6] proposed a double selection inference in a parametric with homoscedastic Gaussian errors, [10] studies the double selection procedure in a non-parametric setting with heteroscedastic errors, =-=[39]-=- and [35] proposed estimators based on ℓ1penalized estimators based on “1-step” correction in parametric models. Going beyond mean models, [35] also provides high level conditions for the one-step est...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University