Cross Validation And Regression Analysis In High Dimensional Sparse Linear Models PDF Download

Are you looking for read ebook online? Search for your book and save it on your Kindle device, PC, phones or tablets. Download Cross Validation And Regression Analysis In High Dimensional Sparse Linear Models PDF full book. Access full book title Cross Validation And Regression Analysis In High Dimensional Sparse Linear Models.

Cross-validation and Regression Analysis in High-dimensional Sparse Linear Models

Cross-validation and Regression Analysis in High-dimensional Sparse Linear Models
Author: Feng Zhang
Publisher: Stanford University
Total Pages: 91
Release: 2011
Genre:
ISBN:

Download Cross-validation and Regression Analysis in High-dimensional Sparse Linear Models Book in PDF, ePub and Kindle

Modern scientific research often involves experiments with at most hundreds of subjects but with tens of thousands of variables for every subject. The challenge of high dimensionality has reshaped statistical thinking and modeling. Variable selection plays a pivotal role in the high-dimensional data analysis, and the combination of sparsity and accuracy is crucial for statistical theory and practical applications. Regularization methods are attractive for tackling these sparsity and accuracy issues. The first part of this thesis studies two regularization methods. First, we consider the orthogonal greedy algorithm (OGA) used in conjunction with a high-dimensional information criterion introduced by Ing& Lai (2011). Although it has been shown to have excellent performance for weakly sparse regression models, one does not know a priori in practice that the actual model is weakly sparse, and we address this problem by developing a new cross-validation approach. OGA can be viewed as L0 regularization for weakly sparse regression models. When such sparsity fails, as revealed by the cross-validation analysis, we propose to use a new way to combine L1 and L2 penalties, which we show to have important advantages over previous regularization methods. The second part of the thesis develops a Monte Carlo Cross-Validation (MCCV) method to estimate the distribution of out-of-sample prediction errors when a training sample is used to build a regression model for prediction. Asymptotic theory and simulation studies show that the proposed MCCV method mimics the actual (but unknown) prediction error distribution even when the number of regressors exceeds the sample size. Therefore MCCV provides a useful tool for comparing the predictive performance of different regularization methods for real (rather than simulated) data sets.


Cross-validation and Regression Analysis in High-dimensional Sparse Linear Models

Cross-validation and Regression Analysis in High-dimensional Sparse Linear Models
Author: Feng Zhang
Publisher:
Total Pages:
Release: 2011
Genre:
ISBN:

Download Cross-validation and Regression Analysis in High-dimensional Sparse Linear Models Book in PDF, ePub and Kindle

Modern scientific research often involves experiments with at most hundreds of subjects but with tens of thousands of variables for every subject. The challenge of high dimensionality has reshaped statistical thinking and modeling. Variable selection plays a pivotal role in the high-dimensional data analysis, and the combination of sparsity and accuracy is crucial for statistical theory and practical applications. Regularization methods are attractive for tackling these sparsity and accuracy issues. The first part of this thesis studies two regularization methods. First, we consider the orthogonal greedy algorithm (OGA) used in conjunction with a high-dimensional information criterion introduced by Ing & Lai (2011). Although it has been shown to have excellent performance for weakly sparse regression models, one does not know a priori in practice that the actual model is weakly sparse, and we address this problem by developing a new cross-validation approach. OGA can be viewed as L0 regularization for weakly sparse regression models. When such sparsity fails, as revealed by the cross-validation analysis, we propose to use a new way to combine L1 and L2 penalties, which we show to have important advantages over previous regularization methods. The second part of the thesis develops a Monte Carlo Cross-Validation (MCCV) method to estimate the distribution of out-of-sample prediction errors when a training sample is used to build a regression model for prediction. Asymptotic theory and simulation studies show that the proposed MCCV method mimics the actual (but unknown) prediction error distribution even when the number of regressors exceeds the sample size. Therefore MCCV provides a useful tool for comparing the predictive performance of different regularization methods for real (rather than simulated) data sets.


Partially Linear Models

Partially Linear Models
Author: Wolfgang Härdle
Publisher: Springer Science & Business Media
Total Pages: 210
Release: 2012-12-06
Genre: Mathematics
ISBN: 3642577008

Download Partially Linear Models Book in PDF, ePub and Kindle

In the last ten years, there has been increasing interest and activity in the general area of partially linear regression smoothing in statistics. Many methods and techniques have been proposed and studied. This monograph hopes to bring an up-to-date presentation of the state of the art of partially linear regression techniques. The emphasis is on methodologies rather than on the theory, with a particular focus on applications of partially linear regression techniques to various statistical problems. These problems include least squares regression, asymptotically efficient estimation, bootstrap resampling, censored data analysis, linear measurement error models, nonlinear measurement models, nonlinear and nonparametric time series models.


Statistical Learning with Sparsity

Statistical Learning with Sparsity
Author: Trevor Hastie
Publisher: CRC Press
Total Pages: 354
Release: 2015-05-07
Genre: Business & Economics
ISBN: 1498712177

Download Statistical Learning with Sparsity Book in PDF, ePub and Kindle

Discover New Methods for Dealing with High-Dimensional DataA sparse statistical model has only a small number of nonzero parameters or weights; therefore, it is much easier to estimate and interpret than a dense model. Statistical Learning with Sparsity: The Lasso and Generalizations presents methods that exploit sparsity to help recover the underl


Machine Learning Techniques for Gait Biometric Recognition

Machine Learning Techniques for Gait Biometric Recognition
Author: James Eric Mason
Publisher: Springer
Total Pages: 247
Release: 2016-02-04
Genre: Technology & Engineering
ISBN: 3319290886

Download Machine Learning Techniques for Gait Biometric Recognition Book in PDF, ePub and Kindle

This book focuses on how machine learning techniques can be used to analyze and make use of one particular category of behavioral biometrics known as the gait biometric. A comprehensive Ground Reaction Force (GRF)-based Gait Biometrics Recognition framework is proposed and validated by experiments. In addition, an in-depth analysis of existing recognition techniques that are best suited for performing footstep GRF-based person recognition is also proposed, as well as a comparison of feature extractors, normalizers, and classifiers configurations that were never directly compared with one another in any previous GRF recognition research. Finally, a detailed theoretical overview of many existing machine learning techniques is presented, leading to a proposal of two novel data processing techniques developed specifically for the purpose of gait biometric recognition using GRF. This book · introduces novel machine-learning-based temporal normalization techniques · bridges research gaps concerning the effect of footwear and stepping speed on footstep GRF-based person recognition · provides detailed discussions of key research challenges and open research issues in gait biometrics recognition · compares biometrics systems trained and tested with the same footwear against those trained and tested with different footwear


Approximate Cross Validation for Sparse Generalized Linear Models

Approximate Cross Validation for Sparse Generalized Linear Models
Author: William Thomas Stephenson
Publisher:
Total Pages: 60
Release: 2019
Genre:
ISBN:

Download Approximate Cross Validation for Sparse Generalized Linear Models Book in PDF, ePub and Kindle

Cross validation (CV) is an effective yet computationally expensive tool for assessing the out of sample error for many methods in machine learning and statistics. Previous work has shown that methods to approximate CV can be very accurate and computationally cheap, but only for low dimensional problems. In this thesis, a modification of existing methods is developed to extend the high accuracy of these techniques to high dimensional settings.


Doctoral Research in Construction Management

Doctoral Research in Construction Management
Author: Zhen Chen
Publisher: Frontiers Media SA
Total Pages: 156
Release: 2023-02-17
Genre: Technology & Engineering
ISBN: 2832515029

Download Doctoral Research in Construction Management Book in PDF, ePub and Kindle


Partial Least Squares Regression

Partial Least Squares Regression
Author: R. Dennis Cook
Publisher: CRC Press
Total Pages: 448
Release: 2024-07-17
Genre: Mathematics
ISBN: 1040051324

Download Partial Least Squares Regression Book in PDF, ePub and Kindle

Partial least squares (PLS) regression is, at its historical core, a black-box algorithmic method for dimension reduction and prediction based on an underlying linear relationship between a possibly vector-valued response and a number of predictors. Through envelopes, much more has been learned about PLS regression, resulting in a mass of information that allows an envelope bridge that takes PLS regression from a black-box algorithm to a core statistical paradigm based on objective function optimization and, more generally, connects the applied sciences and statistics in the context of PLS. This book focuses on developing this bridge. It also covers uses of PLS outside of linear regression, including discriminant analysis, non-linear regression, generalized linear models and dimension reduction generally. Key Features: • Showcases the first serviceable method for studying high-dimensional regressions. • Provides necessary background on PLS and its origin. • R and Python programs are available for nearly all methods discussed in the book. This book can be used as a reference and as a course supplement at the Master's level in Statistics and beyond. It will be of interest to both statisticians and applied scientists.


Data Science for Financial Econometrics

Data Science for Financial Econometrics
Author: Nguyen Ngoc Thach
Publisher: Springer Nature
Total Pages: 633
Release: 2020-11-13
Genre: Computers
ISBN: 3030488535

Download Data Science for Financial Econometrics Book in PDF, ePub and Kindle

This book offers an overview of state-of-the-art econometric techniques, with a special emphasis on financial econometrics. There is a major need for such techniques, since the traditional way of designing mathematical models – based on researchers’ insights – can no longer keep pace with the ever-increasing data flow. To catch up, many application areas have begun relying on data science, i.e., on techniques for extracting models from data, such as data mining, machine learning, and innovative statistics. In terms of capitalizing on data science, many application areas are way ahead of economics. To close this gap, the book provides examples of how data science techniques can be used in economics. Corresponding techniques range from almost traditional statistics to promising novel ideas such as quantum econometrics. Given its scope, the book will appeal to students and researchers interested in state-of-the-art developments, and to practitioners interested in using data science techniques.